Skip to content

Commit

Permalink
test block quote
Browse files Browse the repository at this point in the history
  • Loading branch information
UziTech committed Apr 20, 2024
1 parent c6a98ea commit a988a86
Show file tree
Hide file tree
Showing 5 changed files with 72 additions and 23 deletions.
9 changes: 4 additions & 5 deletions src/Lexer.ts
Original file line number Diff line number Diff line change
Expand Up @@ -101,9 +101,9 @@ export class _Lexer {
/**
* Lexing
*/
blockTokens(src: string, tokens?: Token[]): Token[];
blockTokens(src: string, tokens?: TokensList): TokensList;
blockTokens(src: string, tokens: Token[] = []) {
blockTokens(src: string, tokens?: Token[], lastParagraphClipped?: boolean): Token[];
blockTokens(src: string, tokens?: TokensList, lastParagraphClipped?: boolean): TokensList;
blockTokens(src: string, tokens: Token[] = [], lastParagraphClipped = false) {
if (this.options.pedantic) {
src = src.replace(/\t/g, ' ').replace(/^ +$/gm, '');
} else {
Expand All @@ -115,7 +115,6 @@ export class _Lexer {
let token: Tokens.Generic | undefined;
let lastToken;
let cutSrc;
let lastParagraphClipped;

while (src) {
if (this.options.extensions
Expand Down Expand Up @@ -249,7 +248,7 @@ export class _Lexer {
}
if (this.state.top && (token = this.tokenizer.paragraph(cutSrc))) {
lastToken = tokens[tokens.length - 1];
if (lastParagraphClipped && lastToken.type === 'paragraph') {
if (lastParagraphClipped && lastToken?.type === 'paragraph') {
lastToken.raw += '\n' + token.raw;
lastToken.text += '\n' + token.text;
this.inlineQueue.pop();
Expand Down
71 changes: 62 additions & 9 deletions src/Tokenizer.ts
Original file line number Diff line number Diff line change
Expand Up @@ -7,7 +7,7 @@ import {
} from './helpers.ts';
import type { Rules } from './rules.ts';
import type { _Lexer } from './Lexer.ts';
import type { Links, Tokens } from './Tokens.ts';
import type { Links, Tokens, Token } from './Tokens.ts';
import type { MarkedOptions } from './MarkedOptions.ts';

function outputLink(cap: string[], link: Pick<Tokens.Link, 'href' | 'title'>, raw: string, lexer: _Lexer): Tokens.Link | Tokens.Image {
Expand Down Expand Up @@ -156,16 +156,69 @@ export class _Tokenizer {
blockquote(src: string): Tokens.Blockquote | undefined {
const cap = this.rules.block.blockquote.exec(src);
if (cap) {
// precede setext continuation with 4 spaces so it isn't a setext
let text = cap[0].replace(/\n {0,3}((?:=+|-+) *)(?=\n|$)/g, '\n $1');
text = rtrim(text.replace(/^ *>[ \t]?/gm, ''), '\n');
const top = this.lexer.state.top;
this.lexer.state.top = true;
const tokens = this.lexer.blockTokens(text);
this.lexer.state.top = top;
let lines = rtrim(cap[0], '\n').split('\n');
let raw = '';
let text = '';
const tokens: Token[] = [];

while (lines.length > 0) {
let inBlockquote = false;
const currentLines = [];

while (lines.length > 0) {
if (/^ {0,3}>/.test(lines[0])) {
currentLines.push(lines.shift());
inBlockquote = true;
} else if (!inBlockquote) {
currentLines.push(lines.shift());
} else {
break;
}
}

const currentRaw = currentLines.join('\n');
const currentText = currentRaw
// precede setext continuation with 4 spaces so it isn't a setext
.replace(/\n {0,3}((?:=+|-+) *)(?=\n|$)/g, '\n $1')
.replace(/^ {0,3}>[ \t]?/gm, '');
raw = raw ? `${raw}\n${currentRaw}` : currentRaw;
text = text ? `${text}\n${currentText}` : currentText;
const top = this.lexer.state.top;
this.lexer.state.top = true;
this.lexer.blockTokens(currentText, tokens, true);
this.lexer.state.top = top;

if (lines.length === 0) {
break;
}

const lastToken = tokens[tokens.length - 1];

if (lastToken?.type === 'code') {
break;
} else if (lastToken?.type === 'blockquote') {
const newLines = lastToken.raw + '\n' + lines.join('\n');

tokens[tokens.length - 1] = this.blockquote(newLines)!; // TODO: remove `!`

raw = raw.substring(0, raw.length - lastToken.raw.length) + tokens[tokens.length - 1].raw;
text = text.substring(0, text.length - (lastToken as Tokens.Blockquote).text.length) + (tokens[tokens.length - 1] as Tokens.Blockquote).text;
break;
} else if (lastToken?.type === 'list') {
const newLines = lastToken.raw + '\n' + lines.join('\n');

tokens[tokens.length - 1] = this.list(newLines)!; // TODO: remove `!`

raw = raw.substring(0, raw.length - lastToken.raw.length) + tokens[tokens.length - 1].raw;
text = text.substring(0, text.length - (lastToken as Tokens.List).raw.length) + (tokens[tokens.length - 1] as Tokens.List).raw;
lines = newLines.substring(tokens[tokens.length - 1].raw.length).split('\n');
continue;
}
}

return {
type: 'blockquote',
raw: cap[0],
raw,
tokens,
text
};
Expand Down
6 changes: 2 additions & 4 deletions test/specs/commonmark/commonmark.0.31.json
Original file line number Diff line number Diff line change
Expand Up @@ -1887,17 +1887,15 @@
"example": 236,
"start_line": 3838,
"end_line": 3848,
"section": "Block quotes",
"shouldFail": true
"section": "Block quotes"
},
{
"markdown": "> ```\nfoo\n```\n",
"html": "<blockquote>\n<pre><code></code></pre>\n</blockquote>\n<p>foo</p>\n<pre><code></code></pre>\n",
"example": 237,
"start_line": 3851,
"end_line": 3861,
"section": "Block quotes",
"shouldFail": true
"section": "Block quotes"
},
{
"markdown": "> foo\n - bar\n",
Expand Down
6 changes: 2 additions & 4 deletions test/specs/gfm/commonmark.0.31.json
Original file line number Diff line number Diff line change
Expand Up @@ -1887,17 +1887,15 @@
"example": 236,
"start_line": 3838,
"end_line": 3848,
"section": "Block quotes",
"shouldFail": true
"section": "Block quotes"
},
{
"markdown": "> ```\nfoo\n```\n",
"html": "<blockquote>\n<pre><code></code></pre>\n</blockquote>\n<p>foo</p>\n<pre><code></code></pre>\n",
"example": 237,
"start_line": 3851,
"end_line": 3861,
"section": "Block quotes",
"shouldFail": true
"section": "Block quotes"
},
{
"markdown": "> foo\n - bar\n",
Expand Down
3 changes: 2 additions & 1 deletion test/unit/marked.test.js
Original file line number Diff line number Diff line change
Expand Up @@ -18,7 +18,7 @@ describe('marked unit', () => {

assert.strictEqual(tokens[0].type, 'paragraph');
assert.strictEqual(tokens[2].tokens[0].type, 'paragraph');
assert.strictEqual(tokens[3].items[0].tokens[0].type, 'text');
assert.strictEqual(tokens[4].items[0].tokens[0].type, 'text');
});
});

Expand Down Expand Up @@ -924,6 +924,7 @@ br
['blockquote', '> blockquote'],
['paragraph', 'blockquote'],
['text', 'blockquote'],
['space', ''],
['list', '- list'],
['list_item', '- list'],
['text', 'list'],
Expand Down

0 comments on commit a988a86

Please sign in to comment.