587 lines
		
	
	
		
			21 KiB
		
	
	
	
		
			TypeScript
		
	
	
	
	
	
			
		
		
	
	
			587 lines
		
	
	
		
			21 KiB
		
	
	
	
		
			TypeScript
		
	
	
	
	
	
| import {
 | |
|   ddescribe,
 | |
|   describe,
 | |
|   it,
 | |
|   iit,
 | |
|   xit,
 | |
|   expect,
 | |
|   beforeEach,
 | |
|   afterEach
 | |
| } from 'angular2/testing_internal';
 | |
| import {BaseException} from 'angular2/src/facade/exceptions';
 | |
| 
 | |
| import {
 | |
|   tokenizeHtml,
 | |
|   HtmlToken,
 | |
|   HtmlTokenType,
 | |
|   HtmlTokenError
 | |
| } from 'angular2/src/compiler/html_lexer';
 | |
| import {ParseSourceSpan, ParseLocation, ParseSourceFile} from 'angular2/src/compiler/parse_util';
 | |
| 
 | |
| export function main() {
 | |
|   describe('HtmlLexer', () => {
 | |
|     describe('line/column numbers', () => {
 | |
|       it('should work without newlines', () => {
 | |
|         expect(tokenizeAndHumanizeLineColumn('<t>a</t>'))
 | |
|             .toEqual([
 | |
|               [HtmlTokenType.TAG_OPEN_START, '0:0'],
 | |
|               [HtmlTokenType.TAG_OPEN_END, '0:2'],
 | |
|               [HtmlTokenType.TEXT, '0:3'],
 | |
|               [HtmlTokenType.TAG_CLOSE, '0:4'],
 | |
|               [HtmlTokenType.EOF, '0:8']
 | |
|             ]);
 | |
|       });
 | |
| 
 | |
|       it('should work with one newline', () => {
 | |
|         expect(tokenizeAndHumanizeLineColumn('<t>\na</t>'))
 | |
|             .toEqual([
 | |
|               [HtmlTokenType.TAG_OPEN_START, '0:0'],
 | |
|               [HtmlTokenType.TAG_OPEN_END, '0:2'],
 | |
|               [HtmlTokenType.TEXT, '0:3'],
 | |
|               [HtmlTokenType.TAG_CLOSE, '1:1'],
 | |
|               [HtmlTokenType.EOF, '1:5']
 | |
|             ]);
 | |
|       });
 | |
| 
 | |
|       it('should work with multiple newlines', () => {
 | |
|         expect(tokenizeAndHumanizeLineColumn('<t\n>\na</t>'))
 | |
|             .toEqual([
 | |
|               [HtmlTokenType.TAG_OPEN_START, '0:0'],
 | |
|               [HtmlTokenType.TAG_OPEN_END, '1:0'],
 | |
|               [HtmlTokenType.TEXT, '1:1'],
 | |
|               [HtmlTokenType.TAG_CLOSE, '2:1'],
 | |
|               [HtmlTokenType.EOF, '2:5']
 | |
|             ]);
 | |
|       });
 | |
|     });
 | |
| 
 | |
|     describe('comments', () => {
 | |
|       it('should parse comments', () => {
 | |
|         expect(tokenizeAndHumanizeParts('<!--test-->'))
 | |
|             .toEqual([
 | |
|               [HtmlTokenType.COMMENT_START],
 | |
|               [HtmlTokenType.RAW_TEXT, 'test'],
 | |
|               [HtmlTokenType.COMMENT_END],
 | |
|               [HtmlTokenType.EOF]
 | |
|             ]);
 | |
|       });
 | |
| 
 | |
|       it('should store the locations', () => {expect(tokenizeAndHumanizeSourceSpans('<!--test-->'))
 | |
|                                                   .toEqual([
 | |
|                                                     [HtmlTokenType.COMMENT_START, '<!--'],
 | |
|                                                     [HtmlTokenType.RAW_TEXT, 'test'],
 | |
|                                                     [HtmlTokenType.COMMENT_END, '-->'],
 | |
|                                                     [HtmlTokenType.EOF, '']
 | |
|                                                   ])});
 | |
| 
 | |
|       it('should report <!- without -', () => {
 | |
|         expect(tokenizeAndHumanizeErrors('<!-a'))
 | |
|             .toEqual([[HtmlTokenType.COMMENT_START, 'Unexpected character "a"', '0:3']]);
 | |
|       });
 | |
| 
 | |
|       it('should report missing end comment', () => {
 | |
|         expect(tokenizeAndHumanizeErrors('<!--'))
 | |
|             .toEqual([[HtmlTokenType.RAW_TEXT, 'Unexpected character "EOF"', '0:4']]);
 | |
|       });
 | |
|     });
 | |
| 
 | |
|     describe('doctype', () => {
 | |
|       it('should parse doctypes', () => {
 | |
|         expect(tokenizeAndHumanizeParts('<!doctype html>'))
 | |
|             .toEqual([[HtmlTokenType.DOC_TYPE, 'doctype html'], [HtmlTokenType.EOF]]);
 | |
|       });
 | |
| 
 | |
|       it('should store the locations', () => {
 | |
|         expect(tokenizeAndHumanizeSourceSpans('<!doctype html>'))
 | |
|             .toEqual([[HtmlTokenType.DOC_TYPE, '<!doctype html>'], [HtmlTokenType.EOF, '']]);
 | |
|       });
 | |
| 
 | |
|       it('should report missing end doctype', () => {
 | |
|         expect(tokenizeAndHumanizeErrors('<!'))
 | |
|             .toEqual([[HtmlTokenType.DOC_TYPE, 'Unexpected character "EOF"', '0:2']]);
 | |
|       });
 | |
|     });
 | |
| 
 | |
|     describe('cdata', () => {
 | |
|       it('should parse cdata', () => {
 | |
|         expect(tokenizeAndHumanizeParts('<![cdata[test]]>'))
 | |
|             .toEqual([
 | |
|               [HtmlTokenType.CDATA_START],
 | |
|               [HtmlTokenType.RAW_TEXT, 'test'],
 | |
|               [HtmlTokenType.CDATA_END],
 | |
|               [HtmlTokenType.EOF]
 | |
|             ]);
 | |
|       });
 | |
| 
 | |
|       it('should store the locations', () => {
 | |
|         expect(tokenizeAndHumanizeSourceSpans('<![cdata[test]]>'))
 | |
|             .toEqual([
 | |
|               [HtmlTokenType.CDATA_START, '<![cdata['],
 | |
|               [HtmlTokenType.RAW_TEXT, 'test'],
 | |
|               [HtmlTokenType.CDATA_END, ']]>'],
 | |
|               [HtmlTokenType.EOF, '']
 | |
|             ]);
 | |
|       });
 | |
| 
 | |
|       it('should report <![ without cdata[', () => {
 | |
|         expect(tokenizeAndHumanizeErrors('<![a'))
 | |
|             .toEqual([[HtmlTokenType.CDATA_START, 'Unexpected character "a"', '0:3']]);
 | |
|       });
 | |
| 
 | |
|       it('should report missing end cdata', () => {
 | |
|         expect(tokenizeAndHumanizeErrors('<![cdata['))
 | |
|             .toEqual([[HtmlTokenType.RAW_TEXT, 'Unexpected character "EOF"', '0:9']]);
 | |
|       });
 | |
|     });
 | |
| 
 | |
|     describe('open tags', () => {
 | |
|       it('should parse open tags without prefix', () => {
 | |
|         expect(tokenizeAndHumanizeParts('<test>'))
 | |
|             .toEqual([
 | |
|               [HtmlTokenType.TAG_OPEN_START, null, 'test'],
 | |
|               [HtmlTokenType.TAG_OPEN_END],
 | |
|               [HtmlTokenType.EOF]
 | |
|             ]);
 | |
|       });
 | |
| 
 | |
|       it('should parse namespace prefix', () => {
 | |
|         expect(tokenizeAndHumanizeParts('<ns1:test>'))
 | |
|             .toEqual([
 | |
|               [HtmlTokenType.TAG_OPEN_START, 'ns1', 'test'],
 | |
|               [HtmlTokenType.TAG_OPEN_END],
 | |
|               [HtmlTokenType.EOF]
 | |
|             ]);
 | |
|       });
 | |
| 
 | |
|       it('should parse void tags', () => {
 | |
|         expect(tokenizeAndHumanizeParts('<test/>'))
 | |
|             .toEqual([
 | |
|               [HtmlTokenType.TAG_OPEN_START, null, 'test'],
 | |
|               [HtmlTokenType.TAG_OPEN_END_VOID],
 | |
|               [HtmlTokenType.EOF]
 | |
|             ]);
 | |
|       });
 | |
| 
 | |
|       it('should allow whitespace', () => {
 | |
|         expect(tokenizeAndHumanizeParts('< test >'))
 | |
|             .toEqual([
 | |
|               [HtmlTokenType.TAG_OPEN_START, null, 'test'],
 | |
|               [HtmlTokenType.TAG_OPEN_END],
 | |
|               [HtmlTokenType.EOF]
 | |
|             ]);
 | |
|       });
 | |
| 
 | |
|       it('should store the locations', () => {
 | |
|         expect(tokenizeAndHumanizeSourceSpans('<test>'))
 | |
|             .toEqual([
 | |
|               [HtmlTokenType.TAG_OPEN_START, '<test'],
 | |
|               [HtmlTokenType.TAG_OPEN_END, '>'],
 | |
|               [HtmlTokenType.EOF, '']
 | |
|             ]);
 | |
|       });
 | |
| 
 | |
|       it('should report missing name after <', () => {
 | |
|         expect(tokenizeAndHumanizeErrors('<'))
 | |
|             .toEqual([[HtmlTokenType.TAG_OPEN_START, 'Unexpected character "EOF"', '0:1']]);
 | |
|       });
 | |
| 
 | |
|       it('should report missing >', () => {
 | |
|         expect(tokenizeAndHumanizeErrors('<name'))
 | |
|             .toEqual([[HtmlTokenType.TAG_OPEN_START, 'Unexpected character "EOF"', '0:5']]);
 | |
|       });
 | |
|     });
 | |
| 
 | |
|     describe('attributes', () => {
 | |
|       it('should parse attributes without prefix', () => {
 | |
|         expect(tokenizeAndHumanizeParts('<t a>'))
 | |
|             .toEqual([
 | |
|               [HtmlTokenType.TAG_OPEN_START, null, 't'],
 | |
|               [HtmlTokenType.ATTR_NAME, null, 'a'],
 | |
|               [HtmlTokenType.TAG_OPEN_END],
 | |
|               [HtmlTokenType.EOF]
 | |
|             ]);
 | |
|       });
 | |
| 
 | |
|       it('should parse attributes with prefix', () => {
 | |
|         expect(tokenizeAndHumanizeParts('<t ns1:a>'))
 | |
|             .toEqual([
 | |
|               [HtmlTokenType.TAG_OPEN_START, null, 't'],
 | |
|               [HtmlTokenType.ATTR_NAME, 'ns1', 'a'],
 | |
|               [HtmlTokenType.TAG_OPEN_END],
 | |
|               [HtmlTokenType.EOF]
 | |
|             ]);
 | |
|       });
 | |
| 
 | |
|       it('should parse attributes whose prefix is not valid', () => {
 | |
|         expect(tokenizeAndHumanizeParts('<t (ns1:a)>'))
 | |
|             .toEqual([
 | |
|               [HtmlTokenType.TAG_OPEN_START, null, 't'],
 | |
|               [HtmlTokenType.ATTR_NAME, null, '(ns1:a)'],
 | |
|               [HtmlTokenType.TAG_OPEN_END],
 | |
|               [HtmlTokenType.EOF]
 | |
|             ]);
 | |
|       });
 | |
| 
 | |
|       it('should parse attributes with single quote value', () => {
 | |
|         expect(tokenizeAndHumanizeParts("<t a='b'>"))
 | |
|             .toEqual([
 | |
|               [HtmlTokenType.TAG_OPEN_START, null, 't'],
 | |
|               [HtmlTokenType.ATTR_NAME, null, 'a'],
 | |
|               [HtmlTokenType.ATTR_VALUE, 'b'],
 | |
|               [HtmlTokenType.TAG_OPEN_END],
 | |
|               [HtmlTokenType.EOF]
 | |
|             ]);
 | |
|       });
 | |
| 
 | |
|       it('should parse attributes with double quote value', () => {
 | |
|         expect(tokenizeAndHumanizeParts('<t a="b">'))
 | |
|             .toEqual([
 | |
|               [HtmlTokenType.TAG_OPEN_START, null, 't'],
 | |
|               [HtmlTokenType.ATTR_NAME, null, 'a'],
 | |
|               [HtmlTokenType.ATTR_VALUE, 'b'],
 | |
|               [HtmlTokenType.TAG_OPEN_END],
 | |
|               [HtmlTokenType.EOF]
 | |
|             ]);
 | |
|       });
 | |
| 
 | |
|       it('should parse attributes with unquoted value', () => {
 | |
|         expect(tokenizeAndHumanizeParts('<t a=b>'))
 | |
|             .toEqual([
 | |
|               [HtmlTokenType.TAG_OPEN_START, null, 't'],
 | |
|               [HtmlTokenType.ATTR_NAME, null, 'a'],
 | |
|               [HtmlTokenType.ATTR_VALUE, 'b'],
 | |
|               [HtmlTokenType.TAG_OPEN_END],
 | |
|               [HtmlTokenType.EOF]
 | |
|             ]);
 | |
|       });
 | |
| 
 | |
|       it('should allow whitespace', () => {
 | |
|         expect(tokenizeAndHumanizeParts('<t a = b >'))
 | |
|             .toEqual([
 | |
|               [HtmlTokenType.TAG_OPEN_START, null, 't'],
 | |
|               [HtmlTokenType.ATTR_NAME, null, 'a'],
 | |
|               [HtmlTokenType.ATTR_VALUE, 'b'],
 | |
|               [HtmlTokenType.TAG_OPEN_END],
 | |
|               [HtmlTokenType.EOF]
 | |
|             ]);
 | |
|       });
 | |
| 
 | |
|       it('should parse attributes with entities in values', () => {
 | |
|         expect(tokenizeAndHumanizeParts('<t a="AA">'))
 | |
|             .toEqual([
 | |
|               [HtmlTokenType.TAG_OPEN_START, null, 't'],
 | |
|               [HtmlTokenType.ATTR_NAME, null, 'a'],
 | |
|               [HtmlTokenType.ATTR_VALUE, 'AA'],
 | |
|               [HtmlTokenType.TAG_OPEN_END],
 | |
|               [HtmlTokenType.EOF]
 | |
|             ]);
 | |
|       });
 | |
| 
 | |
|       it('should not decode entities without trailing ";"', () => {
 | |
|         expect(tokenizeAndHumanizeParts('<t a="&" b="c&&d">'))
 | |
|             .toEqual([
 | |
|               [HtmlTokenType.TAG_OPEN_START, null, 't'],
 | |
|               [HtmlTokenType.ATTR_NAME, null, 'a'],
 | |
|               [HtmlTokenType.ATTR_VALUE, '&'],
 | |
|               [HtmlTokenType.ATTR_NAME, null, 'b'],
 | |
|               [HtmlTokenType.ATTR_VALUE, 'c&&d'],
 | |
|               [HtmlTokenType.TAG_OPEN_END],
 | |
|               [HtmlTokenType.EOF]
 | |
|             ]);
 | |
|       });
 | |
| 
 | |
|       it('should parse attributes with "&" in values', () => {
 | |
|         expect(tokenizeAndHumanizeParts('<t a="b && c &">'))
 | |
|             .toEqual([
 | |
|               [HtmlTokenType.TAG_OPEN_START, null, 't'],
 | |
|               [HtmlTokenType.ATTR_NAME, null, 'a'],
 | |
|               [HtmlTokenType.ATTR_VALUE, 'b && c &'],
 | |
|               [HtmlTokenType.TAG_OPEN_END],
 | |
|               [HtmlTokenType.EOF]
 | |
|             ]);
 | |
|       });
 | |
| 
 | |
|       it('should store the locations', () => {
 | |
|         expect(tokenizeAndHumanizeSourceSpans('<t a=b>'))
 | |
|             .toEqual([
 | |
|               [HtmlTokenType.TAG_OPEN_START, '<t'],
 | |
|               [HtmlTokenType.ATTR_NAME, 'a'],
 | |
|               [HtmlTokenType.ATTR_VALUE, 'b'],
 | |
|               [HtmlTokenType.TAG_OPEN_END, '>'],
 | |
|               [HtmlTokenType.EOF, '']
 | |
|             ]);
 | |
|       });
 | |
| 
 | |
|       it('should report missing value after =', () => {
 | |
|         expect(tokenizeAndHumanizeErrors('<name a='))
 | |
|             .toEqual([[HtmlTokenType.ATTR_VALUE, 'Unexpected character "EOF"', '0:8']]);
 | |
|       });
 | |
| 
 | |
|       it('should report missing end quote for \'', () => {
 | |
|         expect(tokenizeAndHumanizeErrors('<name a=\''))
 | |
|             .toEqual([[HtmlTokenType.ATTR_VALUE, 'Unexpected character "EOF"', '0:9']]);
 | |
|       });
 | |
| 
 | |
|       it('should report missing end quote for "', () => {
 | |
|         expect(tokenizeAndHumanizeErrors('<name a="'))
 | |
|             .toEqual([[HtmlTokenType.ATTR_VALUE, 'Unexpected character "EOF"', '0:9']]);
 | |
|       });
 | |
|     });
 | |
| 
 | |
|     describe('closing tags', () => {
 | |
|       it('should parse closing tags without prefix', () => {
 | |
|         expect(tokenizeAndHumanizeParts('</test>'))
 | |
|             .toEqual([[HtmlTokenType.TAG_CLOSE, null, 'test'], [HtmlTokenType.EOF]]);
 | |
|       });
 | |
| 
 | |
|       it('should parse closing tags with prefix', () => {
 | |
|         expect(tokenizeAndHumanizeParts('</ns1:test>'))
 | |
|             .toEqual([[HtmlTokenType.TAG_CLOSE, 'ns1', 'test'], [HtmlTokenType.EOF]]);
 | |
|       });
 | |
| 
 | |
|       it('should allow whitespace', () => {
 | |
|         expect(tokenizeAndHumanizeParts('</ test >'))
 | |
|             .toEqual([[HtmlTokenType.TAG_CLOSE, null, 'test'], [HtmlTokenType.EOF]]);
 | |
|       });
 | |
| 
 | |
|       it('should store the locations', () => {
 | |
|         expect(tokenizeAndHumanizeSourceSpans('</test>'))
 | |
|             .toEqual([[HtmlTokenType.TAG_CLOSE, '</test>'], [HtmlTokenType.EOF, '']]);
 | |
|       });
 | |
| 
 | |
|       it('should report missing name after </', () => {
 | |
|         expect(tokenizeAndHumanizeErrors('</'))
 | |
|             .toEqual([[HtmlTokenType.TAG_CLOSE, 'Unexpected character "EOF"', '0:2']]);
 | |
|       });
 | |
| 
 | |
|       it('should report missing >', () => {
 | |
|         expect(tokenizeAndHumanizeErrors('</test'))
 | |
|             .toEqual([[HtmlTokenType.TAG_CLOSE, 'Unexpected character "EOF"', '0:6']]);
 | |
|       });
 | |
|     });
 | |
| 
 | |
|     describe('entities', () => {
 | |
|       it('should parse named entities', () => {
 | |
|         expect(tokenizeAndHumanizeParts('a&b'))
 | |
|             .toEqual([[HtmlTokenType.TEXT, 'a&b'], [HtmlTokenType.EOF]]);
 | |
|       });
 | |
| 
 | |
|       it('should parse hexadecimal entities', () => {
 | |
|         expect(tokenizeAndHumanizeParts('A'))
 | |
|             .toEqual([[HtmlTokenType.TEXT, 'A'], [HtmlTokenType.EOF]]);
 | |
|       });
 | |
| 
 | |
|       it('should parse decimal entities', () => {
 | |
|         expect(tokenizeAndHumanizeParts('A'))
 | |
|             .toEqual([[HtmlTokenType.TEXT, 'A'], [HtmlTokenType.EOF]]);
 | |
|       });
 | |
| 
 | |
|       it('should store the locations', () => {
 | |
|         expect(tokenizeAndHumanizeSourceSpans('a&b'))
 | |
|             .toEqual([[HtmlTokenType.TEXT, 'a&b'], [HtmlTokenType.EOF, '']]);
 | |
|       });
 | |
| 
 | |
|       it('should report malformed/unknown entities', () => {
 | |
|         expect(tokenizeAndHumanizeErrors('&tbo;'))
 | |
|             .toEqual([
 | |
|               [
 | |
|                 HtmlTokenType.TEXT,
 | |
|                 'Unknown entity "tbo" - use the "&#<decimal>;" or  "&#x<hex>;" syntax',
 | |
|                 '0:0'
 | |
|               ]
 | |
|             ]);
 | |
|         expect(tokenizeAndHumanizeErrors('&#asdf;'))
 | |
|             .toEqual([[HtmlTokenType.TEXT, 'Unexpected character "s"', '0:3']]);
 | |
|         expect(tokenizeAndHumanizeErrors('
sdf;'))
 | |
|             .toEqual([[HtmlTokenType.TEXT, 'Unexpected character "s"', '0:4']]);
 | |
| 
 | |
|         expect(tokenizeAndHumanizeErrors('઼'))
 | |
|             .toEqual([[HtmlTokenType.TEXT, 'Unexpected character "EOF"', '0:6']]);
 | |
|       });
 | |
|     });
 | |
| 
 | |
|     describe('regular text', () => {
 | |
|       it('should parse text', () => {
 | |
|         expect(tokenizeAndHumanizeParts('a'))
 | |
|             .toEqual([[HtmlTokenType.TEXT, 'a'], [HtmlTokenType.EOF]]);
 | |
|       });
 | |
| 
 | |
|       it('should parse entities', () => {
 | |
|         expect(tokenizeAndHumanizeParts('a&b'))
 | |
|             .toEqual([[HtmlTokenType.TEXT, 'a&b'], [HtmlTokenType.EOF]]);
 | |
|       });
 | |
| 
 | |
|       it('should parse text starting with "&"', () => {
 | |
|         expect(tokenizeAndHumanizeParts('a && b &'))
 | |
|             .toEqual([[HtmlTokenType.TEXT, 'a && b &'], [HtmlTokenType.EOF]]);
 | |
|       });
 | |
| 
 | |
|       it('should store the locations', () => {
 | |
|         expect(tokenizeAndHumanizeSourceSpans('a'))
 | |
|             .toEqual([[HtmlTokenType.TEXT, 'a'], [HtmlTokenType.EOF, '']]);
 | |
|       });
 | |
|     });
 | |
| 
 | |
|     describe('raw text', () => {
 | |
|       it('should parse text', () => {
 | |
|         expect(tokenizeAndHumanizeParts(`<script>a</script>`))
 | |
|             .toEqual([
 | |
|               [HtmlTokenType.TAG_OPEN_START, null, 'script'],
 | |
|               [HtmlTokenType.TAG_OPEN_END],
 | |
|               [HtmlTokenType.RAW_TEXT, 'a'],
 | |
|               [HtmlTokenType.TAG_CLOSE, null, 'script'],
 | |
|               [HtmlTokenType.EOF]
 | |
|             ]);
 | |
|       });
 | |
| 
 | |
|       it('should not detect entities', () => {
 | |
|         expect(tokenizeAndHumanizeParts(`<script>&</script>`))
 | |
|             .toEqual([
 | |
|               [HtmlTokenType.TAG_OPEN_START, null, 'script'],
 | |
|               [HtmlTokenType.TAG_OPEN_END],
 | |
|               [HtmlTokenType.RAW_TEXT, '&'],
 | |
|               [HtmlTokenType.TAG_CLOSE, null, 'script'],
 | |
|               [HtmlTokenType.EOF]
 | |
|             ]);
 | |
|       });
 | |
| 
 | |
|       it('should ignore other opening tags', () => {
 | |
|         expect(tokenizeAndHumanizeParts(`<script>a<div></script>`))
 | |
|             .toEqual([
 | |
|               [HtmlTokenType.TAG_OPEN_START, null, 'script'],
 | |
|               [HtmlTokenType.TAG_OPEN_END],
 | |
|               [HtmlTokenType.RAW_TEXT, 'a<div>'],
 | |
|               [HtmlTokenType.TAG_CLOSE, null, 'script'],
 | |
|               [HtmlTokenType.EOF]
 | |
|             ]);
 | |
|       });
 | |
| 
 | |
|       it('should ignore other closing tags', () => {
 | |
|         expect(tokenizeAndHumanizeParts(`<script>a</test></script>`))
 | |
|             .toEqual([
 | |
|               [HtmlTokenType.TAG_OPEN_START, null, 'script'],
 | |
|               [HtmlTokenType.TAG_OPEN_END],
 | |
|               [HtmlTokenType.RAW_TEXT, 'a</test>'],
 | |
|               [HtmlTokenType.TAG_CLOSE, null, 'script'],
 | |
|               [HtmlTokenType.EOF]
 | |
|             ]);
 | |
|       });
 | |
| 
 | |
|       it('should store the locations', () => {
 | |
|         expect(tokenizeAndHumanizeSourceSpans(`<script>a</script>`))
 | |
|             .toEqual([
 | |
|               [HtmlTokenType.TAG_OPEN_START, '<script'],
 | |
|               [HtmlTokenType.TAG_OPEN_END, '>'],
 | |
|               [HtmlTokenType.RAW_TEXT, 'a'],
 | |
|               [HtmlTokenType.TAG_CLOSE, '</script>'],
 | |
|               [HtmlTokenType.EOF, '']
 | |
|             ]);
 | |
|       });
 | |
| 
 | |
|     });
 | |
| 
 | |
|     describe('escapable raw text', () => {
 | |
|       it('should parse text', () => {
 | |
|         expect(tokenizeAndHumanizeParts(`<title>a</title>`))
 | |
|             .toEqual([
 | |
|               [HtmlTokenType.TAG_OPEN_START, null, 'title'],
 | |
|               [HtmlTokenType.TAG_OPEN_END],
 | |
|               [HtmlTokenType.ESCAPABLE_RAW_TEXT, 'a'],
 | |
|               [HtmlTokenType.TAG_CLOSE, null, 'title'],
 | |
|               [HtmlTokenType.EOF]
 | |
|             ]);
 | |
|       });
 | |
| 
 | |
|       it('should detect entities', () => {
 | |
|         expect(tokenizeAndHumanizeParts(`<title>&</title>`))
 | |
|             .toEqual([
 | |
|               [HtmlTokenType.TAG_OPEN_START, null, 'title'],
 | |
|               [HtmlTokenType.TAG_OPEN_END],
 | |
|               [HtmlTokenType.ESCAPABLE_RAW_TEXT, '&'],
 | |
|               [HtmlTokenType.TAG_CLOSE, null, 'title'],
 | |
|               [HtmlTokenType.EOF]
 | |
|             ]);
 | |
|       });
 | |
| 
 | |
|       it('should ignore other opening tags', () => {
 | |
|         expect(tokenizeAndHumanizeParts(`<title>a<div></title>`))
 | |
|             .toEqual([
 | |
|               [HtmlTokenType.TAG_OPEN_START, null, 'title'],
 | |
|               [HtmlTokenType.TAG_OPEN_END],
 | |
|               [HtmlTokenType.ESCAPABLE_RAW_TEXT, 'a<div>'],
 | |
|               [HtmlTokenType.TAG_CLOSE, null, 'title'],
 | |
|               [HtmlTokenType.EOF]
 | |
|             ]);
 | |
|       });
 | |
| 
 | |
|       it('should ignore other closing tags', () => {
 | |
|         expect(tokenizeAndHumanizeParts(`<title>a</test></title>`))
 | |
|             .toEqual([
 | |
|               [HtmlTokenType.TAG_OPEN_START, null, 'title'],
 | |
|               [HtmlTokenType.TAG_OPEN_END],
 | |
|               [HtmlTokenType.ESCAPABLE_RAW_TEXT, 'a</test>'],
 | |
|               [HtmlTokenType.TAG_CLOSE, null, 'title'],
 | |
|               [HtmlTokenType.EOF]
 | |
|             ]);
 | |
|       });
 | |
| 
 | |
|       it('should store the locations', () => {
 | |
|         expect(tokenizeAndHumanizeSourceSpans(`<title>a</title>`))
 | |
|             .toEqual([
 | |
|               [HtmlTokenType.TAG_OPEN_START, '<title'],
 | |
|               [HtmlTokenType.TAG_OPEN_END, '>'],
 | |
|               [HtmlTokenType.ESCAPABLE_RAW_TEXT, 'a'],
 | |
|               [HtmlTokenType.TAG_CLOSE, '</title>'],
 | |
|               [HtmlTokenType.EOF, '']
 | |
|             ]);
 | |
|       });
 | |
| 
 | |
|     });
 | |
| 
 | |
|     describe('errors', () => {
 | |
|       it('should include 2 lines of context in message', () => {
 | |
|         let src = "111\n222\n333\nE\n444\n555\n666\n";
 | |
|         let file = new ParseSourceFile(src, 'file://');
 | |
|         let location = new ParseLocation(file, 12, 123, 456);
 | |
|         let error = new HtmlTokenError('**ERROR**', null, location);
 | |
|         expect(error.toString())
 | |
|             .toEqual(`**ERROR** ("\n222\n333\n[ERROR ->]E\n444\n555\n"): file://@123:456`);
 | |
|       });
 | |
|     });
 | |
| 
 | |
|   });
 | |
| }
 | |
| 
 | |
| function tokenizeWithoutErrors(input: string): HtmlToken[] {
 | |
|   var tokenizeResult = tokenizeHtml(input, 'someUrl');
 | |
|   if (tokenizeResult.errors.length > 0) {
 | |
|     var errorString = tokenizeResult.errors.join('\n');
 | |
|     throw new BaseException(`Unexpected parse errors:\n${errorString}`);
 | |
|   }
 | |
|   return tokenizeResult.tokens;
 | |
| }
 | |
| 
 | |
| function tokenizeAndHumanizeParts(input: string): any[] {
 | |
|   return tokenizeWithoutErrors(input).map(token => [<any>token.type].concat(token.parts));
 | |
| }
 | |
| 
 | |
| function tokenizeAndHumanizeSourceSpans(input: string): any[] {
 | |
|   return tokenizeWithoutErrors(input).map(token => [<any>token.type, token.sourceSpan.toString()]);
 | |
| }
 | |
| 
 | |
| function humanizeLineColumn(location: ParseLocation): string {
 | |
|   return `${location.line}:${location.col}`;
 | |
| }
 | |
| 
 | |
| function tokenizeAndHumanizeLineColumn(input: string): any[] {
 | |
|   return tokenizeWithoutErrors(input)
 | |
|       .map(token => [<any>token.type, humanizeLineColumn(token.sourceSpan.start)]);
 | |
| }
 | |
| 
 | |
| function tokenizeAndHumanizeErrors(input: string): any[] {
 | |
|   return tokenizeHtml(input, 'someUrl')
 | |
|       .errors.map(
 | |
|           tokenError =>
 | |
|               [<any>tokenError.tokenType, tokenError.msg, humanizeLineColumn(tokenError.location)]);
 | |
| }
 |