summaryrefslogtreecommitdiff
path: root/src/tokenizer
diff options
context:
space:
mode:
Diffstat (limited to 'src/tokenizer')
-rw-r--r--src/tokenizer/mod.rs46
1 files changed, 23 insertions, 23 deletions
diff --git a/src/tokenizer/mod.rs b/src/tokenizer/mod.rs
index f5968b2..6a0649a 100644
--- a/src/tokenizer/mod.rs
+++ b/src/tokenizer/mod.rs
@@ -2164,7 +2164,7 @@ mod test {
struct LinesMatch {
current_str: String,
current_str_line: u64,
- lines: Vec<(Token, u64)>,
+ lines: Vec<(u64, Token)>,
}
impl LinesMatch {
@@ -2178,7 +2178,7 @@ mod test {
fn push(&mut self, token: Token, line_number: u64) {
self.finish_str();
- self.lines.push((token, line_number));
+ self.lines.push((line_number, token));
}
fn finish_str(&mut self) {
@@ -2240,7 +2240,7 @@ mod test {
// Take in tokens, process them, and return vector with line
// numbers that each token is on
- fn tokenize(input: Vec<String>, opts: TokenizerOpts) -> Vec<(Token, u64)> {
+ fn tokenize(input: Vec<String>, opts: TokenizerOpts) -> Vec<(u64, Token)> {
let sink = LinesMatch::new();
let mut tok = Tokenizer::new(sink, opts);
let mut buffer = BufferQueue::new();
@@ -2301,14 +2301,14 @@ mod test {
String::from("</a>\n"),
];
let expected = vec![
- (create_tag(String::from("a"), StartTag), 1),
- (CharacterTokens("\n".into()), 2),
- (create_tag(String::from("b"), StartTag), 2),
- (CharacterTokens("\n".into()), 3),
- (create_tag(String::from("b"), EndTag), 3),
- (CharacterTokens("\n".into()), 4),
- (create_tag(String::from("a"), EndTag), 4),
- (CharacterTokens("\n".into()), 5),
+ (1, create_tag(String::from("a"), StartTag)),
+ (2, CharacterTokens("\n".into())),
+ (2, create_tag(String::from("b"), StartTag)),
+ (3, CharacterTokens("\n".into())),
+ (3, create_tag(String::from("b"), EndTag)),
+ (4, CharacterTokens("\n".into())),
+ (4, create_tag(String::from("a"), EndTag)),
+ (5, CharacterTokens("\n".into())),
];
let results = tokenize(vector, opts);
assert_eq!(results, expected);
@@ -2330,14 +2330,14 @@ mod test {
String::from("</a>\r\n"),
];
let expected = vec![
- (create_tag(String::from("a"), StartTag), 1),
- (CharacterTokens("\n".into()), 2),
- (create_tag(String::from("b"), StartTag), 2),
- (CharacterTokens("\n".into()), 3),
- (create_tag(String::from("b"), EndTag), 3),
- (CharacterTokens("\n".into()), 4),
- (create_tag(String::from("a"), EndTag), 4),
- (CharacterTokens("\n".into()), 5),
+ (1, create_tag(String::from("a"), StartTag)),
+ (2, CharacterTokens("\n".into())),
+ (2, create_tag(String::from("b"), StartTag)),
+ (3, CharacterTokens("\n".into())),
+ (3, create_tag(String::from("b"), EndTag)),
+ (4, CharacterTokens("\n".into())),
+ (4, create_tag(String::from("a"), EndTag)),
+ (5, CharacterTokens("\n".into())),
];
let results = tokenize(vector, opts);
assert_eq!(results, expected);
@@ -2348,7 +2348,7 @@ mod test {
fn named_entities() {
let opts = TokenizerOpts::default();
let vector = vec![String::from("&amp;\r\n"), String::from("&aamp;\r\n")];
- let expected = vec![(Token::CharacterTokens("&amp;\n&aamp;\n".into()), 3)];
+ let expected = vec![(3, Token::CharacterTokens("&amp;\n&aamp;\n".into()))];
let results = tokenize(vector, opts);
assert_eq!(results, expected);
}
@@ -2359,9 +2359,9 @@ mod test {
let opts = TokenizerOpts::default();
let vector = vec![String::from("&amp;\r\n"), String::from("&aamp;\r\n")];
let expected = vec![
- (CharacterTokens("&\n".into()), 3),
- (ParseError("Invalid character reference".into()), 3),
- (CharacterTokens("&aamp;\n".into()), 4),
+ (3, CharacterTokens("&\n".into())),
+ (3, ParseError("Invalid character reference".into())),
+ (4, CharacterTokens("&aamp;\n".into())),
];
let results = tokenize(vector, opts);
assert_eq!(results, expected);