From 3ede943556bebb85427415f21ec231973da9e080 Mon Sep 17 00:00:00 2001 From: =?UTF-8?q?Gustav=20S=C3=B6rn=C3=A4s?= Date: Sat, 9 Jan 2021 21:25:39 +0100 Subject: wip line numbers and errors --- src/tokenizer.rs | 6 ++++++ 1 file changed, 6 insertions(+) (limited to 'src/tokenizer.rs') diff --git a/src/tokenizer.rs b/src/tokenizer.rs index a53015d..acebcca 100644 --- a/src/tokenizer.rs +++ b/src/tokenizer.rs @@ -119,6 +119,12 @@ pub type TokenStream = Vec; pub fn file_to_tokens(filename: &str) -> TokenStream { let content = fs::read_to_string(filename).unwrap(); let lexer = Token::lexer(&content); + let mut line = 1; + for (c_idx, c) in content.chars().enumerate() { + if c == '\n' { + line += 1; + } + } lexer.spanned().collect() } -- cgit v1.2.1