From 711968ffb31ee06630ed43740fd2538becc2765b Mon Sep 17 00:00:00 2001 From: =?UTF-8?q?Gustav=20S=C3=B6rn=C3=A4s?= Date: Mon, 11 Jan 2021 16:03:58 +0100 Subject: compile from strings --- src/tokenizer.rs | 7 +++++-- 1 file changed, 5 insertions(+), 2 deletions(-) (limited to 'src/tokenizer.rs') diff --git a/src/tokenizer.rs b/src/tokenizer.rs index 1da9835..e52410c 100644 --- a/src/tokenizer.rs +++ b/src/tokenizer.rs @@ -125,8 +125,7 @@ pub enum Token { pub type PlacedToken = (Token, usize); pub type TokenStream = Vec; -pub fn file_to_tokens(file: &Path) -> TokenStream { - let content = fs::read_to_string(file).unwrap(); +pub fn string_to_tokens(content: &str) -> TokenStream { let lexer = Token::lexer(&content); let mut placed_tokens = lexer.spanned().peekable(); @@ -152,6 +151,10 @@ pub fn file_to_tokens(file: &Path) -> TokenStream { lined_tokens } +pub fn file_to_tokens(file: &Path) -> TokenStream { + string_to_tokens(&fs::read_to_string(file).unwrap()) +} + #[cfg(test)] mod tests { use super::Token; -- cgit v1.2.1