From 78749296776b2d9b7cb1d03b93c97ceb32c8c0b3 Mon Sep 17 00:00:00 2001 From: =?UTF-8?q?Gustav=20S=C3=B6rn=C3=A4s?= Date: Sun, 10 Jan 2021 14:06:38 +0100 Subject: take Paths where files are needed --- src/tokenizer.rs | 5 +++-- 1 file changed, 3 insertions(+), 2 deletions(-) (limited to 'src/tokenizer.rs') diff --git a/src/tokenizer.rs b/src/tokenizer.rs index 9b33a06..e360085 100644 --- a/src/tokenizer.rs +++ b/src/tokenizer.rs @@ -1,4 +1,5 @@ use std::fs; +use std::path::Path; use logos::Logos; #[derive(Logos, Debug, PartialEq, Clone)] @@ -117,8 +118,8 @@ pub enum Token { pub type PlacedToken = (Token, usize); pub type TokenStream = Vec; -pub fn file_to_tokens(filename: &str) -> TokenStream { - let content = fs::read_to_string(filename).unwrap(); +pub fn file_to_tokens(file: &Path) -> TokenStream { + let content = fs::read_to_string(file).unwrap(); let lexer = Token::lexer(&content); let mut placed_tokens = lexer.spanned().peekable(); -- cgit v1.2.1