Change all script strings to Rc<str>
parent
3c348e92ca
commit
1b826430f0
5
build.rs
5
build.rs
|
@ -22,7 +22,8 @@ fn do_ast_codegen<W: Write>(mut output: W) {
|
|||
)
|
||||
.unwrap();
|
||||
|
||||
define_imports(&mut output, &["crate::lex::Token"]).expect("failed to define imports");
|
||||
define_imports(&mut output, &["crate::lex::Token", "std::rc::Rc"])
|
||||
.expect("failed to define imports");
|
||||
writeln!(output).unwrap();
|
||||
|
||||
define_literals(
|
||||
|
@ -30,7 +31,7 @@ fn do_ast_codegen<W: Write>(mut output: W) {
|
|||
"LiteralValue",
|
||||
&BTreeMap::from([
|
||||
("Number", Some("f64")),
|
||||
("String", Some("String")),
|
||||
("String", Some("Rc<str>")),
|
||||
("True", None),
|
||||
("False", None),
|
||||
("Nil", None),
|
||||
|
|
|
@ -25,6 +25,7 @@ impl Environment {
|
|||
const ROOT_KEY: u32 = 0;
|
||||
|
||||
#[must_use]
|
||||
#[allow(dead_code)]
|
||||
pub fn new() -> Self {
|
||||
Self::default()
|
||||
}
|
||||
|
|
26
src/lex.rs
26
src/lex.rs
|
@ -1,5 +1,8 @@
|
|||
use itertools::{FoldWhile, Itertools};
|
||||
use std::iter::{self, Peekable};
|
||||
use std::{
|
||||
iter::{self, Peekable},
|
||||
rc::Rc,
|
||||
};
|
||||
use thiserror::Error;
|
||||
|
||||
use crate::ScriptError;
|
||||
|
@ -28,8 +31,8 @@ pub enum TokenKind {
|
|||
LessEqual,
|
||||
|
||||
// Literals
|
||||
Identifier(String),
|
||||
String(String),
|
||||
Identifier(Rc<str>),
|
||||
String(Rc<str>),
|
||||
Number(f64),
|
||||
|
||||
// Keywords
|
||||
|
@ -92,10 +95,6 @@ impl Token {
|
|||
&self.kind
|
||||
}
|
||||
|
||||
pub fn into_kind(self) -> TokenKind {
|
||||
self.kind
|
||||
}
|
||||
|
||||
pub fn lexeme(&self) -> &str {
|
||||
&self.lexeme
|
||||
}
|
||||
|
@ -326,16 +325,19 @@ fn lex_remaining_string_literal<I: Iterator<Item = char>>(
|
|||
.into_inner();
|
||||
|
||||
if end_found {
|
||||
let lexeme = format!("\"{contents}\"");
|
||||
// must include quotes
|
||||
let length = contents.len() + 2;
|
||||
|
||||
Ok(Consumed {
|
||||
token: Some(Token {
|
||||
kind: TokenKind::String(contents.clone()),
|
||||
lexeme: format!("\"{contents}\""),
|
||||
kind: TokenKind::String(contents.into()),
|
||||
lexeme,
|
||||
line: start_line + newlines,
|
||||
}),
|
||||
span: Span {
|
||||
lines: newlines,
|
||||
// must include the quotes
|
||||
chars: contents.len() + 2,
|
||||
chars: length,
|
||||
},
|
||||
})
|
||||
} else {
|
||||
|
@ -401,6 +403,6 @@ fn tokenize_word(word: String) -> TokenKind {
|
|||
"true" => TokenKind::True,
|
||||
"var" => TokenKind::Var,
|
||||
"while" => TokenKind::While,
|
||||
_other => TokenKind::Identifier(word),
|
||||
_other => TokenKind::Identifier(word.into()),
|
||||
}
|
||||
}
|
||||
|
|
12
src/parse.rs
12
src/parse.rs
|
@ -81,7 +81,6 @@ fn parse<I: Iterator<Item = Token>, F: FnMut(ScriptError)>(
|
|||
line: error.line.unwrap_or_default(),
|
||||
location: String::new(),
|
||||
});
|
||||
todo!("{}", error.message);
|
||||
}
|
||||
}
|
||||
}
|
||||
|
@ -329,17 +328,10 @@ fn parse_primary<I: Iterator<Item = Token>>(iter: &mut Peekable<I>) -> Result<Ex
|
|||
value: LiteralValue::Number(*number),
|
||||
token,
|
||||
}),
|
||||
TokenKind::String(_) => {
|
||||
// TODO: we used to do this with into_kind, but we need to change Token to Rc<str> to fix cloning issues
|
||||
// special case to avoid cloning
|
||||
match token.kind() {
|
||||
TokenKind::String(string) => Ok(Expr::Literal {
|
||||
value: LiteralValue::String(string.clone()),
|
||||
TokenKind::String(value) => Ok(Expr::Literal {
|
||||
value: LiteralValue::String(value.clone()),
|
||||
token,
|
||||
}),
|
||||
_ => unreachable!(),
|
||||
}
|
||||
}
|
||||
TokenKind::LeftParen => {
|
||||
let expr = parse_expression(iter)?;
|
||||
match_token_kind!(iter, TokenKind::RightParen)
|
||||
|
|
Loading…
Reference in New Issue