Skip to content
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
2 changes: 1 addition & 1 deletion crates/djls-templates/src/ast.rs
Original file line number Diff line number Diff line change
Expand Up @@ -153,7 +153,7 @@ impl Span {

#[must_use]
pub fn from_token(token: &Token<'_>, db: &dyn TemplateDb) -> Self {
let start = token.start().unwrap_or(0);
let start = token.offset().unwrap_or(0);
let length = token.length(db);
Span::new(start, length)
}
Expand Down
127 changes: 57 additions & 70 deletions crates/djls-templates/src/lexer.rs
Original file line number Diff line number Diff line change
Expand Up @@ -2,10 +2,16 @@ use crate::db::Db as TemplateDb;
use crate::tokens::Token;
use crate::tokens::TokenContent;

const BLOCK_TAG_START: &str = "{%";
const BLOCK_TAG_END: &str = "%}";
const VARIABLE_TAG_START: &str = "{{";
const VARIABLE_TAG_END: &str = "}}";
const COMMENT_TAG_START: &str = "{#";
const COMMENT_TAG_END: &str = "#}";

pub struct Lexer<'db> {
db: &'db dyn TemplateDb,
source: String,
chars: Vec<char>,
start: usize,
current: usize,
line: usize,
Expand All @@ -17,7 +23,6 @@ impl<'db> Lexer<'db> {
Lexer {
db,
source: String::from(source),
chars: source.chars().collect(),
start: 0,
current: 0,
line: 1,
Expand All @@ -32,22 +37,14 @@ impl<'db> Lexer<'db> {

let token = match self.peek() {
'{' => match self.peek_next() {
'%' => self.lex_django_construct("%}", |content, line, start| Token::Block {
content,
line,
start,
'%' => self.lex_django_construct(BLOCK_TAG_END, |content, offset| {
Token::Block { content, offset }
}),
'{' => {
self.lex_django_construct("}}", |content, line, start| Token::Variable {
content,
line,
start,
})
}
'#' => self.lex_django_construct("#}", |content, line, start| Token::Comment {
content,
line,
start,
'{' => self.lex_django_construct(VARIABLE_TAG_END, |content, offset| {
Token::Variable { content, offset }
}),
'#' => self.lex_django_construct(COMMENT_TAG_END, |content, offset| {
Token::Comment { content, offset }
}),
_ => self.lex_text(),
},
Expand All @@ -69,49 +66,43 @@ impl<'db> Lexer<'db> {
tokens.push(token);
}

tokens.push(Token::Eof { line: self.line });
tokens.push(Token::Eof);

tokens
}

fn lex_django_construct(
&mut self,
end: &str,
token_fn: impl FnOnce(TokenContent<'db>, usize, usize) -> Token<'db>,
token_fn: impl FnOnce(TokenContent<'db>, usize) -> Token<'db>,
) -> Token<'db> {
let line = self.line;
let start = self.start + 3;
let offset = self.start + 3;

self.consume_n(2);

match self.consume_until(end) {
Ok(text) => {
self.consume_n(2);
let content = TokenContent::new(self.db, text);
token_fn(content, line, start)
token_fn(content, offset)
}
Err(err_text) => {
self.synchronize();
let content = TokenContent::new(self.db, err_text);
Token::Error {
content,
line,
start,
}
Token::Error { content, offset }
}
}
}

fn lex_whitespace(&mut self, c: char) -> Token<'db> {
let line = self.line;
let start = self.start;
let offset = self.start;

if c == '\n' || c == '\r' {
self.consume(); // \r or \n
if c == '\r' && self.peek() == '\n' {
self.consume(); // \n of \r\n
}
Token::Newline { line, start }
Token::Newline { offset }
} else {
self.consume(); // Consume the first whitespace
while !self.is_at_end() && self.peek().is_whitespace() {
Expand All @@ -121,67 +112,64 @@ impl<'db> Lexer<'db> {
self.consume();
}
let count = self.current - self.start;
Token::Whitespace { count, line, start }
Token::Whitespace { count, offset }
}
}

fn lex_text(&mut self) -> Token<'db> {
let line = self.line;
let start = self.start;
let text_start = self.current;

let mut text = String::new();
while !self.is_at_end() {
let c = self.peek();

if c == '{' {
let next = self.peek_next();
if next == '%' || next == '{' || next == '#' {
break;
}
} else if c == '\n' {
if self.source[self.current..].starts_with(BLOCK_TAG_START)
|| self.source[self.current..].starts_with(VARIABLE_TAG_START)
|| self.source[self.current..].starts_with(COMMENT_TAG_START)
|| self.source[self.current..].starts_with('\n')
{
break;
}

text.push(c);
self.consume();
}

let content = TokenContent::new(self.db, text);
let text = &self.source[text_start..self.current];
let content = TokenContent::new(self.db, text.to_string());
Token::Text {
content,
line,
start,
offset: self.start,
}
}

#[inline]
fn peek(&self) -> char {
self.peek_at(0)
self.source[self.current..].chars().next().unwrap_or('\0')
}

fn peek_next(&self) -> char {
self.peek_at(1)
let mut chars = self.source[self.current..].chars();
chars.next(); // Skip current
chars.next().unwrap_or('\0')
}

fn peek_previous(&self) -> char {
self.peek_at(-1)
}

fn peek_at(&self, offset: isize) -> char {
let Some(index) = self.current.checked_add_signed(offset) else {
if self.current == 0 {
return '\0';
};
self.chars.get(index).copied().unwrap_or('\0')
}
let mut pos = self.current - 1;
while !self.source.is_char_boundary(pos) && pos > 0 {
pos -= 1;
}
self.source[pos..].chars().next().unwrap_or('\0')
}

#[inline]
fn is_at_end(&self) -> bool {
self.current >= self.source.len()
}

#[inline]
fn consume(&mut self) {
if self.is_at_end() {
return;
if let Some(ch) = self.source[self.current..].chars().next() {
self.current += ch.len_utf8();
}
self.current += 1;
}

fn consume_n(&mut self, count: usize) {
Expand All @@ -190,25 +178,24 @@ impl<'db> Lexer<'db> {
}
}

fn consume_until(&mut self, s: &str) -> Result<String, String> {
let start = self.current;
while !self.is_at_end() {
if self.chars[self.current..self.chars.len()]
.starts_with(s.chars().collect::<Vec<_>>().as_slice())
{
return Ok(self.source[start..self.current].trim().to_string());
fn consume_until(&mut self, delimiter: &str) -> Result<String, String> {
let offset = self.current;

while self.current < self.source.len() {
if self.source[self.current..].starts_with(delimiter) {
return Ok(self.source[offset..self.current].trim().to_string());
}
self.consume();
}
Err(self.source[start..self.current].trim().to_string())

Err(self.source[offset..self.current].trim().to_string())
}

fn synchronize(&mut self) {
let sync_chars = &['{', '\n', '\r'];
const SYNC_POINTS: &[u8] = b"{\n\r";

while !self.is_at_end() {
let current_char = self.peek();
if sync_chars.contains(&current_char) {
if SYNC_POINTS.contains(&self.source.as_bytes()[self.current]) {
return;
}
self.consume();
Expand Down
25 changes: 11 additions & 14 deletions crates/djls-templates/src/parser.rs
Original file line number Diff line number Diff line change
Expand Up @@ -37,9 +37,9 @@ impl<'db> Parser<'db> {
let tokens = self.tokens.stream(self.db);
for token in tokens {
if matches!(token, Token::Newline { .. }) {
let start = token.start();
if let Some(start) = start {
line_offsets.add_line(start + 1);
let offset = token.offset();
if let Some(offset) = offset {
line_offsets.add_line(offset + 1);
}
}
}
Expand Down Expand Up @@ -91,16 +91,13 @@ impl<'db> Parser<'db> {
let token = self.peek_previous()?;

if let Token::Error {
content,
line: _,
start,
..
content, offset, ..
} = token
{
let error_text = content.text(self.db).clone();

Err(ParserError::MalformedConstruct {
position: start,
position: offset,
content: error_text,
})
} else {
Expand Down Expand Up @@ -152,8 +149,8 @@ impl<'db> Parser<'db> {
return self.next_node();
}

let start = first_token.start().unwrap_or(0);
let mut end_position = start + first_token.length(self.db);
let offset = first_token.offset().unwrap_or(0);
let mut end_position = offset + first_token.length(self.db);

while let Ok(token) = self.peek() {
match token {
Expand All @@ -164,16 +161,16 @@ impl<'db> Parser<'db> {
| Token::Eof { .. } => break, // Stop at Django constructs
Token::Text { .. } | Token::Whitespace { .. } | Token::Newline { .. } => {
// Update end position
let token_start = token.start().unwrap_or(end_position);
let token_offset = token.offset().unwrap_or(end_position);
let token_length = token.length(self.db);
end_position = token_start + token_length;
end_position = token_offset + token_length;
self.consume()?;
}
}
}

let length = end_position - start;
let span = Span::new(start, length);
let length = end_position - offset;
let span = Span::new(offset, length);

Ok(Node::Text { span })
}
Expand Down
Loading