1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
|
use nom::{
IResult,
Parser,
branch::alt,
multi::many0,
bytes::complete::{tag, take_while1},
character::complete::multispace1,
};
#[derive(Debug,PartialEq)]
pub enum Token {
ParOpen,
ParClose,
Num(i32),
Sym(String),
Whitespace(String),
}
use Token::*;
use crate::parse::util::*;
fn parse_token(s: &str) -> IResult<&str, Token> {
alt((
tag("(").map(|_| ParOpen),
tag(")").map(|_| ParClose),
multispace1.map(whitespace),
nom::character::complete::i32.map(Num),
take_while1(|c| !(" \n\t()".contains(c))).map(sym),
)).parse(s)
}
fn tokenize(s: &str) -> Result<Vec<Token>, String> {
match many0(parse_token).parse(s) {
Ok(("", res)) => Ok(res),
Ok((rest, _)) => Err(format!("all data should be tokenizable, '{rest}' was not")),
Err(e) => Err(e.to_string()),
}
}
#[cfg(test)]
mod private_parsing_tests {
use super::{*, parse_token};
#[test]
fn test_parse_token() {
assert_eq!(parse_token("()"), Ok((")", ParOpen)));
assert_eq!(parse_token(")"), Ok(("", ParClose)));
assert_eq!(parse_token(" \t\n"), Ok(("", whitespace(" \t\n"))));
assert_eq!(parse_token("1 23"), Ok((" 23", Num(1))));
assert_eq!(parse_token("23"), Ok(("", Num(23))));
assert_eq!(parse_token("Nil a"), Ok((" a", sym("Nil"))));
assert_eq!(parse_token("a"), Ok(("", sym("a"))));
assert!(parse_token("").is_err())
}
#[test]
fn test_tokenize() {
assert_eq!(
tokenize("(+ 1 2 (\t\n a)").unwrap(),
vec![
ParOpen,
sym("+"), whitespace(" "),
Num(1), whitespace(" "),
Num(2), whitespace(" "),
ParOpen, whitespace("\t\n "),
sym("a"),
ParClose
]
);
}
}
|