rust/crates/libsyntax2/tests/test/main.rs

144 lines
3.6 KiB
Rust
Raw Normal View History

2018-08-11 09:03:03 +02:00
extern crate libsyntax2;
2018-01-07 12:56:08 +01:00
extern crate difference;
2018-07-30 14:25:52 +02:00
use std::{
fs,
2018-07-31 14:40:40 +02:00
path::{Path, PathBuf},
2018-08-11 09:03:03 +02:00
fmt::Write,
2018-07-30 14:25:52 +02:00
};
2018-01-07 12:56:08 +01:00
use difference::Changeset;
2018-08-11 09:03:03 +02:00
#[test]
fn lexer_tests() {
dir_tests(&["lexer"], |text| {
let tokens = libsyntax2::tokenize(text);
dump_tokens(&tokens, text)
})
}
#[test]
fn parser_tests() {
dir_tests(&["parser/inline", "parser/ok", "parser/err"], |text| {
let file = libsyntax2::parse(text);
libsyntax2::utils::dump_tree(&file)
})
}
/// Read file and normalize newlines.
///
/// `rustc` seems to always normalize `\r\n` newlines to `\n`:
///
/// ```
/// let s = "
/// ";
/// assert_eq!(s.as_bytes(), &[10]);
/// ```
///
/// so this should always be correct.
fn read_text(path: &Path) -> String {
2018-07-30 14:25:52 +02:00
fs::read_to_string(path).unwrap().replace("\r\n", "\n")
}
2018-02-03 10:51:06 +01:00
pub fn dir_tests<F>(paths: &[&str], f: F)
2018-08-11 09:03:03 +02:00
where
F: Fn(&str) -> String,
2018-01-07 13:34:11 +01:00
{
for path in collect_tests(paths) {
2018-02-11 09:19:54 +01:00
let input_code = read_text(&path);
let parse_tree = f(&input_code);
2018-01-07 13:34:11 +01:00
let path = path.with_extension("txt");
if !path.exists() {
println!("\nfile: {}", path.display());
2018-02-11 09:19:54 +01:00
println!("No .txt file with expected result, creating...\n");
println!("{}\n{}", input_code, parse_tree);
2018-07-30 14:25:52 +02:00
fs::write(&path, parse_tree).unwrap();
2018-01-07 13:34:11 +01:00
panic!("No expected result")
}
let expected = read_text(&path);
2018-01-07 13:34:11 +01:00
let expected = expected.as_str();
2018-02-11 09:19:54 +01:00
let parse_tree = parse_tree.as_str();
assert_equal_text(expected, parse_tree, &path);
2018-01-07 13:34:11 +01:00
}
}
2018-02-03 10:51:06 +01:00
fn assert_equal_text(expected: &str, actual: &str, path: &Path) {
2018-01-07 12:56:08 +01:00
if expected != actual {
print_difference(expected, actual, path)
}
}
2018-01-07 13:34:11 +01:00
fn collect_tests(paths: &[&str]) -> Vec<PathBuf> {
2018-02-03 10:51:06 +01:00
paths
.iter()
.flat_map(|path| {
let path = test_data_dir().join(path);
test_from_dir(&path).into_iter()
})
.collect()
2018-01-07 12:56:08 +01:00
}
fn test_from_dir(dir: &Path) -> Vec<PathBuf> {
let mut acc = Vec::new();
2018-07-30 14:25:52 +02:00
for file in fs::read_dir(&dir).unwrap() {
2018-01-07 12:56:08 +01:00
let file = file.unwrap();
let path = file.path();
if path.extension().unwrap_or_default() == "rs" {
acc.push(path);
}
}
acc.sort();
acc
}
2018-02-10 12:06:47 +01:00
const REWRITE: bool = false;
2018-01-07 12:56:08 +01:00
fn print_difference(expected: &str, actual: &str, path: &Path) {
let dir = project_dir();
let path = path.strip_prefix(&dir).unwrap_or_else(|_| path);
if expected.trim() == actual.trim() {
2018-02-10 10:35:40 +01:00
println!("whitespace difference, rewriting");
2018-02-10 12:06:47 +01:00
println!("file: {}\n", path.display());
2018-07-30 14:25:52 +02:00
fs::write(path, actual).unwrap();
2018-02-10 12:06:47 +01:00
return;
}
if REWRITE {
println!("rewriting {}", path.display());
2018-07-30 14:25:52 +02:00
fs::write(path, actual).unwrap();
2018-02-10 12:06:47 +01:00
return;
2018-01-07 12:56:08 +01:00
}
2018-02-10 12:06:47 +01:00
let changeset = Changeset::new(actual, expected, "\n");
2018-08-04 12:17:24 +02:00
println!("Expected:\n{}\n\nActual:\n{}\n", expected, actual);
2018-02-10 12:06:47 +01:00
print!("{}", changeset);
2018-02-10 10:35:40 +01:00
println!("file: {}\n", path.display());
2018-01-07 12:56:08 +01:00
panic!("Comparison failed")
}
fn project_dir() -> PathBuf {
let dir = env!("CARGO_MANIFEST_DIR");
PathBuf::from(dir)
2018-02-03 10:51:06 +01:00
.parent()
.unwrap()
.parent()
.unwrap()
2018-01-07 12:56:08 +01:00
.to_owned()
}
fn test_data_dir() -> PathBuf {
2018-08-11 09:03:03 +02:00
project_dir().join("crates/libsyntax2/tests/data")
}
fn dump_tokens(tokens: &[libsyntax2::Token], text: &str) -> String {
let mut acc = String::new();
let mut offset = 0;
for token in tokens {
let len: u32 = token.len.into();
let len = len as usize;
let token_text = &text[offset..offset + len];
offset += len;
write!(acc, "{:?} {} {:?}\n", token.kind, token.len, token_text).unwrap()
}
acc
2018-02-03 10:51:06 +01:00
}