A lot of stuff I'm not gonna enumerate

rule fundamentals
This commit is contained in:
2022-07-06 20:13:13 +02:00
parent 49aa73956c
commit ec7ad81fac
21 changed files with 483 additions and 420 deletions

View File

@@ -1,4 +1,4 @@
import std::io::(println, out) -- imports -- import std::io::(println, out) -- imports
-- single word rule (alias) -- single word rule (alias)
greet =1=> (\name. printf out "Hello {}!\n" [name]) greet =1=> (\name. printf out "Hello {}!\n" [name])
@@ -7,12 +7,15 @@ greet =1=> (\name. printf out "Hello {}!\n" [name])
export ;> $a =200=> (greet $a) export ;> $a =200=> (greet $a)
-- single-word exported rule -- single-word exported rule
export main = ( export main == (
print "What is your name?" >> print "What is your name?" >>
readln >>= \name. readln >>= \name.
greet name greet name
) )
export < $a ...$rest /> == (createElement (tok_to_str $a) [(props_carriage ...$rest)])
export (props_carriage $key = $value) == (tok_to_str $key) => $value
-- The broadest trait definition in existence -- The broadest trait definition in existence
Foo = (Bar Baz) Foo == (Bar Baz)
-- default anyFoo = @T. @impl:(T (Bar Baz)). impl:(T Foo) -- default anyFoo = @T. @impl:(T (Bar Baz)). impl:(T Foo)

View File

@@ -1,7 +1,9 @@
use itertools::Itertools;
use ordered_float::NotNan; use ordered_float::NotNan;
use std::{fmt::Debug}; use std::{fmt::Debug};
#[derive(Debug, Clone, PartialEq, Eq, Hash)] /// An exact value
#[derive(Clone, PartialEq, Eq, Hash)]
pub enum Literal { pub enum Literal {
Num(NotNan<f64>), Num(NotNan<f64>),
Int(u64), Int(u64),
@@ -9,10 +11,31 @@ pub enum Literal {
Str(String), Str(String),
} }
impl Debug for Literal {
fn fmt(&self, f: &mut std::fmt::Formatter<'_>) -> std::fmt::Result {
match self {
Self::Num(arg0) => write!(f, "{:?}", arg0),
Self::Int(arg0) => write!(f, "{:?}", arg0),
Self::Char(arg0) => write!(f, "{:?}", arg0),
Self::Str(arg0) => write!(f, "{:?}", arg0),
}
}
}
/// An S-expression with a type /// An S-expression with a type
#[derive(Debug, Clone, PartialEq, Eq, Hash)] #[derive(Clone, PartialEq, Eq, Hash)]
pub struct Expr(pub Clause, pub Option<Box<Expr>>); pub struct Expr(pub Clause, pub Option<Box<Expr>>);
impl Debug for Expr {
fn fmt(&self, f: &mut std::fmt::Formatter<'_>) -> std::fmt::Result {
// f.debug_tuple("Expr").field(&self.0).field(&self.1).finish()
let Expr(val, typ) = self;
write!(f, "{:?}", val)?;
if let Some(typ) = typ { write!(f, "{:?}", typ) }
else { Ok(()) }
}
}
impl Expr { impl Expr {
/// Replace all occurences of a name in the tree with a parameter, to bypass name resolution /// Replace all occurences of a name in the tree with a parameter, to bypass name resolution
pub fn bind_parameter(&mut self, name: &str) { pub fn bind_parameter(&mut self, name: &str) {
@@ -24,7 +47,7 @@ impl Expr {
} }
/// An S-expression as read from a source file /// An S-expression as read from a source file
#[derive(Debug, Clone, PartialEq, Eq, Hash)] #[derive(Clone, PartialEq, Eq, Hash)]
pub enum Clause { pub enum Clause {
Literal(Literal), Literal(Literal),
Name(Vec<String>), Name(Vec<String>),
@@ -34,6 +57,44 @@ pub enum Clause {
Parameter(String) Parameter(String)
} }
fn fmt_expr_seq(it: &mut dyn Iterator<Item = &Expr>, f: &mut std::fmt::Formatter<'_>) -> std::fmt::Result {
for item in Itertools::intersperse(it.map(Some), None) { match item {
Some(expr) => write!(f, "{:?}", expr),
None => f.write_str(" "),
}? }
Ok(())
}
impl Debug for Clause {
fn fmt(&self, f: &mut std::fmt::Formatter<'_>) -> std::fmt::Result {
match self {
Self::Literal(arg0) => write!(f, "{:?}", arg0),
Self::Name(arg0) => write!(f, "{}", arg0.join("::")),
Self::S(del, items) => {
f.write_str(&del.to_string())?;
fmt_expr_seq(&mut items.iter(), f)?;
f.write_str(match del {
'(' => ")", '[' => "]", '{' => "}",
_ => "CLOSING_DELIM"
})
},
Self::Lambda(name, argtyp, body) => {
f.write_str("\\")?;
f.write_str(name)?;
f.write_str(":")?; fmt_expr_seq(&mut argtyp.iter(), f)?; f.write_str(".")?;
fmt_expr_seq(&mut body.iter(), f)
},
Self::Auto(name, argtyp, body) => {
f.write_str("@")?;
f.write_str(&name.clone().unwrap_or_default())?;
f.write_str(":")?; fmt_expr_seq(&mut argtyp.iter(), f)?; f.write_str(".")?;
fmt_expr_seq(&mut body.iter(), f)
},
Self::Parameter(name) => write!(f, "`{}", name)
}
}
}
impl Clause { impl Clause {
/// Replace all occurences of a name in the tree with a parameter, to bypass name resolution /// Replace all occurences of a name in the tree with a parameter, to bypass name resolution
pub fn bind_parameter(&mut self, name: &str) { pub fn bind_parameter(&mut self, name: &str) {
@@ -50,3 +111,17 @@ impl Clause {
} }
} }
} }
/// A substitution rule as read from the source
#[derive(Clone, PartialEq, Eq, Hash)]
pub struct Rule {
pub source: Vec<Expr>,
pub prio: NotNan<f64>,
pub target: Vec<Expr>
}
impl Debug for Rule {
fn fmt(&self, f: &mut std::fmt::Formatter<'_>) -> std::fmt::Result {
write!(f, "{:?} ={}=> {:?}", self.source, self.prio, self.target)
}
}

View File

@@ -1,17 +1,12 @@
use core::ops::Range;
use std::{env::current_dir, fs::read_to_string};
use std::io;
use chumsky::prelude::end; use std::env::current_dir;
use chumsky::{Parser, Stream};
use parse::{LexerEntry, FileEntry};
// use project::{rule_collector, file_loader, Loaded};
mod parse; mod parse;
mod project; mod project;
mod utils; mod utils;
mod expression; mod expression;
pub use expression::*; mod rule;
use project::{rule_collector, Loaded, file_loader};
fn literal(orig: &[&str]) -> Vec<String> { fn literal(orig: &[&str]) -> Vec<String> {
orig.iter().map(|&s| s.to_owned()).collect() orig.iter().map(|&s| s.to_owned()).collect()
@@ -27,37 +22,15 @@ export (match_sequence $lhs) >>= (match_sequence $rhs) =100=> (bind ($lhs) ($rhs
"#; "#;
fn main() { fn main() {
// let mut input = String::new(); let cwd = current_dir().unwrap();
// let mut stdin = io::stdin(); let collect_rules = rule_collector(move |n| {
// stdin.read_to_string(&mut input).unwrap(); if n == vec!["prelude"] { Ok(Loaded::Module(PRELUDE.to_string())) }
let ops: Vec<&str> = vec!["...", ">>", ">>=", "[", "]", ",", "$"]; else { file_loader(cwd.clone())(n) }
let data = read_to_string("./main.orc").unwrap(); }, literal(&["...", ">>", ">>=", "[", "]", ",", "$", "=", "=>"]));
let lexed = parse::lexer(&ops).parse(data).unwrap(); match collect_rules.try_find(&literal(&["main"])) {
println!("Lexed: {:?}", lexed); Ok(rules) => for rule in rules.iter() {
let parsr = parse::line_parser().then_ignore(end()); println!("{rule:?}")
// match parsr.parse(data) { }
// Ok(output) => println!("\nParsed:\n{:?}", output), Err(err) => println!("{:#?}", err)
// Err(e) => println!("\nErrored:\n{:?}", e)
// }
let lines = lexed.iter().filter_map(|v| {
let parse::LexerEntry(_, Range{ end, .. }) = v.last().unwrap().clone();
let tuples = v.into_iter().map(|LexerEntry(l, r)| (l.clone(), r.clone()));
Some(parsr.parse_recovery_verbose(Stream::from_iter(end..end+1, tuples)))
}).collect::<Vec<_>>();
for (id, (out, errs)) in lines.into_iter().enumerate() {
println!("Parsing line {}", id);
if let Some(output) = out { println!("Parsed:\n{:?}", output) }
else { println!("Failed to produce output")}
if errs.len() > 0 { println!("Errored:\n{:?}", errs)}
} }
// let output = parse::file_parser(&ops, &ops).parse(data).unwrap();
// let cwd = current_dir().unwrap();
// let collect_rules = rule_collector(move |n| {
// if n == vec!["prelude"] { Ok(Loaded::Module(PRELUDE.to_string())) }
// else { file_loader(cwd.clone())(n) }
// }, literal(&["...", ">>", ">>=", "[", "]", ","]));
// let rules = collect_rules.try_find(&literal(&["main"])).unwrap();
// for rule in rules.iter() {
// println!("{:?} ={}=> {:?}", rule.source, rule.priority, rule.target)
// }
} }

View File

@@ -1,5 +1,5 @@
use chumsky::{self, prelude::*, Parser}; use chumsky::{self, prelude::*, Parser};
use crate::{Clause, Expr, Literal, enum_parser}; use crate::{enum_parser, expression::{Clause, Expr, Literal}};
use super::{lexer::Lexeme}; use super::{lexer::Lexeme};
@@ -86,5 +86,5 @@ pub fn xpr_parser() -> impl Parser<Lexeme, Expr, Error = Simple<Lexeme>> {
.ignore_then(expr.clone()).or_not() .ignore_then(expr.clone()).or_not()
) )
.map(|(val, typ)| Expr(val, typ.map(Box::new))) .map(|(val, typ)| Expr(val, typ.map(Box::new)))
}) }).labelled("Expression")
} }

View File

@@ -1,6 +1,6 @@
use std::{ops::Range, iter}; use std::{ops::Range, iter, fmt};
use ordered_float::NotNan; use ordered_float::NotNan;
use chumsky::{Parser, prelude::*, text::whitespace}; use chumsky::{Parser, prelude::*};
use std::fmt::Debug; use std::fmt::Debug;
use crate::utils::BoxedIter; use crate::utils::BoxedIter;
@@ -14,6 +14,11 @@ impl Debug for Entry {
// f.debug_tuple("Entry").field(&self.0).field(&self.1).finish() // f.debug_tuple("Entry").field(&self.0).field(&self.1).finish()
} }
} }
impl Into<(Lexeme, Range<usize>)> for Entry {
fn into(self) -> (Lexeme, Range<usize>) {
(self.0, self.1)
}
}
#[derive(Clone, PartialEq, Eq, Hash)] #[derive(Clone, PartialEq, Eq, Hash)]
pub enum Lexeme { pub enum Lexeme {
@@ -61,6 +66,9 @@ impl Lexeme {
pub fn name<T: ToString>(n: T) -> Self { pub fn name<T: ToString>(n: T) -> Self {
Lexeme::Name(n.to_string()) Lexeme::Name(n.to_string())
} }
pub fn rule<T>(prio: T) -> Self where T: Into<f64> {
Lexeme::Rule(NotNan::new(prio.into()).expect("Rule priority cannot be NaN"))
}
pub fn paren_parser<T, P>( pub fn paren_parser<T, P>(
expr: P expr: P
) -> impl Parser<Lexeme, (char, T), Error = Simple<Lexeme>> + Clone ) -> impl Parser<Lexeme, (char, T), Error = Simple<Lexeme>> + Clone
@@ -76,15 +84,20 @@ impl Lexeme {
} }
} }
fn rule_parser() -> impl Parser<char, NotNan<f64>, Error = Simple<char>> { #[derive(Clone, PartialEq, Eq, Hash)]
just('=').ignore_then( pub struct LexedText(pub Vec<Vec<Entry>>);
choice((
none_of("-0123456789").rewind().to(NotNan::new(0f64).unwrap()), impl Debug for LexedText {
number::float_parser().then_ignore(just("=>")) fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result {
)).map_err_with_span(|err, span| { for row in &self.0 {
panic!("Something's up! {:?} {}", span, err) for tok in row {
}) tok.fmt(f)?;
) f.write_str(" ")?
}
f.write_str("\n")?
}
Ok(())
}
} }
type LexSubres<'a> = BoxedIter<'a, Entry>; type LexSubres<'a> = BoxedIter<'a, Entry>;
@@ -104,7 +117,7 @@ fn paren_parser<'a>(
}) })
} }
pub fn lexer<'a, T: 'a>(ops: &[T]) -> impl Parser<char, Vec<Vec<Entry>>, Error=Simple<char>> + 'a pub fn lexer<'a, T: 'a>(ops: &[T]) -> impl Parser<char, LexedText, Error=Simple<char>> + 'a
where T: AsRef<str> + Clone { where T: AsRef<str> + Clone {
let all_ops = ops.iter().map(|o| o.as_ref().to_string()) let all_ops = ops.iter().map(|o| o.as_ref().to_string())
.chain(iter::once(".".to_string())).collect::<Vec<_>>(); .chain(iter::once(".".to_string())).collect::<Vec<_>>();
@@ -114,7 +127,8 @@ where T: AsRef<str> + Clone {
paren_parser(recurse.clone(), '[', ']'), paren_parser(recurse.clone(), '[', ']'),
paren_parser(recurse.clone(), '{', '}'), paren_parser(recurse.clone(), '{', '}'),
choice(( choice((
rule_parser().map(Lexeme::Rule), just("==").padded().to(Lexeme::rule(0f64)),
just("=").ignore_then(number::float_parser()).then_ignore(just("=>")).map(Lexeme::rule),
comment::comment_parser().map(Lexeme::Comment), comment::comment_parser().map(Lexeme::Comment),
just("::").padded().to(Lexeme::NS), just("::").padded().to(Lexeme::NS),
just('\\').padded().to(Lexeme::BS), just('\\').padded().to(Lexeme::BS),
@@ -130,5 +144,5 @@ where T: AsRef<str> + Clone {
}).separated_by(one_of("\t ").repeated()) }).separated_by(one_of("\t ").repeated())
.flatten().collect() .flatten().collect()
.separated_by(just('\n').then(text::whitespace()).ignored()) .separated_by(just('\n').then(text::whitespace()).ignored())
.map(LexedText)
} }

View File

@@ -7,6 +7,7 @@ mod expression;
mod sourcefile; mod sourcefile;
mod import; mod import;
mod enum_parser; mod enum_parser;
mod parse;
pub use sourcefile::FileEntry; pub use sourcefile::FileEntry;
pub use sourcefile::line_parser; pub use sourcefile::line_parser;
@@ -14,3 +15,4 @@ pub use sourcefile::imports;
pub use sourcefile::exported_names; pub use sourcefile::exported_names;
pub use lexer::{lexer, Lexeme, Entry as LexerEntry}; pub use lexer::{lexer, Lexeme, Entry as LexerEntry};
pub use name::is_op; pub use name::is_op;
pub use parse::{parse, reparse, ParseError};

65
src/parse/parse.rs Normal file
View File

@@ -0,0 +1,65 @@
use std::{ops::Range, fmt::Debug};
use chumsky::{prelude::{Simple, end}, Stream, Parser};
use itertools::Itertools;
use thiserror::Error;
use crate::expression::Rule;
use super::{Lexeme, FileEntry, lexer, line_parser, LexerEntry};
#[derive(Error, Debug, Clone)]
pub enum ParseError {
#[error("Could not tokenize {0:?}")]
Lex(Vec<Simple<char>>),
#[error("Could not parse {0:#?}")]
Ast(Vec<Simple<Lexeme>>)
}
pub fn parse<'a, Iter, S, Op>(ops: &[Op], stream: S) -> Result<Vec<FileEntry>, ParseError>
where
Op: 'a + AsRef<str> + Clone,
Iter: Iterator<Item = (char, Range<usize>)> + 'a,
S: Into<Stream<'a, char, Range<usize>, Iter>> {
let lexed = lexer(ops).parse(stream).map_err(ParseError::Lex)?;
println!("Lexed:\n{:?}", lexed);
let parsr = line_parser().then_ignore(end());
let (parsed_lines, errors_per_line) = lexed.0.into_iter().filter_map(|v| {
// Find the first invalid position for Stream::for_iter
let LexerEntry(_, Range{ end, .. }) = v.last().unwrap().clone();
// Stream expects tuples, lexer outputs structs
let tuples = v.into_iter().map_into::<(Lexeme, Range<usize>)>();
Some(parsr.parse(Stream::from_iter(end..end+1, tuples)))
// ^^^^^^^^^^
// I haven't the foggiest idea why this is needed, parsers are supposed to be lazy so the
// end of input should make little difference
}).map(|res| match res {
Ok(r) => (Some(r), vec![]),
Err(e) => (None, e)
}).unzip::<_, _, Vec<_>, Vec<_>>();
let total_err = errors_per_line.into_iter()
.map(Vec::into_iter).flatten()
.collect::<Vec<_>>();
if total_err.len() > 0 { Err(ParseError::Ast(total_err)) }
else { Ok(parsed_lines.into_iter().map(Option::unwrap).collect()) }
}
pub fn reparse<'a, Iter, S, Op>(ops: &[Op], stream: S, pre: &Vec<FileEntry>)
-> Result<Vec<FileEntry>, ParseError>
where
Op: 'a + AsRef<str> + Clone,
Iter: Iterator<Item = (char, Range<usize>)> + 'a,
S: Into<Stream<'a, char, Range<usize>, Iter>> {
let result = parse(ops, stream)?;
Ok(result.into_iter().zip(pre.iter()).map(|(mut output, donor)| {
if let FileEntry::Rule(Rule{source, ..}, _) = &mut output {
if let FileEntry::Rule(Rule{source: s2, ..}, _) = donor {
*source = s2.clone()
} else {
panic!("Preparse and reparse received different row types!")
}
}
output
}).collect())
}

View File

@@ -1,34 +0,0 @@
use chumsky::{self, prelude::*, Parser};
use super::{expression, number::float_parser};
#[derive(Debug, Clone)]
pub struct Rule {
pub source: expression::Expr,
pub priority: f64,
pub target: expression::Expr
}
/// Parses a substitution rule of the forms
///
/// ```orchid
/// main = \x. ...
/// $a + $b = (add $a $b)
/// (foo bar baz) =1.1=> (foo 1 e)
/// reee =2=> shadow_reee
/// shadow_reee =0.9=> reee
/// ```
/// TBD whether this disables reee in the specified range or loops forever
pub fn rule_parser<'a, T: 'a + AsRef<str> + Clone>(
pattern_ops: &[T],
ops: &[T]
) -> impl Parser<char, Rule, Error = Simple<char>> + 'a {
expression::expression_parser(pattern_ops).padded()
.then_ignore(just('='))
.then(
float_parser().then_ignore(just("=>"))
.or_not().map(|prio| prio.unwrap_or(0.0))
).then(expression::expression_parser(ops).padded())
.map(|((source, priority), target)| Rule { source, priority, target })
.labelled("rule")
}

View File

@@ -1,15 +1,13 @@
use std::collections::HashSet; use std::collections::HashSet;
use std::fs::File;
use std::iter; use std::iter;
use crate::{enum_parser, Expr, Clause}; use crate::{enum_parser, expression::{Expr, Clause, Rule}};
use crate::utils::BoxedIter; use crate::utils::BoxedIter;
use super::expression::xpr_parser; use super::expression::xpr_parser;
use super::import; use super::import;
use super::import::import_parser; use super::import::import_parser;
use super::lexer::Lexeme; use super::lexer::Lexeme;
use super::name;
use chumsky::{Parser, prelude::*}; use chumsky::{Parser, prelude::*};
use ordered_float::NotNan; use ordered_float::NotNan;
@@ -18,8 +16,7 @@ use ordered_float::NotNan;
pub enum FileEntry { pub enum FileEntry {
Import(Vec<import::Import>), Import(Vec<import::Import>),
Comment(String), Comment(String),
Rule(Vec<Expr>, NotNan<f64>, Vec<Expr>), Rule(Rule, bool)
Export(Vec<Expr>, NotNan<f64>, Vec<Expr>)
} }
/// Recursively iterate through all "names" in an expression. It also finds a lot of things that /// Recursively iterate through all "names" in an expression. It also finds a lot of things that
@@ -70,49 +67,21 @@ pub fn line_parser() -> impl Parser<Lexeme, FileEntry, Error = Simple<Lexeme>> {
println!("{:?} could not yield an export", s); e println!("{:?} could not yield an export", s); e
}) })
.ignore_then(rule_parser()) .ignore_then(rule_parser())
.map(|(lhs, prio, rhs)| FileEntry::Export(lhs, prio, rhs)), .map(|(source, prio, target)| FileEntry::Rule(Rule{source, prio, target}, true)),
// This could match almost anything so it has to go last // This could match almost anything so it has to go last
rule_parser().map(|(lhs, prio, rhs)| FileEntry::Rule(lhs, prio, rhs)), rule_parser().map(|(source, prio, target)| FileEntry::Rule(Rule{source, prio, target}, false)),
)) ))
} }
/// Collect all exported names (and a lot of other words) from a file /// Collect all exported names (and a lot of other words) from a file
pub fn exported_names(src: &Vec<FileEntry>) -> HashSet<&Vec<String>> { pub fn exported_names(src: &Vec<FileEntry>) -> HashSet<&Vec<String>> {
src.iter().flat_map(|ent| match ent { src.iter().flat_map(|ent| match ent {
FileEntry::Export(s, _, d) => Box::new(s.iter().chain(d.iter())) as BoxedIter<&Expr>, FileEntry::Rule(Rule{source, target, ..}, true) =>
Box::new(source.iter().chain(target.iter())) as BoxedIter<&Expr>,
_ => Box::new(iter::empty()) _ => Box::new(iter::empty())
}).map(find_all_names).flatten().collect() }).map(find_all_names).flatten().collect()
} }
// #[allow(dead_code)]
/// Collect all operators defined in a file (and some other words)
fn defined_ops(src: &Vec<FileEntry>, exported_only: bool) -> Vec<&String> {
let all_names:HashSet<&Vec<String>> = src.iter().flat_map(|ent| match ent {
FileEntry::Rule(s, _, d) =>
if exported_only {Box::new(iter::empty()) as BoxedIter<&Expr>}
else {Box::new(s.iter().chain(d.iter()))}
FileEntry::Export(s, _, d) => Box::new(s.iter().chain(d.iter())),
_ => Box::new(iter::empty())
}).map(find_all_names).flatten().collect();
// Dedupe stage of dubious value; collecting into a hashset may take longer than
// handling duplicates would with a file of sensible size.
all_names.into_iter()
.filter_map(|name|
// If it's namespaced, it's imported.
if name.len() == 1 && name::is_op(&name[0]) {Some(&name[0])}
else {None}
).collect()
}
// #[allow(dead_code)]
/// Collect all operators from a file
pub fn all_ops(src: &Vec<FileEntry>) -> Vec<&String> { defined_ops(src, false) }
// #[allow(dead_code)]
/// Collect exported operators from a file (plus some extra)
pub fn exported_ops(src: &Vec<FileEntry>) -> Vec<&String> { defined_ops(src, true) }
/// Summarize all imports from a file in a single list of qualified names /// Summarize all imports from a file in a single list of qualified names
pub fn imports<'a, 'b, I>( pub fn imports<'a, 'b, I>(
src: I src: I

View File

@@ -1,25 +0,0 @@
#[derive(Debug, Clone)]
pub enum Literal {
Num(f64),
Int(u64),
Char(char),
Str(String),
}
#[derive(Debug, Clone)]
pub enum Token {
Literal(Literal),
Name {
qualified: Vec<String>,
local: Option<String>
},
S(Vec<Expr>),
Lambda(String, Option<Box<Expr>>, Vec<Expr>),
Auto(Option<String>, Option<Box<Expr>>, Vec<Expr>)
}
#[derive(Debug, Clone)]
pub struct Expr {
pub token: Token,
pub typ: Option<Box<Expr>>
}

View File

@@ -40,8 +40,10 @@ pub fn file_loader(proj: PathBuf) -> impl FnMut(Vec<String>) -> Result<Loaded, L
let orcfile = dirpath.with_extension("orc"); let orcfile = dirpath.with_extension("orc");
if orcfile.is_file() { if orcfile.is_file() {
read_to_string(orcfile).map(Loaded::Module).map_err(LoadingError::from) read_to_string(orcfile).map(Loaded::Module).map_err(LoadingError::from)
} else if dirpath.exists() { } else {
Err(LoadingError::UnknownNode(dirpath.to_string_lossy().into_owned())) let pathstr = dirpath.to_string_lossy().into_owned();
} else { Err(LoadingError::Missing(dirpath.to_string_lossy().into_owned())) } Err(if dirpath.exists() { LoadingError::UnknownNode(pathstr) }
else { LoadingError::Missing(pathstr) })
}
} }
} }

View File

@@ -1,23 +1,10 @@
mod rule_collector; mod rule_collector;
// pub use rule_collector::rule_collector; pub use rule_collector::rule_collector;
mod prefix; mod prefix;
mod name_resolver; mod name_resolver;
mod loaded; mod loaded;
pub use loaded::Loaded; pub use loaded::Loaded;
mod parse_error; mod module_error;
mod file_loader; mod file_loader;
pub use file_loader::file_loader; pub use file_loader::file_loader;
use crate::expression::Rule;
#[derive(Debug, Clone)]
pub struct Module {
pub rules: Vec<Rule>,
pub exports: Vec<String>,
pub references: Vec<Vec<String>>
}
#[derive(Debug, Clone)]
pub struct Rule {
pub source: super::Expr,
pub priority: f64,
pub target: super::Expr
}

View File

@@ -0,0 +1,31 @@
use thiserror::Error;
use crate::parse::ParseError;
use super::name_resolver::ResolutionError;
#[derive(Error, Debug, Clone)]
pub enum ModuleError<ELoad> where ELoad: Clone {
#[error("Resolution cycle")]
ResolutionCycle,
#[error("File not found: {0}")]
Load(ELoad),
#[error("Failed to parse: {0:?}")]
Syntax(ParseError),
#[error("Not a module")]
None
}
impl<T> From<ParseError> for ModuleError<T> where T: Clone {
fn from(pars: ParseError) -> Self { Self::Syntax(pars) }
}
impl<T> From<ResolutionError<ModuleError<T>>> for ModuleError<T> where T: Clone {
fn from(res: ResolutionError<ModuleError<T>>) -> Self {
match res {
ResolutionError::Cycle(_) => ModuleError::ResolutionCycle,
ResolutionError::NoModule(_) => ModuleError::None,
ResolutionError::Delegate(d) => d
}
}
}

View File

@@ -3,7 +3,7 @@ use thiserror::Error;
use crate::utils::Substack; use crate::utils::Substack;
use crate::{Expr, Clause, Literal}; use crate::expression::{Expr, Clause};
type ImportMap = HashMap<String, Vec<String>>; type ImportMap = HashMap<String, Vec<String>>;

View File

@@ -1,30 +0,0 @@
use chumsky::prelude::Simple;
use thiserror::Error;
use super::name_resolver::ResolutionError;
#[derive(Error, Debug, Clone)]
pub enum ParseError<ELoad> where ELoad: Clone {
#[error("Resolution cycle")]
ResolutionCycle,
#[error("File not found: {0}")]
Load(ELoad),
#[error("Failed to parse: {0:?}")]
Syntax(Vec<Simple<char>>),
#[error("Not a module")]
None
}
impl<T> From<Vec<Simple<char>>> for ParseError<T> where T: Clone {
fn from(simp: Vec<Simple<char>>) -> Self { Self::Syntax(simp) }
}
impl<T> From<ResolutionError<ParseError<T>>> for ParseError<T> where T: Clone {
fn from(res: ResolutionError<ParseError<T>>) -> Self {
match res {
ResolutionError::Cycle(_) => ParseError::ResolutionCycle,
ResolutionError::NoModule(_) => ParseError::None,
ResolutionError::Delegate(d) => d
}
}
}

View File

@@ -1,4 +1,4 @@
use crate::{Expr, Clause}; use crate::expression::{Expr, Clause};
/// Replaces the first element of a name with the matching prefix from a prefix map /// Replaces the first element of a name with the matching prefix from a prefix map

View File

@@ -1,193 +1,206 @@
// use std::collections::{HashMap, HashSet, VecDeque}; use std::collections::{HashMap, HashSet, VecDeque};
// use std::fmt::Debug; use std::fmt::Debug;
// use std::rc::Rc; use std::rc::Rc;
// use chumsky::Parser; use crate::expression::Rule;
use crate::parse::{self, FileEntry};
use crate::utils::Cache;
// use crate::parse::{self, line_parser, FileEntry}; use super::name_resolver::NameResolver;
// use crate::utils::Cache; use super::module_error::ModuleError;
use super::prefix::prefix_expr;
use super::loaded::Loaded;
type ParseResult<T, ELoad> = Result<T, ModuleError<ELoad>>;
// use super::name_resolver::NameResolver; #[derive(Debug, Clone)]
// use super::parse_error::ParseError; pub struct Module {
// use super::prefix::prefix_expr; pub rules: Vec<Rule>,
// use super::loaded::Loaded; pub exports: Vec<String>,
pub references: Vec<Vec<String>>
}
// type ParseResult<T, ELoad> = Result<T, ParseError<ELoad>>; pub fn rule_collector<F: 'static, ELoad>(
mut load_mod: F,
// pub fn rule_collector<F: 'static, ELoad>( prelude: Vec<String>
// mut load_mod: F, // ) -> impl FnMut(Vec<String>) -> Result<&'a Vec<super::Rule>, ParseError<ELoad>> + 'a
// prelude: Vec<String> ) -> Cache<Vec<String>, Result<Vec<super::Rule>, ModuleError<ELoad>>>
// // ) -> impl FnMut(Vec<String>) -> Result<&'a Vec<super::Rule>, ParseError<ELoad>> + 'a where
// ) -> Cache<Vec<String>, Result<Vec<super::Rule>, ParseError<ELoad>>> F: FnMut(Vec<String>) -> Result<Loaded, ELoad>,
// where ELoad: Clone + Debug
// F: FnMut(Vec<String>) -> Result<Loaded, ELoad>, {
// ELoad: Clone + Debug // Map paths to a namespace with name list (folder) or module with source text (file)
// { let loaded = Rc::new(Cache::new(move |path: Vec<String>|
// // Map paths to a namespace with name list (folder) or module with source text (file) -> ParseResult<Loaded, ELoad> {
// let loaded = Rc::new(Cache::new(move |path: Vec<String>| load_mod(path).map_err(ModuleError::Load)
// -> ParseResult<Loaded, ELoad> { }));
// load_mod(path).map_err(ParseError::Load) // Map names to the longest prefix that points to a valid module
// })); let modname = Rc::new(Cache::new({
// // Map names to the longest prefix that points to a valid module let loaded = Rc::clone(&loaded);
// let modname = Rc::new(Cache::new({ move |symbol: Vec<String>| -> Result<Vec<String>, Vec<ModuleError<ELoad>>> {
// let loaded = Rc::clone(&loaded); let mut errv: Vec<ModuleError<ELoad>> = Vec::new();
// move |symbol: Vec<String>| -> Result<Vec<String>, Vec<ParseError<ELoad>>> { let reg_err = |e, errv: &mut Vec<ModuleError<ELoad>>| {
// let mut errv: Vec<ParseError<ELoad>> = Vec::new(); errv.push(e);
// let reg_err = |e, errv: &mut Vec<ParseError<ELoad>>| { if symbol.len() == errv.len() { Err(errv.clone()) }
// errv.push(e); else { Ok(()) }
// if symbol.len() == errv.len() { Err(errv.clone()) } };
// else { Ok(()) } loop {
// }; let (path, _) = symbol.split_at(symbol.len() - errv.len());
// loop { let pathv = path.to_vec();
// let (path, _) = symbol.split_at(symbol.len() - errv.len()); match loaded.try_find(&pathv) {
// let pathv = path.to_vec(); Ok(imports) => match imports.as_ref() {
// match loaded.try_find(&pathv) { Loaded::Module(_) => break Ok(pathv.clone()),
// Ok(imports) => match imports.as_ref() { _ => reg_err(ModuleError::None, &mut errv)?
// Loaded::Module(_) => break Ok(pathv.clone()), },
// _ => reg_err(ParseError::None, &mut errv)? Err(err) => reg_err(err, &mut errv)?
// }, }
// Err(err) => reg_err(err, &mut errv)? }
// } }
// } }));
// } // Preliminarily parse a file, substitution rules and imports are valid
// })); let preparsed = Rc::new(Cache::new({
// // Preliminarily parse a file, substitution rules and imports are valid let loaded = Rc::clone(&loaded);
// let preparsed = Rc::new(Cache::new({ let prelude2 = prelude.clone();
// let preparser = line_parser(&prelude, &prelude); move |path: Vec<String>| -> ParseResult<Vec<FileEntry>, ELoad> {
// let loaded = Rc::clone(&loaded); let loaded = loaded.try_find(&path)?;
// move |path: Vec<String>| -> ParseResult<Vec<FileEntry>, ELoad> { if let Loaded::Module(source) = loaded.as_ref() {
// let loaded = loaded.try_find(&path)?; Ok(parse::parse(&prelude2, source.as_str())?)
// if let Loaded::Module(source) = loaded.as_ref() { } else {Err(ModuleError::None)}
// Ok(preparser.parse(source.as_str())?) }
// } else {Err(ParseError::None)} }));
// } // Collect all toplevel names exported from a given file
// })); let exports = Rc::new(Cache::new({
// // Collect all toplevel names exported from a given file let loaded = Rc::clone(&loaded);
// let exports = Rc::new(Cache::new({ let preparsed = Rc::clone(&preparsed);
// let loaded = Rc::clone(&loaded); move |path: Vec<String>| -> ParseResult<Vec<String>, ELoad> {
// let preparsed = Rc::clone(&preparsed); let loaded = loaded.try_find(&path)?;
// move |path: Vec<String>| -> ParseResult<Vec<String>, ELoad> { if let Loaded::Namespace(names) = loaded.as_ref() {
// let loaded = loaded.try_find(&path)?; return Ok(names.clone());
// if let Loaded::Namespace(names) = loaded.as_ref() { }
// return Ok(names.clone()); let preparsed = preparsed.try_find(&path)?;
// } Ok(parse::exported_names(&preparsed)
// let preparsed = preparsed.try_find(&path)?; .into_iter()
// Ok(parse::exported_names(&preparsed) .map(|n| n[0].clone())
// .into_iter() .collect())
// .map(|n| n[0].clone()) }
// .collect()) }));
// } // Collect all toplevel names imported by a given file
// })); let imports = Rc::new(Cache::new({
// // Collect all toplevel names imported by a given file let preparsed = Rc::clone(&preparsed);
// let imports = Rc::new(Cache::new({ let exports = Rc::clone(&exports);
// let preparsed = Rc::clone(&preparsed); move |path: Vec<String>| -> ParseResult<HashMap<String, Vec<String>>, ELoad> {
// let exports = Rc::clone(&exports); let entv = preparsed.try_find(&path)?.clone();
// move |path: Vec<String>| -> ParseResult<HashMap<String, Vec<String>>, ELoad> { let import_entries = parse::imports(entv.iter());
// let entv = preparsed.try_find(&path)?.clone(); let mut imported_symbols: HashMap<String, Vec<String>> = HashMap::new();
// let import_entries = parse::imports(entv.iter()); for imp in import_entries {
// let mut imported_symbols: HashMap<String, Vec<String>> = HashMap::new(); let export = exports.try_find(&imp.path)?;
// for imp in import_entries { if let Some(ref name) = imp.name {
// let export = exports.try_find(&imp.path)?; if export.contains(&name) {
// if let Some(ref name) = imp.name { imported_symbols.insert(name.clone(), imp.path.clone());
// if export.contains(&name) { }
// imported_symbols.insert(name.clone(), imp.path.clone()); } else {
// } for exp in export.as_ref() {
// } else { imported_symbols.insert(exp.clone(), imp.path.clone());
// for exp in export.as_ref() { }
// imported_symbols.insert(exp.clone(), imp.path.clone()); }
// } }
// } Ok(imported_symbols)
// } }
// Ok(imported_symbols) }));
// } // Final parse, operators are correctly separated
// })); let parsed = Rc::new(Cache::new({
// // Final parse, operators are correctly separated let preparsed = Rc::clone(&preparsed);
// let parsed = Rc::new(Cache::new({ let imports = Rc::clone(&imports);
// let imports = Rc::clone(&imports); let loaded = Rc::clone(&loaded);
// let loaded = Rc::clone(&loaded); move |path: Vec<String>| -> ParseResult<Vec<FileEntry>, ELoad> {
// move |path: Vec<String>| -> ParseResult<Vec<FileEntry>, ELoad> { let imported_ops: Vec<String> =
// let imported_ops: Vec<String> = imports.try_find(&path)?
// imports.try_find(&path)? .keys()
// .keys() .chain(prelude.iter())
// .chain(prelude.iter()) .filter(|s| parse::is_op(s))
// .filter(|s| parse::is_op(s)) .cloned()
// .cloned() .collect();
// .collect(); // let parser = file_parser(&prelude, &imported_ops);
// let parser = file_parser(&prelude, &imported_ops); let pre = preparsed.try_find(&path)?;
// if let Loaded::Module(source) = loaded.try_find(&path)?.as_ref() { if let Loaded::Module(source) = loaded.try_find(&path)?.as_ref() {
// Ok(parser.parse(source.as_str())?) Ok(parse::reparse(&imported_ops, source.as_str(), &pre)?)
// } else { Err(ParseError::None) } } else { Err(ModuleError::None) }
// } }
// })); }));
// let mut name_resolver = NameResolver::new({ let mut name_resolver = NameResolver::new({
// let modname = Rc::clone(&modname); let modname = Rc::clone(&modname);
// move |path| { move |path| {
// Some(modname.try_find(path).ok()?.as_ref().clone()) Some(modname.try_find(path).ok()?.as_ref().clone())
// } }
// }, { }, {
// let imports = Rc::clone(&imports); let imports = Rc::clone(&imports);
// move |path| { move |path| {
// imports.try_find(path).map(|f| f.as_ref().clone()) imports.try_find(path).map(|f| f.as_ref().clone())
// } }
// }); });
// // Turn parsed files into a bag of rules and a list of toplevel export names // Turn parsed files into a bag of rules and a list of toplevel export names
// let resolved = Rc::new(Cache::new({ let resolved = Rc::new(Cache::new({
// let parsed = Rc::clone(&parsed); let parsed = Rc::clone(&parsed);
// let exports = Rc::clone(&exports); let exports = Rc::clone(&exports);
// let imports = Rc::clone(&imports); let imports = Rc::clone(&imports);
// let modname = Rc::clone(&modname); let modname = Rc::clone(&modname);
// move |path: Vec<String>| -> ParseResult<super::Module, ELoad> { move |path: Vec<String>| -> ParseResult<Module, ELoad> {
// let module = super::Module { let module = Module {
// rules: parsed.try_find(&path)? rules: parsed.try_find(&path)?
// .iter() .iter()
// .filter_map(|ent| { .filter_map(|ent| {
// if let FileEntry::Export(s) | FileEntry::Rule(s) = ent { if let FileEntry::Rule(Rule{source, prio, target}, _) = ent {
// Some(super::Rule { Some(Rule {
// source: prefix_expr(&s.source, &path), source: source.iter().map(|ex| prefix_expr(ex, &path)).collect(),
// target: prefix_expr(&s.target, &path), target: target.iter().map(|ex| prefix_expr(ex, &path)).collect(),
// priority: s.priority, prio: *prio,
// }) })
// } else { None } } else { None }
// }) })
// .map(|rule| Ok(super::Rule { .map(|rule| Ok(super::Rule {
// source: name_resolver.process_expression(&rule.source)?, source: rule.source.iter()
// target: name_resolver.process_expression(&rule.target)?, .map(|ex| name_resolver.process_expression(ex))
// ..rule .collect::<Result<Vec<_>, _>>()?,
// })) target: rule.target.iter()
// .collect::<ParseResult<Vec<super::Rule>, ELoad>>()?, .map(|ex| name_resolver.process_expression(ex))
// exports: exports.try_find(&path)?.as_ref().clone(), .collect::<Result<Vec<_>, _>>()?,
// references: imports.try_find(&path)? // source: name_resolver.process_expression(&rule.source)?,
// .values() // target: name_resolver.process_expression(&rule.target)?,
// .filter_map(|imps| { ..rule
// modname.try_find(&imps).ok().map(|r| r.as_ref().clone()) }))
// }) .collect::<ParseResult<Vec<super::Rule>, ELoad>>()?,
// .collect() exports: exports.try_find(&path)?.as_ref().clone(),
// }; references: imports.try_find(&path)?
// Ok(module) .values()
// } .filter_map(|imps| {
// })); modname.try_find(&imps).ok().map(|r| r.as_ref().clone())
// let all_rules = Cache::new({ })
// let resolved = Rc::clone(&resolved); .collect()
// move |path: Vec<String>| -> ParseResult<Vec<super::Rule>, ELoad> { };
// let mut processed: HashSet<Vec<String>> = HashSet::new(); Ok(module)
// let mut rules: Vec<super::Rule> = Vec::new(); }
// let mut pending: VecDeque<Vec<String>> = VecDeque::new(); }));
// pending.push_back(path); let all_rules = Cache::new({
// while let Some(el) = pending.pop_front() { let resolved = Rc::clone(&resolved);
// let resolved = resolved.try_find(&el)?; move |path: Vec<String>| -> ParseResult<Vec<super::Rule>, ELoad> {
// processed.insert(el.clone()); let mut processed: HashSet<Vec<String>> = HashSet::new();
// pending.extend( let mut rules: Vec<super::Rule> = Vec::new();
// resolved.references.iter() let mut pending: VecDeque<Vec<String>> = VecDeque::new();
// .filter(|&v| !processed.contains(v)) pending.push_back(path);
// .cloned() while let Some(el) = pending.pop_front() {
// ); let resolved = resolved.try_find(&el)?;
// rules.extend( processed.insert(el.clone());
// resolved.rules.iter().cloned() pending.extend(
// ) resolved.references.iter()
// }; .filter(|&v| !processed.contains(v))
// Ok(rules) .cloned()
// } );
// }); rules.extend(
// return all_rules; resolved.rules.iter().cloned()
// } )
};
Ok(rules)
}
});
return all_rules;
}

1
src/rule/mod.rs Normal file
View File

@@ -0,0 +1 @@
mod rule;

38
src/rule/rule.rs Normal file
View File

@@ -0,0 +1,38 @@
use std::cmp::{min, max};
use hashbrown::HashSet;
use crate::expression::Expr;
pub trait Rule {
type OutIter: Iterator<Item = Option<Expr>>;
/// The minimum and maximum set of symbols this rule may match.
fn len(&self) -> (Option<usize>, Option<usize>);
/// The exact tokens the pattern consumes (None if varies)
fn consumes(&self) -> Option<HashSet<Vec<String>>>;
/// The exact tokens the pattern produces (None if varies)
fn produces(&self) -> Option<HashSet<Vec<String>>>;
/// Check if the slice matches, and produce the necessary transformations
fn produce(&self, base: &[Expr]) -> Option<Self::OutIter>;
/// Try all subsections of Vec of appropriate size, longest first, front-to-back
/// Match the first, execute the substitution, return the vector and whether any
/// substitutions happened
fn apply(&self, mut base: Vec<Expr>) -> (Vec<Expr>, bool) {
let len_range = self.len();
let lo = max(len_range.0.unwrap_or(1), 1);
let hi = min(len_range.1.unwrap_or(base.len()), base.len());
for width in (lo..hi).rev() {
let starts = (0..base.len() - width).into_iter();
let first_match = starts.filter_map(|start| {
self.produce(&base[start..start+width])
.map(|res| (start, res))
}).next();
if let Some((start, substitution)) = first_match {
let diff = substitution.enumerate().filter_map(|(i, opt)| opt.map(|val| (i, val)));
for (idx, item) in diff { base[start + idx] = item }
return (base, true)
}
}
(base, false)
}
}

View File

@@ -1,8 +1,6 @@
mod cache; mod cache;
mod substack; mod substack;
mod result_iter_collect;
pub use cache::Cache; pub use cache::Cache;
pub use substack::Substack; pub use substack::Substack;
pub use result_iter_collect::result_iter_collect;
pub type BoxedIter<'a, T> = Box<dyn Iterator<Item = T> + 'a>; pub type BoxedIter<'a, T> = Box<dyn Iterator<Item = T> + 'a>;

View File

@@ -1,19 +0,0 @@
pub fn result_iter_collect<T, E>(i: &mut dyn Iterator<Item = Result<T, E>>)
-> (Vec<Option<T>>, Vec<Option<E>>) {
i.fold((Vec::new(), Vec::new()), |(mut succ, mut err), mut next| {
match next {
Ok(res) => succ.push(Some(res)),
Err(e) => err.push(Some(e))
}
(succ, err)
})
}
pub fn recoverable_iter_collect<T, E>(i: &mut dyn Iterator<Item=(Option<T>, Vec<E>)>)
-> (Vec<Option<T>>, Vec<E>) {
i.fold((Vec::new(), Vec::new()), |(mut succ, mut err), (res, mut errv)| {
succ.push(res);
err.append(&mut errv);
(succ, err)
})
}