Skip to content

Commit

Permalink
Enable float32/float64 by default.
Browse files Browse the repository at this point in the history
  • Loading branch information
sunfishcode committed Jan 9, 2024
1 parent 8bb43ec commit a0226e4
Show file tree
Hide file tree
Showing 3 changed files with 17 additions and 6 deletions.
2 changes: 1 addition & 1 deletion crates/wit-component/src/printing.rs
Original file line number Diff line number Diff line change
Expand Up @@ -6,7 +6,7 @@ use wit_parser::*;

// NB: keep in sync with `crates/wit-parser/src/ast/lex.rs`
const PRINT_SEMICOLONS_DEFAULT: bool = true;
const PRINT_F32_F64_DEFAULT: bool = true;
const PRINT_F32_F64_DEFAULT: bool = false;

/// A utility for printing WebAssembly interface definitions to a string.
pub struct WitPrinter {
Expand Down
17 changes: 14 additions & 3 deletions crates/wit-parser/src/ast.rs
Original file line number Diff line number Diff line change
Expand Up @@ -1160,6 +1160,7 @@ pub struct SourceMap {
sources: Vec<Source>,
offset: u32,
require_semicolons: Option<bool>,
require_f32_f64: Option<bool>,
}

#[derive(Clone)]
Expand All @@ -1180,6 +1181,11 @@ impl SourceMap {
self.require_semicolons = Some(enable);
}

#[doc(hidden)] // NB: only here for a transitionary period
pub fn set_require_f32_f64(&mut self, enable: bool) {
self.require_f32_f64 = Some(enable);
}

/// Reads the file `path` on the filesystem and appends its contents to this
/// [`SourceMap`].
pub fn push_file(&mut self, path: &Path) -> Result<()> {
Expand Down Expand Up @@ -1214,8 +1220,13 @@ impl SourceMap {
let mut srcs = self.sources.iter().collect::<Vec<_>>();
srcs.sort_by_key(|src| &src.path);
for src in srcs {
let mut tokens = Tokenizer::new(&src.contents, src.offset, self.require_semicolons)
.with_context(|| format!("failed to tokenize path: {}", src.path.display()))?;
let mut tokens = Tokenizer::new(
&src.contents,
src.offset,
self.require_semicolons,
self.require_f32_f64,
)
.with_context(|| format!("failed to tokenize path: {}", src.path.display()))?;
let ast = Ast::parse(&mut tokens)?;
resolver.push(ast).with_context(|| {
format!("failed to start resolving path: {}", src.path.display())
Expand Down Expand Up @@ -1324,7 +1335,7 @@ pub(crate) enum AstUsePath {
}

pub(crate) fn parse_use_path(s: &str) -> Result<AstUsePath> {
let mut tokens = Tokenizer::new(s, 0, Some(true))?;
let mut tokens = Tokenizer::new(s, 0, Some(true), None)?;
let path = UsePath::parse(&mut tokens)?;
if tokens.next()?.is_some() {
bail!("trailing tokens in path specifier");
Expand Down
4 changes: 2 additions & 2 deletions crates/wit-parser/src/ast/lex.rs
Original file line number Diff line number Diff line change
Expand Up @@ -119,7 +119,7 @@ pub enum Error {

// NB: keep in sync with `crates/wit-component/src/printing.rs`.
const REQUIRE_SEMICOLONS_BY_DEFAULT: bool = true;
const REQUIRE_F32_F64_BY_DEFAULT: bool = true;
const REQUIRE_F32_F64_BY_DEFAULT: bool = false;

impl<'a> Tokenizer<'a> {
pub fn new(
Expand Down Expand Up @@ -665,7 +665,7 @@ fn test_validate_id() {
#[test]
fn test_tokenizer() {
fn collect(s: &str) -> Result<Vec<Token>> {
let mut t = Tokenizer::new(s, 0, Some(true))?;
let mut t = Tokenizer::new(s, 0, Some(true), None)?;
let mut tokens = Vec::new();
while let Some(token) = t.next()? {
tokens.push(token.1);
Expand Down

0 comments on commit a0226e4

Please sign in to comment.