[INFO] fetching crate oak-typst 0.0.11... [INFO] testing oak-typst-0.0.11 against try#dec9417b8611e34e787a3e4c37686b5131f9e5c5 for pr-154210-2 [INFO] extracting crate oak-typst 0.0.11 into /workspace/builds/worker-5-tc2/source [INFO] started tweaking crates.io crate oak-typst 0.0.11 [INFO] removed 0 missing tests [INFO] finished tweaking crates.io crate oak-typst 0.0.11 [INFO] tweaked toml for crates.io crate oak-typst 0.0.11 written to /workspace/builds/worker-5-tc2/source/Cargo.toml [INFO] validating manifest of crates.io crate oak-typst 0.0.11 on toolchain dec9417b8611e34e787a3e4c37686b5131f9e5c5 [INFO] running `Command { std: CARGO_HOME="/workspace/cargo-home" RUSTUP_HOME="/workspace/rustup-home" "/workspace/cargo-home/bin/cargo" "+dec9417b8611e34e787a3e4c37686b5131f9e5c5" "metadata" "--manifest-path" "Cargo.toml" "--no-deps", kill_on_drop: false }` [INFO] crate crates.io crate oak-typst 0.0.11 already has a lockfile, it will not be regenerated [INFO] running `Command { std: CARGO_HOME="/workspace/cargo-home" RUSTUP_HOME="/workspace/rustup-home" "/workspace/cargo-home/bin/cargo" "+dec9417b8611e34e787a3e4c37686b5131f9e5c5" "fetch" "--manifest-path" "Cargo.toml", kill_on_drop: false }` [INFO] running `Command { std: "docker" "create" "-v" "/var/lib/crater-agent-workspace/builds/worker-5-tc2/target:/opt/rustwide/target:rw,Z" "-v" "/var/lib/crater-agent-workspace/builds/worker-5-tc2/source:/opt/rustwide/workdir:ro,Z" "-v" "/var/lib/crater-agent-workspace/cargo-home:/opt/rustwide/cargo-home:ro,Z" "-v" "/var/lib/crater-agent-workspace/rustup-home:/opt/rustwide/rustup-home:ro,Z" "-e" "SOURCE_DIR=/opt/rustwide/workdir" "-e" "CARGO_TARGET_DIR=/opt/rustwide/target" "-e" "CARGO_HOME=/opt/rustwide/cargo-home" "-e" "RUSTUP_HOME=/opt/rustwide/rustup-home" "-w" "/opt/rustwide/workdir" "-m" "1610612736" "--user" "0:0" "--network" "none" "ghcr.io/rust-lang/crates-build-env/linux@sha256:d429b63d4308055ea97f60fb1d3dfca48854a00942f1bd2ad806beaf015945ec" "/opt/rustwide/cargo-home/bin/cargo" "+dec9417b8611e34e787a3e4c37686b5131f9e5c5" "metadata" "--no-deps" "--format-version=1", kill_on_drop: false }` [INFO] [stdout] 427a96d3f460a6be09c9dc3578cc6662ceddc9ba7f16f5016a1727db89117066 [INFO] running `Command { std: "docker" "start" "-a" "427a96d3f460a6be09c9dc3578cc6662ceddc9ba7f16f5016a1727db89117066", kill_on_drop: false }` [INFO] running `Command { std: "docker" "inspect" "427a96d3f460a6be09c9dc3578cc6662ceddc9ba7f16f5016a1727db89117066", kill_on_drop: false }` [INFO] running `Command { std: "docker" "rm" "-f" "427a96d3f460a6be09c9dc3578cc6662ceddc9ba7f16f5016a1727db89117066", kill_on_drop: false }` [INFO] [stdout] 427a96d3f460a6be09c9dc3578cc6662ceddc9ba7f16f5016a1727db89117066 [INFO] running `Command { std: "docker" "create" "-v" "/var/lib/crater-agent-workspace/builds/worker-5-tc2/target:/opt/rustwide/target:rw,Z" "-v" "/var/lib/crater-agent-workspace/builds/worker-5-tc2/source:/opt/rustwide/workdir:ro,Z" "-v" "/var/lib/crater-agent-workspace/cargo-home:/opt/rustwide/cargo-home:ro,Z" "-v" "/var/lib/crater-agent-workspace/rustup-home:/opt/rustwide/rustup-home:ro,Z" "-e" "SOURCE_DIR=/opt/rustwide/workdir" "-e" "CARGO_TARGET_DIR=/opt/rustwide/target" "-e" "CARGO_INCREMENTAL=0" "-e" "RUST_BACKTRACE=full" "-e" "RUSTFLAGS=--cap-lints=forbid" "-e" "RUSTDOCFLAGS=--cap-lints=forbid" "-e" "CARGO_HOME=/opt/rustwide/cargo-home" "-e" "RUSTUP_HOME=/opt/rustwide/rustup-home" "-w" "/opt/rustwide/workdir" "-m" "1610612736" "--user" "0:0" "--network" "none" "ghcr.io/rust-lang/crates-build-env/linux@sha256:d429b63d4308055ea97f60fb1d3dfca48854a00942f1bd2ad806beaf015945ec" "/opt/rustwide/cargo-home/bin/cargo" "+dec9417b8611e34e787a3e4c37686b5131f9e5c5" "build" "--frozen" "--message-format=json", kill_on_drop: false }` [INFO] [stdout] f32855073e7e6d98db54f206a71904739d5701649aa4f2e4a9d6a1e9ee8b33ef [INFO] running `Command { std: "docker" "start" "-a" "f32855073e7e6d98db54f206a71904739d5701649aa4f2e4a9d6a1e9ee8b33ef", kill_on_drop: false }` [INFO] [stderr] Compiling libc v0.2.182 [INFO] [stderr] Compiling zerocopy v0.8.40 [INFO] [stderr] Compiling serde_core v1.0.228 [INFO] [stderr] Compiling serde v1.0.228 [INFO] [stderr] Compiling smallvec v1.15.1 [INFO] [stderr] Compiling syn v2.0.117 [INFO] [stderr] Compiling getrandom v0.3.4 [INFO] [stderr] Compiling rand_core v0.9.5 [INFO] [stderr] Compiling serde_derive v1.0.228 [INFO] [stderr] Compiling tracing-attributes v0.1.31 [INFO] [stderr] Compiling ppv-lite86 v0.2.21 [INFO] [stderr] Compiling rand_chacha v0.9.0 [INFO] [stderr] Compiling tracing v0.1.44 [INFO] [stderr] Compiling rand v0.9.2 [INFO] [stderr] Compiling twox-hash v2.1.2 [INFO] [stderr] Compiling triomphe v0.1.15 [INFO] [stderr] Compiling oak-core v0.0.11 [INFO] [stderr] Compiling oak-typst v0.0.11 (/opt/rustwide/workdir) [INFO] [stdout] warning: unused imports: `ast::TypstRoot` and `lexer::token_type::TypstTokenType` [INFO] [stdout] --> src/language/mod.rs:2:13 [INFO] [stdout] | [INFO] [stdout] 2 | use crate::{ast::TypstRoot, lexer::token_type::TypstTokenType}; [INFO] [stdout] | ^^^^^^^^^^^^^^ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] | [INFO] [stdout] = note: `#[warn(unused_imports)]` (part of `#[warn(unused)]`) on by default [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused imports: `Source` and `UniversalElementRole` [INFO] [stdout] --> src/lexer/token_type.rs:1:16 [INFO] [stdout] | [INFO] [stdout] 1 | use oak_core::{Source, Token, TokenType, UniversalElementRole, UniversalTokenRole}; [INFO] [stdout] | ^^^^^^ ^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `Parser` [INFO] [stdout] --> src/parser/element_type.rs:1:29 [INFO] [stdout] | [INFO] [stdout] 1 | use oak_core::{ElementType, Parser, UniversalElementRole}; [INFO] [stdout] | ^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `GreenNode` [INFO] [stdout] --> src/parser/mod.rs:10:5 [INFO] [stdout] | [INFO] [stdout] 10 | GreenNode, OakError, [INFO] [stdout] | ^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: variable `count` is assigned to, but never used [INFO] [stdout] --> src/lexer/mod.rs:155:25 [INFO] [stdout] | [INFO] [stdout] 155 | let mut count = 0; [INFO] [stdout] | ^^^^^^^^^ [INFO] [stdout] | [INFO] [stdout] = note: consider using `_count` instead [INFO] [stdout] = note: `#[warn(unused_variables)]` (part of `#[warn(unused)]`) on by default [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: value assigned to `count` is never read [INFO] [stdout] --> src/lexer/mod.rs:157:25 [INFO] [stdout] | [INFO] [stdout] 157 | count += 1; [INFO] [stdout] | ^^^^^^^^^^ [INFO] [stdout] | [INFO] [stdout] = help: maybe it is overwritten before being read? [INFO] [stdout] = note: `#[warn(unused_assignments)]` (part of `#[warn(unused)]`) on by default [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: field `config` is never read [INFO] [stdout] --> src/lexer/mod.rs:22:5 [INFO] [stdout] | [INFO] [stdout] 21 | pub struct TypstLexer<'config> { [INFO] [stdout] | ---------- field in this struct [INFO] [stdout] 22 | config: &'config TypstLanguage, [INFO] [stdout] | ^^^^^^ [INFO] [stdout] | [INFO] [stdout] = note: `TypstLexer` has derived impls for the traits `Debug` and `Clone`, but these are intentionally ignored during dead code analysis [INFO] [stdout] = note: `#[warn(dead_code)]` (part of `#[warn(unused)]`) on by default [INFO] [stdout] [INFO] [stdout] [INFO] [stderr] Finished `dev` profile [unoptimized + debuginfo] target(s) in 15.11s [INFO] running `Command { std: "docker" "inspect" "f32855073e7e6d98db54f206a71904739d5701649aa4f2e4a9d6a1e9ee8b33ef", kill_on_drop: false }` [INFO] running `Command { std: "docker" "rm" "-f" "f32855073e7e6d98db54f206a71904739d5701649aa4f2e4a9d6a1e9ee8b33ef", kill_on_drop: false }` [INFO] [stdout] f32855073e7e6d98db54f206a71904739d5701649aa4f2e4a9d6a1e9ee8b33ef [INFO] running `Command { std: "docker" "create" "-v" "/var/lib/crater-agent-workspace/builds/worker-5-tc2/target:/opt/rustwide/target:rw,Z" "-v" "/var/lib/crater-agent-workspace/builds/worker-5-tc2/source:/opt/rustwide/workdir:ro,Z" "-v" "/var/lib/crater-agent-workspace/cargo-home:/opt/rustwide/cargo-home:ro,Z" "-v" "/var/lib/crater-agent-workspace/rustup-home:/opt/rustwide/rustup-home:ro,Z" "-e" "SOURCE_DIR=/opt/rustwide/workdir" "-e" "CARGO_TARGET_DIR=/opt/rustwide/target" "-e" "CARGO_INCREMENTAL=0" "-e" "RUST_BACKTRACE=full" "-e" "RUSTFLAGS=--cap-lints=forbid" "-e" "RUSTDOCFLAGS=--cap-lints=forbid" "-e" "CARGO_HOME=/opt/rustwide/cargo-home" "-e" "RUSTUP_HOME=/opt/rustwide/rustup-home" "-w" "/opt/rustwide/workdir" "-m" "1610612736" "--user" "0:0" "--network" "none" "ghcr.io/rust-lang/crates-build-env/linux@sha256:d429b63d4308055ea97f60fb1d3dfca48854a00942f1bd2ad806beaf015945ec" "/opt/rustwide/cargo-home/bin/cargo" "+dec9417b8611e34e787a3e4c37686b5131f9e5c5" "test" "--frozen" "--no-run" "--message-format=json", kill_on_drop: false }` [INFO] [stdout] b48ed4120de350c70dd5c90096a9aa5b33c3e51a18466fc75f7c6d77f5a4257b [INFO] running `Command { std: "docker" "start" "-a" "b48ed4120de350c70dd5c90096a9aa5b33c3e51a18466fc75f7c6d77f5a4257b", kill_on_drop: false }` [INFO] [stderr] Compiling memchr v2.8.0 [INFO] [stdout] warning: unused imports: `ast::TypstRoot` and `lexer::token_type::TypstTokenType` [INFO] [stdout] --> src/language/mod.rs:2:13 [INFO] [stdout] | [INFO] [stdout] 2 | use crate::{ast::TypstRoot, lexer::token_type::TypstTokenType}; [INFO] [stdout] | ^^^^^^^^^^^^^^ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] | [INFO] [stdout] = note: `#[warn(unused_imports)]` (part of `#[warn(unused)]`) on by default [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused imports: `Source` and `UniversalElementRole` [INFO] [stdout] --> src/lexer/token_type.rs:1:16 [INFO] [stdout] | [INFO] [stdout] 1 | use oak_core::{Source, Token, TokenType, UniversalElementRole, UniversalTokenRole}; [INFO] [stdout] | ^^^^^^ ^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `Parser` [INFO] [stdout] --> src/parser/element_type.rs:1:29 [INFO] [stdout] | [INFO] [stdout] 1 | use oak_core::{ElementType, Parser, UniversalElementRole}; [INFO] [stdout] | ^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `GreenNode` [INFO] [stdout] --> src/parser/mod.rs:10:5 [INFO] [stdout] | [INFO] [stdout] 10 | GreenNode, OakError, [INFO] [stdout] | ^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: variable `count` is assigned to, but never used [INFO] [stdout] --> src/lexer/mod.rs:155:25 [INFO] [stdout] | [INFO] [stdout] 155 | let mut count = 0; [INFO] [stdout] | ^^^^^^^^^ [INFO] [stdout] | [INFO] [stdout] = note: consider using `_count` instead [INFO] [stdout] = note: `#[warn(unused_variables)]` (part of `#[warn(unused)]`) on by default [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: value assigned to `count` is never read [INFO] [stdout] --> src/lexer/mod.rs:157:25 [INFO] [stdout] | [INFO] [stdout] 157 | count += 1; [INFO] [stdout] | ^^^^^^^^^^ [INFO] [stdout] | [INFO] [stdout] = help: maybe it is overwritten before being read? [INFO] [stdout] = note: `#[warn(unused_assignments)]` (part of `#[warn(unused)]`) on by default [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: field `config` is never read [INFO] [stdout] --> src/lexer/mod.rs:22:5 [INFO] [stdout] | [INFO] [stdout] 21 | pub struct TypstLexer<'config> { [INFO] [stdout] | ---------- field in this struct [INFO] [stdout] 22 | config: &'config TypstLanguage, [INFO] [stdout] | ^^^^^^ [INFO] [stdout] | [INFO] [stdout] = note: `TypstLexer` has derived impls for the traits `Debug` and `Clone`, but these are intentionally ignored during dead code analysis [INFO] [stdout] = note: `#[warn(dead_code)]` (part of `#[warn(unused)]`) on by default [INFO] [stdout] [INFO] [stdout] [INFO] [stderr] Compiling serde_json v1.0.149 [INFO] [stderr] Compiling oak-testing v0.0.11 [INFO] [stderr] Compiling oak-typst v0.0.11 (/opt/rustwide/workdir) [INFO] [stdout] warning: unused imports: `ast::TypstRoot` and `lexer::token_type::TypstTokenType` [INFO] [stdout] --> src/language/mod.rs:2:13 [INFO] [stdout] | [INFO] [stdout] 2 | use crate::{ast::TypstRoot, lexer::token_type::TypstTokenType}; [INFO] [stdout] | ^^^^^^^^^^^^^^ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] | [INFO] [stdout] = note: `#[warn(unused_imports)]` (part of `#[warn(unused)]`) on by default [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused imports: `Source` and `UniversalElementRole` [INFO] [stdout] --> src/lexer/token_type.rs:1:16 [INFO] [stdout] | [INFO] [stdout] 1 | use oak_core::{Source, Token, TokenType, UniversalElementRole, UniversalTokenRole}; [INFO] [stdout] | ^^^^^^ ^^^^^^^^^^^^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `Parser` [INFO] [stdout] --> src/parser/element_type.rs:1:29 [INFO] [stdout] | [INFO] [stdout] 1 | use oak_core::{ElementType, Parser, UniversalElementRole}; [INFO] [stdout] | ^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: unused import: `GreenNode` [INFO] [stdout] --> src/parser/mod.rs:10:5 [INFO] [stdout] | [INFO] [stdout] 10 | GreenNode, OakError, [INFO] [stdout] | ^^^^^^^^^ [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: variable `count` is assigned to, but never used [INFO] [stdout] --> src/lexer/mod.rs:155:25 [INFO] [stdout] | [INFO] [stdout] 155 | let mut count = 0; [INFO] [stdout] | ^^^^^^^^^ [INFO] [stdout] | [INFO] [stdout] = note: consider using `_count` instead [INFO] [stdout] = note: `#[warn(unused_variables)]` (part of `#[warn(unused)]`) on by default [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: value assigned to `count` is never read [INFO] [stdout] --> src/lexer/mod.rs:157:25 [INFO] [stdout] | [INFO] [stdout] 157 | count += 1; [INFO] [stdout] | ^^^^^^^^^^ [INFO] [stdout] | [INFO] [stdout] = help: maybe it is overwritten before being read? [INFO] [stdout] = note: `#[warn(unused_assignments)]` (part of `#[warn(unused)]`) on by default [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] warning: field `config` is never read [INFO] [stdout] --> src/lexer/mod.rs:22:5 [INFO] [stdout] | [INFO] [stdout] 21 | pub struct TypstLexer<'config> { [INFO] [stdout] | ---------- field in this struct [INFO] [stdout] 22 | config: &'config TypstLanguage, [INFO] [stdout] | ^^^^^^ [INFO] [stdout] | [INFO] [stdout] = note: `TypstLexer` has derived impls for the traits `Debug` and `Clone`, but these are intentionally ignored during dead code analysis [INFO] [stdout] = note: `#[warn(dead_code)]` (part of `#[warn(unused)]`) on by default [INFO] [stdout] [INFO] [stdout] [INFO] [stderr] Finished `test` profile [unoptimized + debuginfo] target(s) in 4.07s [INFO] running `Command { std: "docker" "inspect" "b48ed4120de350c70dd5c90096a9aa5b33c3e51a18466fc75f7c6d77f5a4257b", kill_on_drop: false }` [INFO] running `Command { std: "docker" "rm" "-f" "b48ed4120de350c70dd5c90096a9aa5b33c3e51a18466fc75f7c6d77f5a4257b", kill_on_drop: false }` [INFO] [stdout] b48ed4120de350c70dd5c90096a9aa5b33c3e51a18466fc75f7c6d77f5a4257b [INFO] running `Command { std: "docker" "create" "-v" "/var/lib/crater-agent-workspace/builds/worker-5-tc2/target:/opt/rustwide/target:rw,Z" "-v" "/var/lib/crater-agent-workspace/builds/worker-5-tc2/source:/opt/rustwide/workdir:ro,Z" "-v" "/var/lib/crater-agent-workspace/cargo-home:/opt/rustwide/cargo-home:ro,Z" "-v" "/var/lib/crater-agent-workspace/rustup-home:/opt/rustwide/rustup-home:ro,Z" "-e" "SOURCE_DIR=/opt/rustwide/workdir" "-e" "CARGO_TARGET_DIR=/opt/rustwide/target" "-e" "CARGO_INCREMENTAL=0" "-e" "RUST_BACKTRACE=full" "-e" "RUSTFLAGS=--cap-lints=forbid" "-e" "RUSTDOCFLAGS=--cap-lints=forbid" "-e" "CARGO_HOME=/opt/rustwide/cargo-home" "-e" "RUSTUP_HOME=/opt/rustwide/rustup-home" "-w" "/opt/rustwide/workdir" "-m" "1610612736" "--user" "0:0" "--network" "none" "ghcr.io/rust-lang/crates-build-env/linux@sha256:d429b63d4308055ea97f60fb1d3dfca48854a00942f1bd2ad806beaf015945ec" "/opt/rustwide/cargo-home/bin/cargo" "+dec9417b8611e34e787a3e4c37686b5131f9e5c5" "test" "--frozen", kill_on_drop: false }` [INFO] [stdout] b7715e2b4235a6f5ecc8ec59e20d28380f7d9580eb337b45d2fb81e1e3adab27 [INFO] running `Command { std: "docker" "start" "-a" "b7715e2b4235a6f5ecc8ec59e20d28380f7d9580eb337b45d2fb81e1e3adab27", kill_on_drop: false }` [INFO] [stderr] warning: unused imports: `ast::TypstRoot` and `lexer::token_type::TypstTokenType` [INFO] [stderr] --> src/language/mod.rs:2:13 [INFO] [stderr] | [INFO] [stderr] 2 | use crate::{ast::TypstRoot, lexer::token_type::TypstTokenType}; [INFO] [stderr] | ^^^^^^^^^^^^^^ ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ [INFO] [stderr] | [INFO] [stderr] = note: `#[warn(unused_imports)]` (part of `#[warn(unused)]`) on by default [INFO] [stderr] [INFO] [stderr] warning: unused imports: `Source` and `UniversalElementRole` [INFO] [stderr] --> src/lexer/token_type.rs:1:16 [INFO] [stderr] | [INFO] [stderr] 1 | use oak_core::{Source, Token, TokenType, UniversalElementRole, UniversalTokenRole}; [INFO] [stderr] | ^^^^^^ ^^^^^^^^^^^^^^^^^^^^ [INFO] [stderr] [INFO] [stderr] warning: unused import: `Parser` [INFO] [stderr] --> src/parser/element_type.rs:1:29 [INFO] [stderr] | [INFO] [stderr] 1 | use oak_core::{ElementType, Parser, UniversalElementRole}; [INFO] [stderr] | ^^^^^^ [INFO] [stderr] [INFO] [stderr] warning: unused import: `GreenNode` [INFO] [stderr] --> src/parser/mod.rs:10:5 [INFO] [stderr] | [INFO] [stderr] 10 | GreenNode, OakError, [INFO] [stderr] | ^^^^^^^^^ [INFO] [stderr] [INFO] [stderr] warning: variable `count` is assigned to, but never used [INFO] [stderr] --> src/lexer/mod.rs:155:25 [INFO] [stderr] | [INFO] [stderr] 155 | let mut count = 0; [INFO] [stderr] | ^^^^^^^^^ [INFO] [stderr] | [INFO] [stderr] = note: consider using `_count` instead [INFO] [stderr] = note: `#[warn(unused_variables)]` (part of `#[warn(unused)]`) on by default [INFO] [stderr] [INFO] [stderr] warning: value assigned to `count` is never read [INFO] [stderr] --> src/lexer/mod.rs:157:25 [INFO] [stderr] | [INFO] [stderr] 157 | count += 1; [INFO] [stderr] | ^^^^^^^^^^ [INFO] [stderr] | [INFO] [stderr] = help: maybe it is overwritten before being read? [INFO] [stderr] = note: `#[warn(unused_assignments)]` (part of `#[warn(unused)]`) on by default [INFO] [stderr] [INFO] [stderr] warning: field `config` is never read [INFO] [stderr] --> src/lexer/mod.rs:22:5 [INFO] [stderr] | [INFO] [stderr] 21 | pub struct TypstLexer<'config> { [INFO] [stderr] | ---------- field in this struct [INFO] [stderr] 22 | config: &'config TypstLanguage, [INFO] [stderr] | ^^^^^^ [INFO] [stderr] | [INFO] [stderr] = note: `TypstLexer` has derived impls for the traits `Debug` and `Clone`, but these are intentionally ignored during dead code analysis [INFO] [stderr] = note: `#[warn(dead_code)]` (part of `#[warn(unused)]`) on by default [INFO] [stderr] [INFO] [stderr] warning: `oak-typst` (lib) generated 7 warnings (run `cargo fix --lib -p oak-typst` to apply 4 suggestions) [INFO] [stderr] warning: `oak-typst` (lib test) generated 7 warnings (7 duplicates) [INFO] [stderr] Finished `test` profile [unoptimized + debuginfo] target(s) in 0.11s [INFO] [stderr] Running unittests src/lib.rs (/opt/rustwide/target/debug/deps/oak_typst-e0addcf48e35b27d) [INFO] [stdout] [INFO] [stdout] running 0 tests [INFO] [stdout] [INFO] [stdout] test result: ok. 0 passed; 0 failed; 0 ignored; 0 measured; 0 filtered out; finished in 0.00s [INFO] [stdout] [INFO] [stderr] Running tests/main.rs (/opt/rustwide/target/debug/deps/main-9c2a5d74f30ba4bc) [INFO] [stdout] [INFO] [stdout] running 2 tests [INFO] [stdout] test ready ... ok [INFO] [stdout] test lexer::test_typst_lexer ... FAILED [INFO] [stdout] [INFO] [stdout] failures: [INFO] [stdout] [INFO] [stdout] ---- lexer::test_typst_lexer stdout ---- [INFO] [stdout] Testing file: /opt/rustwide/workdir/tests/lexer/basic.typ [INFO] [stdout] Error: TestFailure { path: "/opt/rustwide/workdir/tests/lexer/basic.typ", expected: "LexerTestExpected {\n success: true,\n count: 21,\n tokens: [\n TokenData {\n kind: \"Equal\",\n text: \"=\",\n start: 0,\n end: 1,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"Heading\",\n start: 2,\n end: 9,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"This\",\n start: 13,\n end: 17,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"is\",\n start: 18,\n end: 20,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"a\",\n start: 21,\n end: 22,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"paragraph\",\n start: 23,\n end: 32,\n },\n TokenData {\n kind: \"Dot\",\n text: \".\",\n start: 32,\n end: 33,\n },\n TokenData {\n kind: \"Star\",\n text: \"*\",\n start: 37,\n end: 38,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"Bold\",\n start: 38,\n end: 42,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"text\",\n start: 43,\n end: 47,\n },\n TokenData {\n kind: \"Star\",\n text: \"*\",\n start: 47,\n end: 48,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"and\",\n start: 49,\n end: 52,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"_italic\",\n start: 53,\n end: 60,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"text_\",\n start: 61,\n end: 66,\n },\n TokenData {\n kind: \"Dot\",\n text: \".\",\n start: 66,\n end: 67,\n },\n TokenData {\n kind: \"Hash\",\n text: \"#\",\n start: 71,\n end: 72,\n },\n TokenData {\n kind: \"Let\",\n text: \"let\",\n start: 72,\n end: 75,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"x\",\n start: 76,\n end: 77,\n },\n TokenData {\n kind: \"Equal\",\n text: \"=\",\n start: 78,\n end: 79,\n },\n TokenData {\n kind: \"NumericLiteral\",\n text: \"42\",\n start: 80,\n end: 82,\n },\n TokenData {\n kind: \"Eof\",\n text: \"\",\n start: 82,\n end: 82,\n },\n ],\n errors: [],\n}", actual: "LexerTestExpected {\n success: true,\n count: 565,\n tokens: [\n TokenData {\n kind: \"Hash\",\n text: \"#\",\n start: 37,\n end: 38,\n },\n TokenData {\n kind: \"Set\",\n text: \"set\",\n start: 38,\n end: 41,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"page\",\n start: 42,\n end: 46,\n },\n TokenData {\n kind: \"LeftParen\",\n text: \"(\",\n start: 46,\n end: 47,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"paper\",\n start: 53,\n end: 58,\n },\n TokenData {\n kind: \"Colon\",\n text: \":\",\n start: 58,\n end: 59,\n },\n TokenData {\n kind: \"StringLiteral\",\n text: \"\\\"a4\\\"\",\n start: 60,\n end: 64,\n },\n TokenData {\n kind: \"Comma\",\n text: \",\",\n start: 64,\n end: 65,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"margin\",\n start: 71,\n end: 77,\n },\n TokenData {\n kind: \"Colon\",\n text: \":\",\n start: 77,\n end: 78,\n },\n TokenData {\n kind: \"LeftParen\",\n text: \"(\",\n start: 79,\n end: 80,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"x\",\n start: 80,\n end: 81,\n },\n TokenData {\n kind: \"Colon\",\n text: \":\",\n start: 81,\n end: 82,\n },\n TokenData {\n kind: \"NumericLiteral\",\n text: \"2\",\n start: 83,\n end: 84,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"cm\",\n start: 84,\n end: 86,\n },\n TokenData {\n kind: \"Comma\",\n text: \",\",\n start: 86,\n end: 87,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"y\",\n start: 88,\n end: 89,\n },\n TokenData {\n kind: \"Colon\",\n text: \":\",\n start: 89,\n end: 90,\n },\n TokenData {\n kind: \"NumericLiteral\",\n text: \"2\",\n start: 91,\n end: 92,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"cm\",\n start: 92,\n end: 94,\n },\n TokenData {\n kind: \"RightParen\",\n text: \")\",\n start: 94,\n end: 95,\n },\n TokenData {\n kind: \"Comma\",\n text: \",\",\n start: 95,\n end: 96,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"numbering\",\n start: 102,\n end: 111,\n },\n TokenData {\n kind: \"Colon\",\n text: \":\",\n start: 111,\n end: 112,\n },\n TokenData {\n kind: \"StringLiteral\",\n text: \"\\\"1\\\"\",\n start: 113,\n end: 116,\n },\n TokenData {\n kind: \"Comma\",\n text: \",\",\n start: 116,\n end: 117,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"header\",\n start: 123,\n end: 129,\n },\n TokenData {\n kind: \"Colon\",\n text: \":\",\n start: 129,\n end: 130,\n },\n TokenData {\n kind: \"LeftBracket\",\n text: \"[\",\n start: 131,\n end: 132,\n },\n TokenData {\n kind: \"Hash\",\n text: \"#\",\n start: 142,\n end: 143,\n },\n TokenData {\n kind: \"Set\",\n text: \"set\",\n start: 143,\n end: 146,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"text\",\n start: 147,\n end: 151,\n },\n TokenData {\n kind: \"LeftParen\",\n text: \"(\",\n start: 151,\n end: 152,\n },\n TokenData {\n kind: \"NumericLiteral\",\n text: \"8\",\n start: 152,\n end: 153,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"pt\",\n start: 153,\n end: 155,\n },\n TokenData {\n kind: \"RightParen\",\n text: \")\",\n start: 155,\n end: 156,\n },\n TokenData {\n kind: \"Hash\",\n text: \"#\",\n start: 166,\n end: 167,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"smallcaps\",\n start: 167,\n end: 176,\n },\n TokenData {\n kind: \"LeftParen\",\n text: \"(\",\n start: 176,\n end: 177,\n },\n TokenData {\n kind: \"StringLiteral\",\n text: \"\\\"Typst Documentation\\\"\",\n start: 177,\n end: 198,\n },\n TokenData {\n kind: \"RightParen\",\n text: \")\",\n start: 198,\n end: 199,\n },\n TokenData {\n kind: \"Hash\",\n text: \"#\",\n start: 209,\n end: 210,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"h\",\n start: 210,\n end: 211,\n },\n TokenData {\n kind: \"LeftParen\",\n text: \"(\",\n start: 211,\n end: 212,\n },\n TokenData {\n kind: \"NumericLiteral\",\n text: \"1\",\n start: 212,\n end: 213,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"fr\",\n start: 213,\n end: 215,\n },\n TokenData {\n kind: \"RightParen\",\n text: \")\",\n start: 215,\n end: 216,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"Page\",\n start: 226,\n end: 230,\n },\n TokenData {\n kind: \"Hash\",\n text: \"#\",\n start: 231,\n end: 232,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"counter\",\n start: 232,\n end: 239,\n },\n TokenData {\n kind: \"LeftParen\",\n text: \"(\",\n start: 239,\n end: 240,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"page\",\n start: 240,\n end: 244,\n },\n TokenData {\n kind: \"RightParen\",\n text: \")\",\n start: 244,\n end: 245,\n },\n TokenData {\n kind: \"Dot\",\n text: \".\",\n start: 245,\n end: 246,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"display\",\n start: 246,\n end: 253,\n },\n TokenData {\n kind: \"LeftParen\",\n text: \"(\",\n start: 253,\n end: 254,\n },\n TokenData {\n kind: \"RightParen\",\n text: \")\",\n start: 254,\n end: 255,\n },\n TokenData {\n kind: \"RightBracket\",\n text: \"]\",\n start: 261,\n end: 262,\n },\n TokenData {\n kind: \"RightParen\",\n text: \")\",\n start: 264,\n end: 265,\n },\n TokenData {\n kind: \"Hash\",\n text: \"#\",\n start: 269,\n end: 270,\n },\n TokenData {\n kind: \"Set\",\n text: \"set\",\n start: 270,\n end: 273,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"text\",\n start: 274,\n end: 278,\n },\n TokenData {\n kind: \"LeftParen\",\n text: \"(\",\n start: 278,\n end: 279,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"font\",\n start: 285,\n end: 289,\n },\n TokenData {\n kind: \"Colon\",\n text: \":\",\n start: 289,\n end: 290,\n },\n TokenData {\n kind: \"StringLiteral\",\n text: \"\\\"New Computer Modern\\\"\",\n start: 291,\n end: 312,\n },\n TokenData {\n kind: \"Comma\",\n text: \",\",\n start: 312,\n end: 313,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"size\",\n start: 319,\n end: 323,\n },\n TokenData {\n kind: \"Colon\",\n text: \":\",\n start: 323,\n end: 324,\n },\n TokenData {\n kind: \"NumericLiteral\",\n text: \"11\",\n start: 325,\n end: 327,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"pt\",\n start: 327,\n end: 329,\n },\n TokenData {\n kind: \"Comma\",\n text: \",\",\n start: 329,\n end: 330,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"lang\",\n start: 336,\n end: 340,\n },\n TokenData {\n kind: \"Colon\",\n text: \":\",\n start: 340,\n end: 341,\n },\n TokenData {\n kind: \"StringLiteral\",\n text: \"\\\"en\\\"\",\n start: 342,\n end: 346,\n },\n TokenData {\n kind: \"RightParen\",\n text: \")\",\n start: 348,\n end: 349,\n },\n TokenData {\n kind: \"Hash\",\n text: \"#\",\n start: 353,\n end: 354,\n },\n TokenData {\n kind: \"Let\",\n text: \"let\",\n start: 354,\n end: 357,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"title\",\n start: 358,\n end: 363,\n },\n TokenData {\n kind: \"Equal\",\n text: \"=\",\n start: 364,\n end: 365,\n },\n TokenData {\n kind: \"StringLiteral\",\n text: \"\\\"Typst Document\\\"\",\n start: 366,\n end: 382,\n },\n TokenData {\n kind: \"Heading\",\n text: \"=\",\n start: 386,\n end: 387,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"Introduction\",\n start: 388,\n end: 400,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"This\",\n start: 402,\n end: 406,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"is\",\n start: 407,\n end: 409,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"a\",\n start: 410,\n end: 411,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"comprehensive\",\n start: 412,\n end: 425,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"test\",\n start: 426,\n end: 430,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"of\",\n start: 431,\n end: 433,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"Typst\",\n start: 434,\n end: 439,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"syntax\",\n start: 440,\n end: 446,\n },\n TokenData {\n kind: \"Dot\",\n text: \".\",\n start: 446,\n end: 447,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"We\",\n start: 449,\n end: 451,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"can\",\n start: 452,\n end: 455,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"use\",\n start: 456,\n end: 459,\n },\n TokenData {\n kind: \"Strong\",\n text: \"*\",\n start: 460,\n end: 461,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"bold\",\n start: 461,\n end: 465,\n },\n TokenData {\n kind: \"Strong\",\n text: \"*\",\n start: 465,\n end: 466,\n },\n TokenData {\n kind: \"Comma\",\n text: \",\",\n start: 466,\n end: 467,\n },\n TokenData {\n kind: \"Emphasis\",\n text: \"_\",\n start: 468,\n end: 469,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"italic\",\n start: 469,\n end: 475,\n },\n TokenData {\n kind: \"Emphasis\",\n text: \"_\",\n start: 475,\n end: 476,\n },\n TokenData {\n kind: \"Comma\",\n text: \",\",\n start: 476,\n end: 477,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"and\",\n start: 478,\n end: 481,\n },\n TokenData {\n kind: \"Backtick\",\n text: \"`\",\n start: 482,\n end: 483,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"monospaced\",\n start: 483,\n end: 493,\n },\n TokenData {\n kind: \"Backtick\",\n text: \"`\",\n start: 493,\n end: 494,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"text\",\n start: 495,\n end: 499,\n },\n TokenData {\n kind: \"Dot\",\n text: \".\",\n start: 499,\n end: 500,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"Escaped\",\n start: 502,\n end: 509,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"symbols\",\n start: 510,\n end: 517,\n },\n TokenData {\n kind: \"Colon\",\n text: \":\",\n start: 517,\n end: 518,\n },\n TokenData {\n kind: \"Hash\",\n text: \"#\",\n start: 520,\n end: 521,\n },\n TokenData {\n kind: \"Dollar\",\n text: \"$\",\n start: 523,\n end: 524,\n },\n TokenData {\n kind: \"Percent\",\n text: \"%\",\n start: 526,\n end: 527,\n },\n TokenData {\n kind: \"Dot\",\n text: \".\",\n start: 527,\n end: 528,\n },\n TokenData {\n kind: \"Heading\",\n text: \"==\",\n start: 532,\n end: 534,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"Math\",\n start: 535,\n end: 539,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"Mode\",\n start: 540,\n end: 544,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"Inline\",\n start: 546,\n end: 552,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"math\",\n start: 553,\n end: 557,\n },\n TokenData {\n kind: \"Colon\",\n text: \":\",\n start: 557,\n end: 558,\n },\n TokenData {\n kind: \"Dollar\",\n text: \"$\",\n start: 559,\n end: 560,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"E\",\n start: 560,\n end: 561,\n },\n TokenData {\n kind: \"Equal\",\n text: \"=\",\n start: 562,\n end: 563,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"mc\",\n start: 564,\n end: 566,\n },\n TokenData {\n kind: \"Text\",\n text: \"^2\",\n start: 566,\n end: 568,\n },\n TokenData {\n kind: \"Dollar\",\n text: \"$\",\n start: 568,\n end: 569,\n },\n TokenData {\n kind: \"Dot\",\n text: \".\",\n start: 569,\n end: 570,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"Display\",\n start: 572,\n end: 579,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"math\",\n start: 580,\n end: 584,\n },\n TokenData {\n kind: \"Colon\",\n text: \":\",\n start: 584,\n end: 585,\n },\n TokenData {\n kind: \"Dollar\",\n text: \"$\",\n start: 587,\n end: 588,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"sum\",\n start: 594,\n end: 597,\n },\n TokenData {\n kind: \"Emphasis\",\n text: \"_\",\n start: 597,\n end: 598,\n },\n TokenData {\n kind: \"LeftParen\",\n text: \"(\",\n start: 598,\n end: 599,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"i\",\n start: 599,\n end: 600,\n },\n TokenData {\n kind: \"Equal\",\n text: \"=\",\n start: 600,\n end: 601,\n },\n TokenData {\n kind: \"NumericLiteral\",\n text: \"0\",\n start: 601,\n end: 602,\n },\n TokenData {\n kind: \"RightParen\",\n text: \")\",\n start: 602,\n end: 603,\n },\n TokenData {\n kind: \"Text\",\n text: \"^n\",\n start: 603,\n end: 605,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"i\",\n start: 606,\n end: 607,\n },\n TokenData {\n kind: \"Equal\",\n text: \"=\",\n start: 608,\n end: 609,\n },\n TokenData {\n kind: \"LeftParen\",\n text: \"(\",\n start: 610,\n end: 611,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"n\",\n start: 611,\n end: 612,\n },\n TokenData {\n kind: \"LeftParen\",\n text: \"(\",\n start: 612,\n end: 613,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"n\",\n start: 613,\n end: 614,\n },\n TokenData {\n kind: \"Plus\",\n text: \"+\",\n start: 614,\n end: 615,\n },\n TokenData {\n kind: \"NumericLiteral\",\n text: \"1\",\n start: 615,\n end: 616,\n },\n TokenData {\n kind: \"RightParen\",\n text: \")\",\n start: 616,\n end: 617,\n },\n TokenData {\n kind: \"RightParen\",\n text: \")\",\n start: 617,\n end: 618,\n },\n TokenData {\n kind: \"Slash\",\n text: \"/\",\n start: 619,\n end: 620,\n },\n TokenData {\n kind: \"NumericLiteral\",\n text: \"2\",\n start: 621,\n end: 622,\n },\n TokenData {\n kind: \"Dollar\",\n text: \"$\",\n start: 624,\n end: 625,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"Complex\",\n start: 629,\n end: 636,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"Math\",\n start: 637,\n end: 641,\n },\n TokenData {\n kind: \"Colon\",\n text: \":\",\n start: 641,\n end: 642,\n },\n TokenData {\n kind: \"Dollar\",\n text: \"$\",\n start: 644,\n end: 645,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"cal\",\n start: 651,\n end: 654,\n },\n TokenData {\n kind: \"LeftParen\",\n text: \"(\",\n start: 654,\n end: 655,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"A\",\n start: 655,\n end: 656,\n },\n TokenData {\n kind: \"RightParen\",\n text: \")\",\n start: 656,\n end: 657,\n },\n TokenData {\n kind: \"Colon\",\n text: \":\",\n start: 658,\n end: 659,\n },\n TokenData {\n kind: \"Equal\",\n text: \"=\",\n start: 659,\n end: 660,\n },\n TokenData {\n kind: \"LeftBrace\",\n text: \"{\",\n start: 661,\n end: 662,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"x\",\n start: 663,\n end: 664,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"in\",\n start: 665,\n end: 667,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"RR\",\n start: 668,\n end: 670,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"x\",\n start: 673,\n end: 674,\n },\n TokenData {\n kind: \"StringLiteral\",\n text: \"\\\"is natural\\\"\",\n start: 675,\n end: 687,\n },\n TokenData {\n kind: \"RightBrace\",\n text: \"}\",\n start: 688,\n end: 689,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"mat\",\n start: 695,\n end: 698,\n },\n TokenData {\n kind: \"LeftParen\",\n text: \"(\",\n start: 698,\n end: 699,\n },\n TokenData {\n kind: \"NumericLiteral\",\n text: \"1\",\n start: 699,\n end: 700,\n },\n TokenData {\n kind: \"Comma\",\n text: \",\",\n start: 700,\n end: 701,\n },\n TokenData {\n kind: \"NumericLiteral\",\n text: \"2\",\n start: 702,\n end: 703,\n },\n TokenData {\n kind: \"Semicolon\",\n text: \";\",\n start: 703,\n end: 704,\n },\n TokenData {\n kind: \"NumericLiteral\",\n text: \"3\",\n start: 705,\n end: 706,\n },\n TokenData {\n kind: \"Comma\",\n text: \",\",\n start: 706,\n end: 707,\n },\n TokenData {\n kind: \"NumericLiteral\",\n text: \"4\",\n start: 708,\n end: 709,\n },\n TokenData {\n kind: \"RightParen\",\n text: \")\",\n start: 709,\n end: 710,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"vec\",\n start: 716,\n end: 719,\n },\n TokenData {\n kind: \"LeftParen\",\n text: \"(\",\n start: 719,\n end: 720,\n },\n TokenData {\n kind: \"NumericLiteral\",\n text: \"1\",\n start: 720,\n end: 721,\n },\n TokenData {\n kind: \"Comma\",\n text: \",\",\n start: 721,\n end: 722,\n },\n TokenData {\n kind: \"NumericLiteral\",\n text: \"2\",\n start: 723,\n end: 724,\n },\n TokenData {\n kind: \"RightParen\",\n text: \")\",\n start: 724,\n end: 725,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"cases\",\n start: 731,\n end: 736,\n },\n TokenData {\n kind: \"LeftParen\",\n text: \"(\",\n start: 736,\n end: 737,\n },\n TokenData {\n kind: \"NumericLiteral\",\n text: \"1\",\n start: 747,\n end: 748,\n },\n TokenData {\n kind: \"StringLiteral\",\n text: \"\\\"if\\\"\",\n start: 749,\n end: 753,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"x\",\n start: 754,\n end: 755,\n },\n TokenData {\n kind: \"Greater\",\n text: \">\",\n start: 756,\n end: 757,\n },\n TokenData {\n kind: \"NumericLiteral\",\n text: \"0\",\n start: 758,\n end: 759,\n },\n TokenData {\n kind: \"Comma\",\n text: \",\",\n start: 759,\n end: 760,\n },\n TokenData {\n kind: \"NumericLiteral\",\n text: \"0\",\n start: 770,\n end: 771,\n },\n TokenData {\n kind: \"StringLiteral\",\n text: \"\\\"otherwise\\\"\",\n start: 772,\n end: 783,\n },\n TokenData {\n kind: \"RightParen\",\n text: \")\",\n start: 789,\n end: 790,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"integral\",\n start: 796,\n end: 804,\n },\n TokenData {\n kind: \"Emphasis\",\n text: \"_\",\n start: 804,\n end: 805,\n },\n TokenData {\n kind: \"NumericLiteral\",\n text: \"0\",\n start: 805,\n end: 806,\n },\n TokenData {\n kind: \"Text\",\n text: \"^oo\",\n start: 806,\n end: 809,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"e\",\n start: 810,\n end: 811,\n },\n TokenData {\n kind: \"Text\",\n text: \"^\",\n start: 811,\n end: 812,\n },\n TokenData {\n kind: \"LeftParen\",\n text: \"(\",\n start: 812,\n end: 813,\n },\n TokenData {\n kind: \"Minus\",\n text: \"-\",\n start: 813,\n end: 814,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"x\",\n start: 814,\n end: 815,\n },\n TokenData {\n kind: \"RightParen\",\n text: \")\",\n start: 815,\n end: 816,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"dx\",\n start: 817,\n end: 819,\n },\n TokenData {\n kind: \"Equal\",\n text: \"=\",\n start: 820,\n end: 821,\n },\n TokenData {\n kind: \"NumericLiteral\",\n text: \"1\",\n start: 822,\n end: 823,\n },\n TokenData {\n kind: \"Dollar\",\n text: \"$\",\n start: 825,\n end: 826,\n },\n TokenData {\n kind: \"Heading\",\n text: \"==\",\n start: 830,\n end: 832,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"Lists\",\n start: 833,\n end: 838,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"and\",\n start: 839,\n end: 842,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"Enumerations\",\n start: 843,\n end: 855,\n },\n TokenData {\n kind: \"ListItem\",\n text: \"-\",\n start: 857,\n end: 858,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"Unordered\",\n start: 859,\n end: 868,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"item\",\n start: 869,\n end: 873,\n },\n TokenData {\n kind: \"NumericLiteral\",\n text: \"1\",\n start: 874,\n end: 875,\n },\n TokenData {\n kind: \"ListItem\",\n text: \"-\",\n start: 877,\n end: 878,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"Unordered\",\n start: 879,\n end: 888,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"item\",\n start: 889,\n end: 893,\n },\n TokenData {\n kind: \"NumericLiteral\",\n text: \"2\",\n start: 894,\n end: 895,\n },\n TokenData {\n kind: \"Minus\",\n text: \"-\",\n start: 901,\n end: 902,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"Nested\",\n start: 903,\n end: 909,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"item\",\n start: 910,\n end: 914,\n },\n TokenData {\n kind: \"Minus\",\n text: \"-\",\n start: 924,\n end: 925,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"Deeply\",\n start: 926,\n end: 932,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"nested\",\n start: 933,\n end: 939,\n },\n TokenData {\n kind: \"ListItem\",\n text: \"+\",\n start: 943,\n end: 944,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"Ordered\",\n start: 945,\n end: 952,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"item\",\n start: 953,\n end: 957,\n },\n TokenData {\n kind: \"NumericLiteral\",\n text: \"1\",\n start: 958,\n end: 959,\n },\n TokenData {\n kind: \"ListItem\",\n text: \"+\",\n start: 961,\n end: 962,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"Ordered\",\n start: 963,\n end: 970,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"item\",\n start: 971,\n end: 975,\n },\n TokenData {\n kind: \"NumericLiteral\",\n text: \"2\",\n start: 976,\n end: 977,\n },\n TokenData {\n kind: \"Plus\",\n text: \"+\",\n start: 983,\n end: 984,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"Nested\",\n start: 985,\n end: 991,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"ordered\",\n start: 992,\n end: 999,\n },\n TokenData {\n kind: \"Slash\",\n text: \"/\",\n start: 1003,\n end: 1004,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"Term\",\n start: 1005,\n end: 1009,\n },\n TokenData {\n kind: \"NumericLiteral\",\n text: \"1\",\n start: 1010,\n end: 1011,\n },\n TokenData {\n kind: \"Colon\",\n text: \":\",\n start: 1011,\n end: 1012,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"Definition\",\n start: 1013,\n end: 1023,\n },\n TokenData {\n kind: \"NumericLiteral\",\n text: \"1\",\n start: 1024,\n end: 1025,\n },\n TokenData {\n kind: \"Slash\",\n text: \"/\",\n start: 1027,\n end: 1028,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"Term\",\n start: 1029,\n end: 1033,\n },\n TokenData {\n kind: \"NumericLiteral\",\n text: \"2\",\n start: 1034,\n end: 1035,\n },\n TokenData {\n kind: \"Colon\",\n text: \":\",\n start: 1035,\n end: 1036,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"Definition\",\n start: 1037,\n end: 1047,\n },\n TokenData {\n kind: \"NumericLiteral\",\n text: \"2\",\n start: 1048,\n end: 1049,\n },\n TokenData {\n kind: \"LeftParen\",\n text: \"(\",\n start: 1050,\n end: 1051,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"long\",\n start: 1051,\n end: 1055,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"definition\",\n start: 1056,\n end: 1066,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"that\",\n start: 1067,\n end: 1071,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"spans\",\n start: 1075,\n end: 1080,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"multiple\",\n start: 1081,\n end: 1089,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"lines\",\n start: 1090,\n end: 1095,\n },\n TokenData {\n kind: \"RightParen\",\n text: \")\",\n start: 1095,\n end: 1096,\n },\n TokenData {\n kind: \"Heading\",\n text: \"==\",\n start: 1100,\n end: 1102,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"Functions\",\n start: 1103,\n end: 1112,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"and\",\n start: 1113,\n end: 1116,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"Scripting\",\n start: 1117,\n end: 1126,\n },\n TokenData {\n kind: \"Hash\",\n text: \"#\",\n start: 1128,\n end: 1129,\n },\n TokenData {\n kind: \"Let\",\n text: \"let\",\n start: 1129,\n end: 1132,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"add\",\n start: 1133,\n end: 1136,\n },\n TokenData {\n kind: \"LeftParen\",\n text: \"(\",\n start: 1136,\n end: 1137,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"x\",\n start: 1137,\n end: 1138,\n },\n TokenData {\n kind: \"Comma\",\n text: \",\",\n start: 1138,\n end: 1139,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"y\",\n start: 1140,\n end: 1141,\n },\n TokenData {\n kind: \"RightParen\",\n text: \")\",\n start: 1141,\n end: 1142,\n },\n TokenData {\n kind: \"Equal\",\n text: \"=\",\n start: 1143,\n end: 1144,\n },\n TokenData {\n kind: \"LeftBrace\",\n text: \"{\",\n start: 1145,\n end: 1146,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"x\",\n start: 1152,\n end: 1153,\n },\n TokenData {\n kind: \"Plus\",\n text: \"+\",\n start: 1154,\n end: 1155,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"y\",\n start: 1156,\n end: 1157,\n },\n TokenData {\n kind: \"RightBrace\",\n text: \"}\",\n start: 1159,\n end: 1160,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"The\",\n start: 1164,\n end: 1167,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"sum\",\n start: 1168,\n end: 1171,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"of\",\n start: 1172,\n end: 1174,\n },\n TokenData {\n kind: \"NumericLiteral\",\n text: \"1\",\n start: 1175,\n end: 1176,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"and\",\n start: 1177,\n end: 1180,\n },\n TokenData {\n kind: \"NumericLiteral\",\n text: \"2\",\n start: 1181,\n end: 1182,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"is\",\n start: 1183,\n end: 1185,\n },\n TokenData {\n kind: \"Hash\",\n text: \"#\",\n start: 1186,\n end: 1187,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"add\",\n start: 1187,\n end: 1190,\n },\n TokenData {\n kind: \"LeftParen\",\n text: \"(\",\n start: 1190,\n end: 1191,\n },\n TokenData {\n kind: \"NumericLiteral\",\n text: \"1\",\n start: 1191,\n end: 1192,\n },\n TokenData {\n kind: \"Comma\",\n text: \",\",\n start: 1192,\n end: 1193,\n },\n TokenData {\n kind: \"NumericLiteral\",\n text: \"2\",\n start: 1194,\n end: 1195,\n },\n TokenData {\n kind: \"RightParen\",\n text: \")\",\n start: 1195,\n end: 1196,\n },\n TokenData {\n kind: \"Dot\",\n text: \".\",\n start: 1196,\n end: 1197,\n },\n TokenData {\n kind: \"Hash\",\n text: \"#\",\n start: 1201,\n end: 1202,\n },\n TokenData {\n kind: \"Let\",\n text: \"let\",\n start: 1202,\n end: 1205,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"values\",\n start: 1206,\n end: 1212,\n },\n TokenData {\n kind: \"Equal\",\n text: \"=\",\n start: 1213,\n end: 1214,\n },\n TokenData {\n kind: \"LeftParen\",\n text: \"(\",\n start: 1215,\n end: 1216,\n },\n TokenData {\n kind: \"NumericLiteral\",\n text: \"1\",\n start: 1216,\n end: 1217,\n },\n TokenData {\n kind: \"Comma\",\n text: \",\",\n start: 1217,\n end: 1218,\n },\n TokenData {\n kind: \"NumericLiteral\",\n text: \"2\",\n start: 1219,\n end: 1220,\n },\n TokenData {\n kind: \"Comma\",\n text: \",\",\n start: 1220,\n end: 1221,\n },\n TokenData {\n kind: \"NumericLiteral\",\n text: \"3\",\n start: 1222,\n end: 1223,\n },\n TokenData {\n kind: \"Comma\",\n text: \",\",\n start: 1223,\n end: 1224,\n },\n TokenData {\n kind: \"NumericLiteral\",\n text: \"4\",\n start: 1225,\n end: 1226,\n },\n TokenData {\n kind: \"Comma\",\n text: \",\",\n start: 1226,\n end: 1227,\n },\n TokenData {\n kind: \"NumericLiteral\",\n text: \"5\",\n start: 1228,\n end: 1229,\n },\n TokenData {\n kind: \"RightParen\",\n text: \")\",\n start: 1229,\n end: 1230,\n },\n TokenData {\n kind: \"Hash\",\n text: \"#\",\n start: 1232,\n end: 1233,\n },\n TokenData {\n kind: \"For\",\n text: \"for\",\n start: 1233,\n end: 1236,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"v\",\n start: 1237,\n end: 1238,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"in\",\n start: 1239,\n end: 1241,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"values\",\n start: 1242,\n end: 1248,\n },\n TokenData {\n kind: \"LeftBracket\",\n text: \"[\",\n start: 1249,\n end: 1250,\n },\n TokenData {\n kind: \"Minus\",\n text: \"-\",\n start: 1256,\n end: 1257,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"Value\",\n start: 1258,\n end: 1263,\n },\n TokenData {\n kind: \"Colon\",\n text: \":\",\n start: 1263,\n end: 1264,\n },\n TokenData {\n kind: \"Hash\",\n text: \"#\",\n start: 1265,\n end: 1266,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"v\",\n start: 1266,\n end: 1267,\n },\n TokenData {\n kind: \"RightBracket\",\n text: \"]\",\n start: 1269,\n end: 1270,\n },\n TokenData {\n kind: \"Hash\",\n text: \"#\",\n start: 1274,\n end: 1275,\n },\n TokenData {\n kind: \"Let\",\n text: \"let\",\n start: 1275,\n end: 1278,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"dict\",\n start: 1279,\n end: 1283,\n },\n TokenData {\n kind: \"Equal\",\n text: \"=\",\n start: 1284,\n end: 1285,\n },\n TokenData {\n kind: \"LeftParen\",\n text: \"(\",\n start: 1286,\n end: 1287,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"name\",\n start: 1287,\n end: 1291,\n },\n TokenData {\n kind: \"Colon\",\n text: \":\",\n start: 1291,\n end: 1292,\n },\n TokenData {\n kind: \"StringLiteral\",\n text: \"\\\"Typst\\\"\",\n start: 1293,\n end: 1300,\n },\n TokenData {\n kind: \"Comma\",\n text: \",\",\n start: 1300,\n end: 1301,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"version\",\n start: 1302,\n end: 1309,\n },\n TokenData {\n kind: \"Colon\",\n text: \":\",\n start: 1309,\n end: 1310,\n },\n TokenData {\n kind: \"NumericLiteral\",\n text: \"0.1\",\n start: 1311,\n end: 1314,\n },\n TokenData {\n kind: \"RightParen\",\n text: \")\",\n start: 1314,\n end: 1315,\n },\n TokenData {\n kind: \"Hash\",\n text: \"#\",\n start: 1317,\n end: 1318,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"dict\",\n start: 1318,\n end: 1322,\n },\n TokenData {\n kind: \"Dot\",\n text: \".\",\n start: 1322,\n end: 1323,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"name\",\n start: 1323,\n end: 1327,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"is\",\n start: 1328,\n end: 1330,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"version\",\n start: 1331,\n end: 1338,\n },\n TokenData {\n kind: \"Hash\",\n text: \"#\",\n start: 1339,\n end: 1340,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"dict\",\n start: 1340,\n end: 1344,\n },\n TokenData {\n kind: \"Dot\",\n text: \".\",\n start: 1344,\n end: 1345,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"version\",\n start: 1345,\n end: 1352,\n },\n TokenData {\n kind: \"Hash\",\n text: \"#\",\n start: 1356,\n end: 1357,\n },\n TokenData {\n kind: \"If\",\n text: \"if\",\n start: 1357,\n end: 1359,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"values\",\n start: 1360,\n end: 1366,\n },\n TokenData {\n kind: \"Dot\",\n text: \".\",\n start: 1366,\n end: 1367,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"len\",\n start: 1367,\n end: 1370,\n },\n TokenData {\n kind: \"LeftParen\",\n text: \"(\",\n start: 1370,\n end: 1371,\n },\n TokenData {\n kind: \"RightParen\",\n text: \")\",\n start: 1371,\n end: 1372,\n },\n TokenData {\n kind: \"Greater\",\n text: \">\",\n start: 1373,\n end: 1374,\n },\n TokenData {\n kind: \"NumericLiteral\",\n text: \"0\",\n start: 1375,\n end: 1376,\n },\n TokenData {\n kind: \"LeftBracket\",\n text: \"[\",\n start: 1377,\n end: 1378,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"The\",\n start: 1384,\n end: 1387,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"list\",\n start: 1388,\n end: 1392,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"is\",\n start: 1393,\n end: 1395,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"not\",\n start: 1396,\n end: 1399,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"empty\",\n start: 1400,\n end: 1405,\n },\n TokenData {\n kind: \"Dot\",\n text: \".\",\n start: 1405,\n end: 1406,\n },\n TokenData {\n kind: \"RightBracket\",\n text: \"]\",\n start: 1408,\n end: 1409,\n },\n TokenData {\n kind: \"Else\",\n text: \"else\",\n start: 1410,\n end: 1414,\n },\n TokenData {\n kind: \"LeftBracket\",\n text: \"[\",\n start: 1415,\n end: 1416,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"The\",\n start: 1422,\n end: 1425,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"list\",\n start: 1426,\n end: 1430,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"is\",\n start: 1431,\n end: 1433,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"empty\",\n start: 1434,\n end: 1439,\n },\n TokenData {\n kind: \"Dot\",\n text: \".\",\n start: 1439,\n end: 1440,\n },\n TokenData {\n kind: \"RightBracket\",\n text: \"]\",\n start: 1442,\n end: 1443,\n },\n TokenData {\n kind: \"Hash\",\n text: \"#\",\n start: 1447,\n end: 1448,\n },\n TokenData {\n kind: \"Show\",\n text: \"show\",\n start: 1448,\n end: 1452,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"heading\",\n start: 1453,\n end: 1460,\n },\n TokenData {\n kind: \"Colon\",\n text: \":\",\n start: 1460,\n end: 1461,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"it\",\n start: 1462,\n end: 1464,\n },\n TokenData {\n kind: \"Equal\",\n text: \"=\",\n start: 1465,\n end: 1466,\n },\n TokenData {\n kind: \"Greater\",\n text: \">\",\n start: 1466,\n end: 1467,\n },\n TokenData {\n kind: \"LeftBracket\",\n text: \"[\",\n start: 1468,\n end: 1469,\n },\n TokenData {\n kind: \"Hash\",\n text: \"#\",\n start: 1475,\n end: 1476,\n },\n TokenData {\n kind: \"Set\",\n text: \"set\",\n start: 1476,\n end: 1479,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"text\",\n start: 1480,\n end: 1484,\n },\n TokenData {\n kind: \"LeftParen\",\n text: \"(\",\n start: 1484,\n end: 1485,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"fill\",\n start: 1485,\n end: 1489,\n },\n TokenData {\n kind: \"Colon\",\n text: \":\",\n start: 1489,\n end: 1490,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"blue\",\n start: 1491,\n end: 1495,\n },\n TokenData {\n kind: \"RightParen\",\n text: \")\",\n start: 1495,\n end: 1496,\n },\n TokenData {\n kind: \"Hash\",\n text: \"#\",\n start: 1502,\n end: 1503,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"block\",\n start: 1503,\n end: 1508,\n },\n TokenData {\n kind: \"LeftParen\",\n text: \"(\",\n start: 1508,\n end: 1509,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"below\",\n start: 1509,\n end: 1514,\n },\n TokenData {\n kind: \"Colon\",\n text: \":\",\n start: 1514,\n end: 1515,\n },\n TokenData {\n kind: \"NumericLiteral\",\n text: \"1e\",\n start: 1516,\n end: 1518,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"m\",\n start: 1518,\n end: 1519,\n },\n TokenData {\n kind: \"RightParen\",\n text: \")\",\n start: 1519,\n end: 1520,\n },\n TokenData {\n kind: \"LeftBracket\",\n text: \"[\",\n start: 1520,\n end: 1521,\n },\n TokenData {\n kind: \"Hash\",\n text: \"#\",\n start: 1531,\n end: 1532,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"it\",\n start: 1532,\n end: 1534,\n },\n TokenData {\n kind: \"RightBracket\",\n text: \"]\",\n start: 1540,\n end: 1541,\n },\n TokenData {\n kind: \"RightBracket\",\n text: \"]\",\n start: 1543,\n end: 1544,\n },\n TokenData {\n kind: \"Heading\",\n text: \"==\",\n start: 1548,\n end: 1550,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"Figures\",\n start: 1551,\n end: 1558,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"and\",\n start: 1559,\n end: 1562,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"Images\",\n start: 1563,\n end: 1569,\n },\n TokenData {\n kind: \"Hash\",\n text: \"#\",\n start: 1571,\n end: 1572,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"figure\",\n start: 1572,\n end: 1578,\n },\n TokenData {\n kind: \"LeftParen\",\n text: \"(\",\n start: 1578,\n end: 1579,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"image\",\n start: 1585,\n end: 1590,\n },\n TokenData {\n kind: \"LeftParen\",\n text: \"(\",\n start: 1590,\n end: 1591,\n },\n TokenData {\n kind: \"StringLiteral\",\n text: \"\\\"test.png\\\"\",\n start: 1591,\n end: 1601,\n },\n TokenData {\n kind: \"Comma\",\n text: \",\",\n start: 1601,\n end: 1602,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"width\",\n start: 1603,\n end: 1608,\n },\n TokenData {\n kind: \"Colon\",\n text: \":\",\n start: 1608,\n end: 1609,\n },\n TokenData {\n kind: \"NumericLiteral\",\n text: \"80\",\n start: 1610,\n end: 1612,\n },\n TokenData {\n kind: \"Percent\",\n text: \"%\",\n start: 1612,\n end: 1613,\n },\n TokenData {\n kind: \"RightParen\",\n text: \")\",\n start: 1613,\n end: 1614,\n },\n TokenData {\n kind: \"Comma\",\n text: \",\",\n start: 1614,\n end: 1615,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"caption\",\n start: 1621,\n end: 1628,\n },\n TokenData {\n kind: \"Colon\",\n text: \":\",\n start: 1628,\n end: 1629,\n },\n TokenData {\n kind: \"LeftBracket\",\n text: \"[\",\n start: 1630,\n end: 1631,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"A\",\n start: 1631,\n end: 1632,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"test\",\n start: 1633,\n end: 1637,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"image\",\n start: 1638,\n end: 1643,\n },\n TokenData {\n kind: \"RightBracket\",\n text: \"]\",\n start: 1643,\n end: 1644,\n },\n TokenData {\n kind: \"Comma\",\n text: \",\",\n start: 1644,\n end: 1645,\n },\n TokenData {\n kind: \"RightParen\",\n text: \")\",\n start: 1647,\n end: 1648,\n },\n TokenData {\n kind: \"Heading\",\n text: \"==\",\n start: 1652,\n end: 1654,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"References\",\n start: 1655,\n end: 1665,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"See\",\n start: 1667,\n end: 1670,\n },\n TokenData {\n kind: \"At\",\n text: \"@\",\n start: 1671,\n end: 1672,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"intro\",\n start: 1672,\n end: 1677,\n },\n TokenData {\n kind: \"For\",\n text: \"for\",\n start: 1678,\n end: 1681,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"more\",\n start: 1682,\n end: 1686,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"details\",\n start: 1687,\n end: 1694,\n },\n TokenData {\n kind: \"Dot\",\n text: \".\",\n start: 1694,\n end: 1695,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"See\",\n start: 1697,\n end: 1700,\n },\n TokenData {\n kind: \"At\",\n text: \"@\",\n start: 1701,\n end: 1702,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"math\",\n start: 1702,\n end: 1706,\n },\n TokenData {\n kind: \"Minus\",\n text: \"-\",\n start: 1706,\n end: 1707,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"section\",\n start: 1707,\n end: 1714,\n },\n TokenData {\n kind: \"Dot\",\n text: \".\",\n start: 1714,\n end: 1715,\n },\n TokenData {\n kind: \"Heading\",\n text: \"==\",\n start: 1719,\n end: 1721,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"Code\",\n start: 1722,\n end: 1726,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"Blocks\",\n start: 1727,\n end: 1733,\n },\n TokenData {\n kind: \"Backtick\",\n text: \"`\",\n start: 1735,\n end: 1736,\n },\n TokenData {\n kind: \"Backtick\",\n text: \"`\",\n start: 1736,\n end: 1737,\n },\n TokenData {\n kind: \"Backtick\",\n text: \"`\",\n start: 1737,\n end: 1738,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"rust\",\n start: 1738,\n end: 1742,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"fn\",\n start: 1744,\n end: 1746,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"main\",\n start: 1747,\n end: 1751,\n },\n TokenData {\n kind: \"LeftParen\",\n text: \"(\",\n start: 1751,\n end: 1752,\n },\n TokenData {\n kind: \"RightParen\",\n text: \")\",\n start: 1752,\n end: 1753,\n },\n TokenData {\n kind: \"LeftBrace\",\n text: \"{\",\n start: 1754,\n end: 1755,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"println\",\n start: 1761,\n end: 1768,\n },\n TokenData {\n kind: \"Not\",\n text: \"!\",\n start: 1768,\n end: 1769,\n },\n TokenData {\n kind: \"LeftParen\",\n text: \"(\",\n start: 1769,\n end: 1770,\n },\n TokenData {\n kind: \"StringLiteral\",\n text: \"\\\"Hello from Rust code block!\\\"\",\n start: 1770,\n end: 1799,\n },\n TokenData {\n kind: \"RightParen\",\n text: \")\",\n start: 1799,\n end: 1800,\n },\n TokenData {\n kind: \"Semicolon\",\n text: \";\",\n start: 1800,\n end: 1801,\n },\n TokenData {\n kind: \"RightBrace\",\n text: \"}\",\n start: 1803,\n end: 1804,\n },\n TokenData {\n kind: \"Backtick\",\n text: \"`\",\n start: 1806,\n end: 1807,\n },\n TokenData {\n kind: \"Backtick\",\n text: \"`\",\n start: 1807,\n end: 1808,\n },\n TokenData {\n kind: \"Backtick\",\n text: \"`\",\n start: 1808,\n end: 1809,\n },\n TokenData {\n kind: \"Backtick\",\n text: \"`\",\n start: 1813,\n end: 1814,\n },\n TokenData {\n kind: \"Backtick\",\n text: \"`\",\n start: 1814,\n end: 1815,\n },\n TokenData {\n kind: \"Backtick\",\n text: \"`\",\n start: 1815,\n end: 1816,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"python\",\n start: 1816,\n end: 1822,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"def\",\n start: 1824,\n end: 1827,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"hello\",\n start: 1828,\n end: 1833,\n },\n TokenData {\n kind: \"LeftParen\",\n text: \"(\",\n start: 1833,\n end: 1834,\n },\n TokenData {\n kind: \"RightParen\",\n text: \")\",\n start: 1834,\n end: 1835,\n },\n TokenData {\n kind: \"Colon\",\n text: \":\",\n start: 1835,\n end: 1836,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"print\",\n start: 1842,\n end: 1847,\n },\n TokenData {\n kind: \"LeftParen\",\n text: \"(\",\n start: 1847,\n end: 1848,\n },\n TokenData {\n kind: \"StringLiteral\",\n text: \"\\\"Hello from Python\\\"\",\n start: 1848,\n end: 1867,\n },\n TokenData {\n kind: \"RightParen\",\n text: \")\",\n start: 1867,\n end: 1868,\n },\n TokenData {\n kind: \"Backtick\",\n text: \"`\",\n start: 1870,\n end: 1871,\n },\n TokenData {\n kind: \"Backtick\",\n text: \"`\",\n start: 1871,\n end: 1872,\n },\n TokenData {\n kind: \"Backtick\",\n text: \"`\",\n start: 1872,\n end: 1873,\n },\n TokenData {\n kind: \"Heading\",\n text: \"==\",\n start: 1877,\n end: 1879,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"Comments\",\n start: 1880,\n end: 1888,\n },\n TokenData {\n kind: \"Heading\",\n text: \"==\",\n start: 1994,\n end: 1996,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"Labels\",\n start: 1997,\n end: 2003,\n },\n TokenData {\n kind: \"Less\",\n text: \"<\",\n start: 2005,\n end: 2006,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"intro\",\n start: 2006,\n end: 2011,\n },\n TokenData {\n kind: \"Greater\",\n text: \">\",\n start: 2011,\n end: 2012,\n },\n TokenData {\n kind: \"Less\",\n text: \"<\",\n start: 2014,\n end: 2015,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"math\",\n start: 2015,\n end: 2019,\n },\n TokenData {\n kind: \"Minus\",\n text: \"-\",\n start: 2019,\n end: 2020,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"section\",\n start: 2020,\n end: 2027,\n },\n TokenData {\n kind: \"Greater\",\n text: \">\",\n start: 2027,\n end: 2028,\n },\n TokenData {\n kind: \"Heading\",\n text: \"==\",\n start: 2032,\n end: 2034,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"String\",\n start: 2035,\n end: 2041,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"Operations\",\n start: 2042,\n end: 2052,\n },\n TokenData {\n kind: \"Hash\",\n text: \"#\",\n start: 2054,\n end: 2055,\n },\n TokenData {\n kind: \"Let\",\n text: \"let\",\n start: 2055,\n end: 2058,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"name\",\n start: 2059,\n end: 2063,\n },\n TokenData {\n kind: \"Equal\",\n text: \"=\",\n start: 2064,\n end: 2065,\n },\n TokenData {\n kind: \"StringLiteral\",\n text: \"\\\"World\\\"\",\n start: 2066,\n end: 2073,\n },\n TokenData {\n kind: \"Hash\",\n text: \"#\",\n start: 2075,\n end: 2076,\n },\n TokenData {\n kind: \"LeftParen\",\n text: \"(\",\n start: 2076,\n end: 2077,\n },\n TokenData {\n kind: \"StringLiteral\",\n text: \"\\\"Hello \\\"\",\n start: 2077,\n end: 2085,\n },\n TokenData {\n kind: \"Plus\",\n text: \"+\",\n start: 2086,\n end: 2087,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"name\",\n start: 2088,\n end: 2092,\n },\n TokenData {\n kind: \"RightParen\",\n text: \")\",\n start: 2092,\n end: 2093,\n },\n TokenData {\n kind: \"Heading\",\n text: \"==\",\n start: 2097,\n end: 2099,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"Raw\",\n start: 2100,\n end: 2103,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"Text\",\n start: 2104,\n end: 2108,\n },\n TokenData {\n kind: \"Backtick\",\n text: \"`\",\n start: 2110,\n end: 2111,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"raw\",\n start: 2111,\n end: 2114,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"text\",\n start: 2115,\n end: 2119,\n },\n TokenData {\n kind: \"Backtick\",\n text: \"`\",\n start: 2119,\n end: 2120,\n },\n TokenData {\n kind: \"Backtick\",\n text: \"`\",\n start: 2122,\n end: 2123,\n },\n TokenData {\n kind: \"Backtick\",\n text: \"`\",\n start: 2123,\n end: 2124,\n },\n TokenData {\n kind: \"Backtick\",\n text: \"`\",\n start: 2124,\n end: 2125,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"block\",\n start: 2127,\n end: 2132,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"raw\",\n start: 2133,\n end: 2136,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"text\",\n start: 2137,\n end: 2141,\n },\n TokenData {\n kind: \"Backtick\",\n text: \"`\",\n start: 2143,\n end: 2144,\n },\n TokenData {\n kind: \"Backtick\",\n text: \"`\",\n start: 2144,\n end: 2145,\n },\n TokenData {\n kind: \"Backtick\",\n text: \"`\",\n start: 2145,\n end: 2146,\n },\n TokenData {\n kind: \"Heading\",\n text: \"==\",\n start: 2150,\n end: 2152,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"Shapes\",\n start: 2153,\n end: 2159,\n },\n TokenData {\n kind: \"Hash\",\n text: \"#\",\n start: 2161,\n end: 2162,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"rect\",\n start: 2162,\n end: 2166,\n },\n TokenData {\n kind: \"LeftParen\",\n text: \"(\",\n start: 2166,\n end: 2167,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"width\",\n start: 2167,\n end: 2172,\n },\n TokenData {\n kind: \"Colon\",\n text: \":\",\n start: 2172,\n end: 2173,\n },\n TokenData {\n kind: \"NumericLiteral\",\n text: \"100\",\n start: 2174,\n end: 2177,\n },\n TokenData {\n kind: \"Percent\",\n text: \"%\",\n start: 2177,\n end: 2178,\n },\n TokenData {\n kind: \"Comma\",\n text: \",\",\n start: 2178,\n end: 2179,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"height\",\n start: 2180,\n end: 2186,\n },\n TokenData {\n kind: \"Colon\",\n text: \":\",\n start: 2186,\n end: 2187,\n },\n TokenData {\n kind: \"NumericLiteral\",\n text: \"20\",\n start: 2188,\n end: 2190,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"pt\",\n start: 2190,\n end: 2192,\n },\n TokenData {\n kind: \"Comma\",\n text: \",\",\n start: 2192,\n end: 2193,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"fill\",\n start: 2194,\n end: 2198,\n },\n TokenData {\n kind: \"Colon\",\n text: \":\",\n start: 2198,\n end: 2199,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"blue\",\n start: 2200,\n end: 2204,\n },\n TokenData {\n kind: \"RightParen\",\n text: \")\",\n start: 2204,\n end: 2205,\n },\n TokenData {\n kind: \"Hash\",\n text: \"#\",\n start: 2207,\n end: 2208,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"circle\",\n start: 2208,\n end: 2214,\n },\n TokenData {\n kind: \"LeftParen\",\n text: \"(\",\n start: 2214,\n end: 2215,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"radius\",\n start: 2215,\n end: 2221,\n },\n TokenData {\n kind: \"Colon\",\n text: \":\",\n start: 2221,\n end: 2222,\n },\n TokenData {\n kind: \"NumericLiteral\",\n text: \"10\",\n start: 2223,\n end: 2225,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"pt\",\n start: 2225,\n end: 2227,\n },\n TokenData {\n kind: \"RightParen\",\n text: \")\",\n start: 2227,\n end: 2228,\n },\n TokenData {\n kind: \"Hash\",\n text: \"#\",\n start: 2230,\n end: 2231,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"line\",\n start: 2231,\n end: 2235,\n },\n TokenData {\n kind: \"LeftParen\",\n text: \"(\",\n start: 2235,\n end: 2236,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"length\",\n start: 2236,\n end: 2242,\n },\n TokenData {\n kind: \"Colon\",\n text: \":\",\n start: 2242,\n end: 2243,\n },\n TokenData {\n kind: \"NumericLiteral\",\n text: \"100\",\n start: 2244,\n end: 2247,\n },\n TokenData {\n kind: \"Percent\",\n text: \"%\",\n start: 2247,\n end: 2248,\n },\n TokenData {\n kind: \"RightParen\",\n text: \")\",\n start: 2248,\n end: 2249,\n },\n TokenData {\n kind: \"Heading\",\n text: \"==\",\n start: 2253,\n end: 2255,\n },\n TokenData {\n kind: \"Identifier\",\n text: \"Imports\",\n start: 2256,\n end: 2263,\n },\n TokenData {\n kind: \"Hash\",\n text: \"#\",\n start: 2265,\n end: 2266,\n },\n TokenData {\n kind: \"Import\",\n text: \"import\",\n start: 2266,\n end: 2272,\n },\n TokenData {\n kind: \"StringLiteral\",\n text: \"\\\"utils.typ\\\"\",\n start: 2273,\n end: 2284,\n },\n TokenData {\n kind: \"Colon\",\n text: \":\",\n start: 2284,\n end: 2285,\n },\n TokenData {\n kind: \"Strong\",\n text: \"*\",\n start: 2286,\n end: 2287,\n },\n TokenData {\n kind: \"Hash\",\n text: \"#\",\n start: 2289,\n end: 2290,\n },\n TokenData {\n kind: \"Include\",\n text: \"include\",\n start: 2290,\n end: 2297,\n },\n TokenData {\n kind: \"StringLiteral\",\n text: \"\\\"chapter1.typ\\\"\",\n start: 2298,\n end: 2312,\n },\n TokenData {\n kind: \"Eof\",\n text: \"\",\n start: 2314,\n end: 2314,\n },\n ],\n errors: [],\n}" } [INFO] [stderr] error: test failed, to rerun pass `--test main` [INFO] [stdout] [INFO] [stdout] [INFO] [stdout] failures: [INFO] [stdout] lexer::test_typst_lexer [INFO] [stdout] [INFO] [stdout] test result: FAILED. 1 passed; 1 failed; 0 ignored; 0 measured; 0 filtered out; finished in 0.02s [INFO] [stdout] [INFO] running `Command { std: "docker" "inspect" "b7715e2b4235a6f5ecc8ec59e20d28380f7d9580eb337b45d2fb81e1e3adab27", kill_on_drop: false }` [INFO] running `Command { std: "docker" "rm" "-f" "b7715e2b4235a6f5ecc8ec59e20d28380f7d9580eb337b45d2fb81e1e3adab27", kill_on_drop: false }` [INFO] [stdout] b7715e2b4235a6f5ecc8ec59e20d28380f7d9580eb337b45d2fb81e1e3adab27