mirror of
https://github.com/nushell/nushell
synced 2025-01-12 21:29:07 +00:00
214714e0ab
<!-- if this PR closes one or more issues, you can automatically link the PR with them by using one of the [*linking keywords*](https://docs.github.com/en/issues/tracking-your-work-with-issues/linking-a-pull-request-to-an-issue#linking-a-pull-request-to-an-issue-using-a-keyword), e.g. - this PR should close #xxxx - fixes #xxxx you can also mention related issues, PRs or discussions! --> # Description <!-- Thank you for improving Nushell. Please, check our [contributing guide](../CONTRIBUTING.md) and talk to the core team before making major changes. Description of your pull request goes here. **Provide examples and/or screenshots** if your changes affect the user experience. --> This PR adds type checking of all command input types at run-time. Generally, these errors should be caught by the parser, but sometimes we can't know the type of a value at parse-time. The simplest example is using the `echo` command, which has an output type of `any`, so prefixing a literal with `echo` will bypass parse-time type checking. Before this PR, each command has to individually check its input types. This can result in scenarios where the input/output types don't match the actual command behavior. This can cause valid usage with an non-`any` type to become a parse-time error if a command is missing that type in its pipeline input/output (`drop nth` and `history import` do this before this PR). Alternatively, a command may not list a type in its input/output types, but doesn't actually reject that type in its code, which can have unintended side effects (`get` does this on an empty pipeline input, and `sort` used to before #13154). After this PR, the type of the pipeline input is checked to ensure it matches one of the input types listed in the proceeding command's input/output types. While each of the issues in the "before this PR" section could be addressed with each command individually, this PR solves this issue for _all_ commands. **This will likely cause some breakage**, as some commands have incorrect input/output types, and should be adjusted. Also, some scripts may have erroneous usage of commands. In writing this PR, I discovered that `toolkit.nu` was passing `null` values to `str join`, which doesn't accept nothing types (if folks think it should, we can adjust it in this PR or in a different PR). I found some issues in the standard library and its tests. I also found that carapace's vendor script had an incorrect chaining of `get -i`: ```nushell let expanded_alias = (scope aliases | where name == $spans.0 | get -i 0 | get -i expansion) ``` Before this PR, if the `get -i 0` ever actually did evaluate to `null`, the second `get` invocation would error since `get` doesn't operate on `null` values. After this PR, this is immediately a run-time error, alerting the user to the problematic code. As a side note, we'll need to PR this fix (`get -i 0 | get -i expansion` -> `get -i 0.expansion`) to carapace. A notable exception to the type checking is commands with input type of `nothing -> <type>`. In this case, any input type is allowed. This allows piping values into the command without an error being thrown. For example, `123 | echo $in` would be an error without this exception. Additionally, custom types bypass type checking (I believe this also happens during parsing, but not certain) I added a `is_subtype` method to `Value` and `PipelineData`. It functions slightly differently than `get_type().is_subtype()`, as noted in the doccomments. Notably, it respects structural typing of lists and tables. For example, the type of a value `[{a: 123} {a: 456, b: 789}]` is a subtype of `table<a: int>`, whereas the type returned by `Value::get_type` is a `list<any>`. Similarly, `PipelineData` has some special handling for `ListStream`s and `ByteStream`s. The latter was needed for this PR to work properly with external commands. Here's some examples. Before: ```nu 1..2 | drop nth 1 Error: nu::parser::input_type_mismatch × Command does not support range input. ╭─[entry #9:1:8] 1 │ 1..2 | drop nth 1 · ────┬─── · ╰── command doesn't support range input ╰──── echo 1..2 | drop nth 1 # => ╭───┬───╮ # => │ 0 │ 1 │ # => ╰───┴───╯ ``` After this PR, I've adjusted `drop nth`'s input/output types to accept range input. Before this PR, zip accepted any value despite not being listed in its input/output types. This caused different behavior depending on if you triggered a parse error or not: ```nushell 1 | zip [2] # => Error: nu::parser::input_type_mismatch # => # => × Command does not support int input. # => ╭─[entry #3:1:5] # => 1 │ 1 | zip [2] # => · ─┬─ # => · ╰── command doesn't support int input # => ╰──── echo 1 | zip [2] # => ╭───┬───────────╮ # => │ 0 │ ╭───┬───╮ │ # => │ │ │ 0 │ 1 │ │ # => │ │ │ 1 │ 2 │ │ # => │ │ ╰───┴───╯ │ # => ╰───┴───────────╯ ``` After this PR, it works the same in both cases. For cases like this, if we do decide we want `zip` or other commands to accept any input value, then we should explicitly add that to the input types. ```nushell 1 | zip [2] # => Error: nu::parser::input_type_mismatch # => # => × Command does not support int input. # => ╭─[entry #3:1:5] # => 1 │ 1 | zip [2] # => · ─┬─ # => · ╰── command doesn't support int input # => ╰──── echo 1 | zip [2] # => Error: nu:🐚:only_supports_this_input_type # => # => × Input type not supported. # => ╭─[entry #14:2:6] # => 2 │ echo 1 | zip [2] # => · ┬ ─┬─ # => · │ ╰── only list<any> and range input data is supported # => · ╰── input type: int # => ╰──── ``` # User-Facing Changes <!-- List of all changes that impact the user experience here. This helps us keep track of breaking changes. --> **Breaking change**: The type of a command's input is now checked against the input/output types of that command at run-time. While these errors should mostly be caught at parse-time, in cases where they can't be detected at parse-time they will be caught at run-time instead. This applies to both internal commands and custom commands. Example function and corresponding parse-time error (same before and after PR): ```nushell def foo []: int -> nothing { print $"my cool int is ($in)" } 1 | foo # => my cool int is 1 "evil string" | foo # => Error: nu::parser::input_type_mismatch # => # => × Command does not support string input. # => ╭─[entry #16:1:17] # => 1 │ "evil string" | foo # => · ─┬─ # => · ╰── command doesn't support string input # => ╰──── # => ``` Before: ```nu echo "evil string" | foo # => my cool int is evil string ``` After: ```nu echo "evil string" | foo # => Error: nu:🐚:only_supports_this_input_type # => # => × Input type not supported. # => ╭─[entry #17:1:6] # => 1 │ echo "evil string" | foo # => · ──────┬────── ─┬─ # => · │ ╰── only int input data is supported # => · ╰── input type: string # => ╰──── ``` Known affected internal commands which erroneously accepted any type: * `str join` * `zip` * `reduce` # Tests + Formatting <!-- Don't forget to add tests that cover your changes. Make sure you've run and fixed any issues with these commands: - `cargo fmt --all -- --check` to check standard code formatting (`cargo fmt --all` applies these changes) - `cargo clippy --workspace -- -D warnings -D clippy::unwrap_used` to check that you're using the standard code style - `cargo test --workspace` to check that all tests pass (on Windows make sure to [enable developer mode](https://learn.microsoft.com/en-us/windows/apps/get-started/developer-mode-features-and-debugging)) - `cargo run -- -c "use toolkit.nu; toolkit test stdlib"` to run the tests for the standard library > **Note** > from `nushell` you can also use the `toolkit` as follows > ```bash > use toolkit.nu # or use an `env_change` hook to activate it automatically > toolkit check pr > ``` --> - 🟢 `toolkit fmt` - 🟢 `toolkit clippy` - 🟢 `toolkit test` - 🟢 `toolkit test stdlib` # After Submitting <!-- If your PR had any user-facing changes, update [the documentation](https://github.com/nushell/nushell.github.io) after the PR is merged, if necessary. This will help us keep the docs up to date. --> * Play whack-a-mole with the commands and scripts this will inevitably break
385 lines
13 KiB
Text
385 lines
13 KiB
Text
use std/log
|
|
export-env {
|
|
# Place NU_FORMAT... environment variables in module-scope
|
|
export use std/log *
|
|
}
|
|
|
|
def "nu-complete threads" [] {
|
|
seq 1 (sys cpu | length)
|
|
}
|
|
|
|
# Here we store the map of annotations internal names and the annotation actually used during test creation
|
|
# The reason we do that is to allow annotations to be easily renamed without modifying rest of the code
|
|
# Functions with no annotations or with annotations not on the list are rejected during module evaluation
|
|
# test and test-skip annotations may be used multiple times throughout the module as the function names are stored in a list
|
|
# Other annotations should only be used once within a module file
|
|
# If you find yourself in need of multiple before- or after- functions it's a sign your test suite probably needs redesign
|
|
def valid-annotations [] {
|
|
{
|
|
"#[test]": "test",
|
|
"#[ignore]": "test-skip",
|
|
"#[before-each]": "before-each"
|
|
"#[before-all]": "before-all"
|
|
"#[after-each]": "after-each"
|
|
"#[after-all]": "after-all"
|
|
}
|
|
}
|
|
|
|
# Returns a table containing the list of function names together with their annotations (comments above the declaration)
|
|
def get-annotated [
|
|
file: path
|
|
]: nothing -> table<function_name: string, annotation: string> {
|
|
let raw_file = (
|
|
open $file
|
|
| lines
|
|
| enumerate
|
|
| flatten
|
|
)
|
|
|
|
$raw_file
|
|
| where item starts-with def and index > 0
|
|
| insert annotation {|x|
|
|
$raw_file
|
|
| get ($x.index - 1)
|
|
| get item
|
|
| str trim
|
|
}
|
|
| where annotation in (valid-annotations|columns)
|
|
| reject index
|
|
| update item {
|
|
split column --collapse-empty ' '
|
|
| get column2.0
|
|
}
|
|
| rename function_name
|
|
}
|
|
|
|
# Takes table of function names and their annotations such as the one returned by get-annotated
|
|
#
|
|
# Returns a record where keys are internal names of valid annotations and values are corresponding function names
|
|
# Annotations that allow multiple functions are of type list<string>
|
|
# Other annotations are of type string
|
|
# Result gets merged with the template record so that the output shape remains consistent regardless of the table content
|
|
def create-test-record []: nothing -> record<before-each: string, after-each: string, before-all: string, after-all: string, test: list<string>, test-skip: list<string>> {
|
|
let input = $in
|
|
|
|
let template_record = {
|
|
before-each: '',
|
|
before-all: '',
|
|
after-each: '',
|
|
after-all: '',
|
|
test-skip: []
|
|
}
|
|
|
|
let test_record = (
|
|
$input
|
|
| update annotation {|x|
|
|
valid-annotations
|
|
| get $x.annotation
|
|
}
|
|
| group-by --to-table annotation
|
|
| update items {|x|
|
|
$x.items.function_name
|
|
| if $x.annotation in ["test", "test-skip"] {
|
|
$in
|
|
} else {
|
|
get 0
|
|
}
|
|
}
|
|
| transpose --ignore-titles -r -d
|
|
)
|
|
|
|
$template_record
|
|
| merge $test_record
|
|
|
|
}
|
|
|
|
def throw-error [error: record] {
|
|
error make {
|
|
msg: $"(ansi red)($error.msg)(ansi reset)"
|
|
label: {
|
|
text: ($error.label)
|
|
span: $error.span
|
|
}
|
|
}
|
|
}
|
|
|
|
# show a test record in a pretty way
|
|
#
|
|
# `$in` must be a `record<file: string, module: string, name: string, pass: bool>`.
|
|
#
|
|
# the output would be like
|
|
# - "<indentation> x <module> <test>" all in red if failed
|
|
# - "<indentation> s <module> <test>" all in yellow if skipped
|
|
# - "<indentation> <module> <test>" all in green if passed
|
|
def show-pretty-test [indent: int = 4] {
|
|
let test = $in
|
|
|
|
[
|
|
(1..$indent | each {" "} | str join)
|
|
(match $test.result {
|
|
"pass" => { ansi green },
|
|
"skip" => { ansi yellow },
|
|
_ => { ansi red }
|
|
})
|
|
(match $test.result {
|
|
"pass" => " ",
|
|
"skip" => "s",
|
|
_ => { char failed }
|
|
})
|
|
" "
|
|
$"($test.name) ($test.test)"
|
|
(ansi reset)
|
|
] | str join
|
|
}
|
|
|
|
# Takes a test record and returns the execution result
|
|
# Test is executed via following steps:
|
|
# * Public function with random name is generated that runs specified test in try/catch block
|
|
# * Module file is opened
|
|
# * Random public function is appended to the end of the file
|
|
# * Modified file is saved under random name
|
|
# * Nu subprocess is spawned
|
|
# * Inside subprocess the modified file is imported and random function called
|
|
# * Output of the random function is serialized into nuon and returned to parent process
|
|
# * Modified file is removed
|
|
def run-test [
|
|
test: record
|
|
] {
|
|
let test_file_name = (random chars --length 10)
|
|
let test_function_name = (random chars --length 10)
|
|
let rendered_module_path = ({parent: ($test.file|path dirname), stem: $test_file_name, extension: nu}| path join)
|
|
|
|
let test_function = $"
|
|
export def ($test_function_name) [] {
|
|
($test.before-each)
|
|
try {
|
|
$context | ($test.test)
|
|
($test.after-each)
|
|
} catch { |err|
|
|
($test.after-each)
|
|
$err | get raw
|
|
}
|
|
}
|
|
"
|
|
open $test.file
|
|
| lines
|
|
| append ($test_function)
|
|
| str join (char nl)
|
|
| save $rendered_module_path
|
|
|
|
let result = (
|
|
^$nu.current-exe --no-config-file -c $"use ($rendered_module_path) *; ($test_function_name)|to nuon"
|
|
| complete
|
|
)
|
|
|
|
rm $rendered_module_path
|
|
|
|
return $result
|
|
}
|
|
|
|
|
|
# Takes a module record and returns a table with following columns:
|
|
#
|
|
# * file - path to file under test
|
|
# * name - name of the module under test
|
|
# * test - name of specific test
|
|
# * result - test execution result
|
|
def run-tests-for-module [
|
|
module: record<file: path name: string before-each: string after-each: string before-all: string after-all: string test: list test-skip: list>
|
|
threads: int
|
|
]: nothing -> table<file: path, name: string, test: string, result: string> {
|
|
let global_context = if not ($module.before-all|is-empty) {
|
|
log info $"Running before-all for module ($module.name)"
|
|
run-test {
|
|
file: $module.file,
|
|
before-each: 'let context = {}',
|
|
after-each: '',
|
|
test: $module.before-all
|
|
}
|
|
| if $in.exit_code == 0 {
|
|
$in.stdout
|
|
} else {
|
|
throw-error {
|
|
msg: "Before-all failed"
|
|
label: "Failure in test setup"
|
|
span: (metadata $in | get span)
|
|
}
|
|
}
|
|
} else {
|
|
{}
|
|
}
|
|
|
|
# since tests are skipped based on their annotation and never actually executed we can generate their list in advance
|
|
let skipped_tests = (
|
|
if not ($module.test-skip|is-empty) {
|
|
$module
|
|
| update test $module.test-skip
|
|
| reject test-skip
|
|
| flatten
|
|
| insert result 'skip'
|
|
} else {
|
|
[]
|
|
}
|
|
)
|
|
|
|
let tests = (
|
|
$module
|
|
| reject test-skip
|
|
| flatten test
|
|
| update before-each {|x|
|
|
if not ($module.before-each|is-empty) {
|
|
$"let context = \(($global_context)|merge \(($module.before-each)\)\)"
|
|
} else {
|
|
$"let context = ($global_context)"
|
|
}
|
|
}
|
|
| update after-each {|x|
|
|
if not ($module.after-each|is-empty) {
|
|
$"$context | ($module.after-each)"
|
|
} else {
|
|
''
|
|
}
|
|
}
|
|
| par-each --threads $threads {|test|
|
|
log info $"Running ($test.test) in module ($module.name)"
|
|
log debug $"Global context is ($global_context)"
|
|
|
|
$test|insert result {|x|
|
|
run-test $test
|
|
| if $in.exit_code == 0 {
|
|
'pass'
|
|
} else {
|
|
'fail'
|
|
}
|
|
}
|
|
}
|
|
| append $skipped_tests
|
|
| select file name test result
|
|
)
|
|
|
|
if not ($module.after-all|is-empty) {
|
|
log info $"Running after-all for module ($module.name)"
|
|
|
|
run-test {
|
|
file: $module.file,
|
|
before-each: $"let context = ($global_context)",
|
|
after-each: '',
|
|
test: $module.after-all
|
|
}
|
|
}
|
|
return $tests
|
|
}
|
|
|
|
# Run tests for nushell code
|
|
#
|
|
# By default all detected tests are executed
|
|
# Test list can be filtered out by specifying either path to search for, name of the module to run tests for or specific test name
|
|
# In order for a function to be recognized as a test by the test runner it needs to be annotated with # test
|
|
# Following annotations are supported by the test runner:
|
|
# * test - test case to be executed during test run
|
|
# * test-skip - test case to be skipped during test run
|
|
# * before-all - function to run at the beginning of test run. Returns a global context record that is piped into every test function
|
|
# * before-each - function to run before every test case. Returns a per-test context record that is merged with global context and piped into test functions
|
|
# * after-each - function to run after every test case. Receives the context record just like the test cases
|
|
# * after-all - function to run after all test cases have been executed. Receives the global context record
|
|
export def run-tests [
|
|
--path: path, # Path to look for tests. Default: current directory.
|
|
--module: string, # Test module to run. Default: all test modules found.
|
|
--test: string, # Pattern to use to include tests. Default: all tests found in the files.
|
|
--exclude: string, # Pattern to use to exclude tests. Default: no tests are excluded
|
|
--exclude-module: string, # Pattern to use to exclude test modules. Default: No modules are excluded
|
|
--list, # list the selected tests without running them.
|
|
--threads: int@"nu-complete threads", # Amount of threads to use for parallel execution. Default: All threads are utilized
|
|
] {
|
|
let available_threads = (sys cpu | length)
|
|
|
|
# Can't use pattern matching here due to https://github.com/nushell/nushell/issues/9198
|
|
let threads = (if $threads == null {
|
|
$available_threads
|
|
} else if $threads < 1 {
|
|
1
|
|
} else if $threads <= $available_threads {
|
|
$threads
|
|
} else {
|
|
$available_threads
|
|
})
|
|
|
|
let module_search_pattern = ('**' | path join ({
|
|
stem: ($module | default "*")
|
|
extension: nu
|
|
} | path join))
|
|
|
|
let path = if $path == null {
|
|
$env.PWD
|
|
} else {
|
|
if not ($path | path exists) {
|
|
throw-error {
|
|
msg: "directory_not_found"
|
|
label: "no such directory"
|
|
span: (metadata $path | get span)
|
|
}
|
|
}
|
|
$path
|
|
}
|
|
|
|
if not ($module | is-empty) {
|
|
try { ls ($path | path join $module_search_pattern) | null } catch {
|
|
throw-error {
|
|
msg: "module_not_found"
|
|
label: $"no such module in ($path)"
|
|
span: (metadata $module | get span)
|
|
}
|
|
}
|
|
}
|
|
|
|
let modules = (
|
|
ls ($path | path join $module_search_pattern | into glob)
|
|
| par-each --threads $threads {|row|
|
|
{
|
|
file: $row.name
|
|
name: ($row.name | path parse | get stem)
|
|
commands: (get-annotated $row.name)
|
|
}
|
|
}
|
|
| filter {|x| ($x.commands|length) > 0}
|
|
| upsert commands {|module|
|
|
$module.commands
|
|
| create-test-record
|
|
}
|
|
| flatten
|
|
| filter {|x| ($x.test|length) > 0}
|
|
| filter {|x| if ($exclude_module|is-empty) {true} else {$x.name !~ $exclude_module}}
|
|
| filter {|x| if ($test|is-empty) {true} else {$x.test|any {|y| $y =~ $test}}}
|
|
| filter {|x| if ($module|is-empty) {true} else {$module == $x.name}}
|
|
| update test {|x|
|
|
$x.test
|
|
| filter {|y| if ($test|is-empty) {true} else {$y =~ $test}}
|
|
| filter {|y| if ($exclude|is-empty) {true} else {$y !~ $exclude}}
|
|
}
|
|
)
|
|
if $list {
|
|
return $modules
|
|
}
|
|
|
|
if ($modules | is-empty) {
|
|
error make --unspanned {msg: "no test to run"}
|
|
}
|
|
|
|
let results = (
|
|
$modules
|
|
| par-each --threads $threads {|module|
|
|
run-tests-for-module $module $threads
|
|
}
|
|
| flatten
|
|
)
|
|
if ($results | any {|x| $x.result == fail}) {
|
|
let text = ([
|
|
$"(ansi purple)some tests did not pass (char lparen)see complete errors below(char rparen):(ansi reset)"
|
|
""
|
|
($results | par-each --threads $threads {|test| ($test | show-pretty-test 4)} | str join "\n")
|
|
""
|
|
] | str join "\n")
|
|
|
|
error make --unspanned { msg: $text }
|
|
}
|
|
}
|