mirror of
https://github.com/nushell/nushell
synced 2024-12-28 05:53:09 +00:00
13cbfd3cee
- Refactor command benchmark - should simplify addition of more command benchmarks - just provide the pipeline you want to run/ choose potential scaling - Bench engine state with ctrlc set - `None` vs `AtomicBool` behind `Arc`
311 lines
9.9 KiB
Rust
311 lines
9.9 KiB
Rust
use nu_cli::{eval_source, evaluate_commands};
|
|
use nu_parser::parse;
|
|
use nu_plugin::{Encoder, EncodingType, PluginCallResponse, PluginOutput};
|
|
use nu_protocol::{
|
|
engine::EngineState, eval_const::create_nu_constant, PipelineData, Span, Spanned, Value,
|
|
NU_VARIABLE_ID,
|
|
};
|
|
use nu_std::load_standard_library;
|
|
use nu_utils::{get_default_config, get_default_env};
|
|
use std::path::{Path, PathBuf};
|
|
|
|
fn main() {
|
|
// Run registered benchmarks.
|
|
divan::main();
|
|
}
|
|
|
|
fn load_bench_commands() -> EngineState {
|
|
nu_command::add_shell_command_context(nu_cmd_lang::create_default_context())
|
|
}
|
|
|
|
fn canonicalize_path(engine_state: &EngineState, path: &Path) -> PathBuf {
|
|
let cwd = engine_state.current_work_dir();
|
|
|
|
if path.exists() {
|
|
match nu_path::canonicalize_with(path, cwd) {
|
|
Ok(canon_path) => canon_path,
|
|
Err(_) => path.to_owned(),
|
|
}
|
|
} else {
|
|
path.to_owned()
|
|
}
|
|
}
|
|
|
|
fn get_home_path(engine_state: &EngineState) -> PathBuf {
|
|
nu_path::home_dir()
|
|
.map(|path| canonicalize_path(engine_state, &path))
|
|
.unwrap_or_default()
|
|
}
|
|
|
|
fn setup_engine() -> EngineState {
|
|
let mut engine_state = load_bench_commands();
|
|
let home_path = get_home_path(&engine_state);
|
|
|
|
// parsing config.nu breaks without PWD set, so set a valid path
|
|
engine_state.add_env_var(
|
|
"PWD".into(),
|
|
Value::string(home_path.to_string_lossy(), Span::test_data()),
|
|
);
|
|
|
|
let nu_const = create_nu_constant(&engine_state, Span::unknown())
|
|
.expect("Failed to create nushell constant.");
|
|
engine_state.set_variable_const_val(NU_VARIABLE_ID, nu_const);
|
|
|
|
engine_state
|
|
}
|
|
|
|
// FIXME: All benchmarks live in this 1 file to speed up build times when benchmarking.
|
|
// When the *_benchmarks functions were in different files, `cargo bench` would build
|
|
// an executable for every single one - incredibly slowly. Would be nice to figure out
|
|
// a way to split things up again.
|
|
|
|
#[divan::bench]
|
|
fn load_standard_lib(bencher: divan::Bencher) {
|
|
let engine = setup_engine();
|
|
bencher
|
|
.with_inputs(|| engine.clone())
|
|
.bench_values(|mut engine| {
|
|
load_standard_library(&mut engine).unwrap();
|
|
})
|
|
}
|
|
|
|
#[divan::bench_group]
|
|
mod eval_commands {
|
|
use super::*;
|
|
|
|
fn bench_command(bencher: divan::Bencher, scaled_command: String) {
|
|
let mut engine = setup_engine();
|
|
load_standard_library(&mut engine).unwrap();
|
|
let commands = Spanned {
|
|
span: Span::unknown(),
|
|
item: scaled_command,
|
|
};
|
|
|
|
bencher
|
|
.with_inputs(|| engine.clone())
|
|
.bench_values(|mut engine| {
|
|
evaluate_commands(
|
|
&commands,
|
|
&mut engine,
|
|
&mut nu_protocol::engine::Stack::new(),
|
|
PipelineData::empty(),
|
|
None,
|
|
)
|
|
.unwrap();
|
|
})
|
|
}
|
|
|
|
#[divan::bench(args = [100, 1_000, 10_000])]
|
|
fn interleave(bencher: divan::Bencher, n: i32) {
|
|
bench_command(
|
|
bencher,
|
|
format!("seq 1 {n} | wrap a | interleave {{ seq 1 {n} | wrap b }} | ignore"),
|
|
)
|
|
}
|
|
|
|
#[divan::bench(args = [100, 1_000, 10_000])]
|
|
fn interleave_with_ctrlc(bencher: divan::Bencher, n: i32) {
|
|
let mut engine = setup_engine();
|
|
engine.ctrlc = Some(std::sync::Arc::new(std::sync::atomic::AtomicBool::new(
|
|
false,
|
|
)));
|
|
load_standard_library(&mut engine).unwrap();
|
|
let commands = Spanned {
|
|
span: Span::unknown(),
|
|
item: format!("seq 1 {n} | wrap a | interleave {{ seq 1 {n} | wrap b }} | ignore"),
|
|
};
|
|
|
|
bencher
|
|
.with_inputs(|| engine.clone())
|
|
.bench_values(|mut engine| {
|
|
evaluate_commands(
|
|
&commands,
|
|
&mut engine,
|
|
&mut nu_protocol::engine::Stack::new(),
|
|
PipelineData::empty(),
|
|
None,
|
|
)
|
|
.unwrap();
|
|
})
|
|
}
|
|
|
|
#[divan::bench(args = [1, 5, 10, 100, 1_000])]
|
|
fn for_range(bencher: divan::Bencher, n: i32) {
|
|
bench_command(bencher, format!("(for $x in (1..{}) {{ sleep 50ns }})", n))
|
|
}
|
|
|
|
#[divan::bench(args = [1, 5, 10, 100, 1_000])]
|
|
fn each(bencher: divan::Bencher, n: i32) {
|
|
bench_command(
|
|
bencher,
|
|
format!("(1..{}) | each {{|_| sleep 50ns }} | ignore", n),
|
|
)
|
|
}
|
|
|
|
#[divan::bench(args = [1, 5, 10, 100, 1_000])]
|
|
fn par_each_1t(bencher: divan::Bencher, n: i32) {
|
|
bench_command(
|
|
bencher,
|
|
format!("(1..{}) | par-each -t 1 {{|_| sleep 50ns }} | ignore", n),
|
|
)
|
|
}
|
|
|
|
#[divan::bench(args = [1, 5, 10, 100, 1_000])]
|
|
fn par_each_2t(bencher: divan::Bencher, n: i32) {
|
|
bench_command(
|
|
bencher,
|
|
format!("(1..{}) | par-each -t 2 {{|_| sleep 50ns }} | ignore", n),
|
|
)
|
|
}
|
|
}
|
|
|
|
#[divan::bench_group()]
|
|
mod parser_benchmarks {
|
|
use super::*;
|
|
|
|
#[divan::bench()]
|
|
fn parse_default_config_file(bencher: divan::Bencher) {
|
|
let engine_state = setup_engine();
|
|
let default_env = get_default_config().as_bytes();
|
|
|
|
bencher
|
|
.with_inputs(|| nu_protocol::engine::StateWorkingSet::new(&engine_state))
|
|
.bench_refs(|working_set| parse(working_set, None, default_env, false))
|
|
}
|
|
|
|
#[divan::bench()]
|
|
fn parse_default_env_file(bencher: divan::Bencher) {
|
|
let engine_state = setup_engine();
|
|
let default_env = get_default_env().as_bytes();
|
|
|
|
bencher
|
|
.with_inputs(|| nu_protocol::engine::StateWorkingSet::new(&engine_state))
|
|
.bench_refs(|working_set| parse(working_set, None, default_env, false))
|
|
}
|
|
}
|
|
|
|
#[divan::bench_group()]
|
|
mod eval_benchmarks {
|
|
use super::*;
|
|
|
|
#[divan::bench()]
|
|
fn eval_default_env(bencher: divan::Bencher) {
|
|
let default_env = get_default_env().as_bytes();
|
|
let fname = "default_env.nu";
|
|
bencher
|
|
.with_inputs(|| (setup_engine(), nu_protocol::engine::Stack::new()))
|
|
.bench_values(|(mut engine_state, mut stack)| {
|
|
eval_source(
|
|
&mut engine_state,
|
|
&mut stack,
|
|
default_env,
|
|
fname,
|
|
PipelineData::empty(),
|
|
false,
|
|
)
|
|
})
|
|
}
|
|
|
|
#[divan::bench()]
|
|
fn eval_default_config(bencher: divan::Bencher) {
|
|
let default_env = get_default_config().as_bytes();
|
|
let fname = "default_config.nu";
|
|
bencher
|
|
.with_inputs(|| (setup_engine(), nu_protocol::engine::Stack::new()))
|
|
.bench_values(|(mut engine_state, mut stack)| {
|
|
eval_source(
|
|
&mut engine_state,
|
|
&mut stack,
|
|
default_env,
|
|
fname,
|
|
PipelineData::empty(),
|
|
false,
|
|
)
|
|
})
|
|
}
|
|
}
|
|
|
|
// generate a new table data with `row_cnt` rows, `col_cnt` columns.
|
|
fn encoding_test_data(row_cnt: usize, col_cnt: usize) -> Value {
|
|
let record = Value::test_record(
|
|
(0..col_cnt)
|
|
.map(|x| (format!("col_{x}"), Value::test_int(x as i64)))
|
|
.collect(),
|
|
);
|
|
|
|
Value::list(vec![record; row_cnt], Span::test_data())
|
|
}
|
|
|
|
#[divan::bench_group()]
|
|
mod encoding_benchmarks {
|
|
use super::*;
|
|
|
|
#[divan::bench(args = [(100, 5), (10000, 15)])]
|
|
fn json_encode(bencher: divan::Bencher, (row_cnt, col_cnt): (usize, usize)) {
|
|
let test_data = PluginOutput::CallResponse(
|
|
0,
|
|
PluginCallResponse::value(encoding_test_data(row_cnt, col_cnt)),
|
|
);
|
|
let encoder = EncodingType::try_from_bytes(b"json").unwrap();
|
|
bencher
|
|
.with_inputs(Vec::new)
|
|
.bench_values(|mut res| encoder.encode(&test_data, &mut res))
|
|
}
|
|
|
|
#[divan::bench(args = [(100, 5), (10000, 15)])]
|
|
fn msgpack_encode(bencher: divan::Bencher, (row_cnt, col_cnt): (usize, usize)) {
|
|
let test_data = PluginOutput::CallResponse(
|
|
0,
|
|
PluginCallResponse::value(encoding_test_data(row_cnt, col_cnt)),
|
|
);
|
|
let encoder = EncodingType::try_from_bytes(b"msgpack").unwrap();
|
|
bencher
|
|
.with_inputs(Vec::new)
|
|
.bench_values(|mut res| encoder.encode(&test_data, &mut res))
|
|
}
|
|
}
|
|
|
|
#[divan::bench_group()]
|
|
mod decoding_benchmarks {
|
|
use super::*;
|
|
|
|
#[divan::bench(args = [(100, 5), (10000, 15)])]
|
|
fn json_decode(bencher: divan::Bencher, (row_cnt, col_cnt): (usize, usize)) {
|
|
let test_data = PluginOutput::CallResponse(
|
|
0,
|
|
PluginCallResponse::value(encoding_test_data(row_cnt, col_cnt)),
|
|
);
|
|
let encoder = EncodingType::try_from_bytes(b"json").unwrap();
|
|
let mut res = vec![];
|
|
encoder.encode(&test_data, &mut res).unwrap();
|
|
bencher
|
|
.with_inputs(|| {
|
|
let mut binary_data = std::io::Cursor::new(res.clone());
|
|
binary_data.set_position(0);
|
|
binary_data
|
|
})
|
|
.bench_values(|mut binary_data| -> Result<Option<PluginOutput>, _> {
|
|
encoder.decode(&mut binary_data)
|
|
})
|
|
}
|
|
|
|
#[divan::bench(args = [(100, 5), (10000, 15)])]
|
|
fn msgpack_decode(bencher: divan::Bencher, (row_cnt, col_cnt): (usize, usize)) {
|
|
let test_data = PluginOutput::CallResponse(
|
|
0,
|
|
PluginCallResponse::value(encoding_test_data(row_cnt, col_cnt)),
|
|
);
|
|
let encoder = EncodingType::try_from_bytes(b"msgpack").unwrap();
|
|
let mut res = vec![];
|
|
encoder.encode(&test_data, &mut res).unwrap();
|
|
bencher
|
|
.with_inputs(|| {
|
|
let mut binary_data = std::io::Cursor::new(res.clone());
|
|
binary_data.set_position(0);
|
|
binary_data
|
|
})
|
|
.bench_values(|mut binary_data| -> Result<Option<PluginOutput>, _> {
|
|
encoder.decode(&mut binary_data)
|
|
})
|
|
}
|
|
}
|