diff --git a/tests/test_api.py b/tests/test_api.py index 01851c34..34598a02 100644 --- a/tests/test_api.py +++ b/tests/test_api.py @@ -127,9 +127,16 @@ def test_api_launch(device, scenario, library, task, model): def test_api_push_to_hub_mixin(): benchmark_name = "test_api_push_to_hub_mixin" - scenario_config = InferenceConfig(memory=True, latency=True, duration=1, iterations=1, warmup_runs=1) backend_config = PyTorchConfig(model="google-bert/bert-base-uncased", device="cpu") launcher_config = ProcessConfig(device_isolation=False) + scenario_config = InferenceConfig( + duration=1, + iterations=1, + warmup_runs=1, + memory=True, + latency=True, + input_shapes=INPUT_SHAPES, + ) benchmark_config = BenchmarkConfig( name=benchmark_name, scenario=scenario_config, diff --git a/tests/test_cli.py b/tests/test_cli.py index 3a510806..0ef49651 100644 --- a/tests/test_cli.py +++ b/tests/test_cli.py @@ -65,6 +65,8 @@ def test_cli_exit_code_0(launcher): "name=test", "launcher=" + launcher, # compatible task and model + "scenario.input_shapes.batch_size=1", + "scenario.input_shapes.sequence_length=16", "backend.task=text-classification", "backend.model=bert-base-uncased", "backend.device=cpu", @@ -88,6 +90,8 @@ def test_cli_exit_code_1(launcher): "name=test", "launcher=" + launcher, # incompatible task and model to trigger an error + "scenario.input_shapes.batch_size=1", + "scenario.input_shapes.sequence_length=16", "backend.task=image-classification", "backend.model=bert-base-uncased", "backend.device=cpu", @@ -111,6 +115,8 @@ def test_cli_numactl(launcher): "name=test", "launcher=" + launcher, "launcher.numactl=True", + "scenario.input_shapes.batch_size=1", + "scenario.input_shapes.sequence_length=16", "backend.task=text-classification", "backend.model=bert-base-uncased", "backend.device=cpu",