diff --git a/backends/xnnpack/test/ops/test_check_quant_params.py b/backends/xnnpack/test/ops/test_check_quant_params.py index a04751ca5f1..d05b1fce540 100644 --- a/backends/xnnpack/test/ops/test_check_quant_params.py +++ b/backends/xnnpack/test/ops/test_check_quant_params.py @@ -65,7 +65,7 @@ def _test_check_quant_message(self, ep_modifier, expected_message): with self.assertRaises(ValueError) as context: to_edge_transform_and_lower(aten, partitioner=[XnnpackPartitioner()]) - self.assertEquals(str(context.exception), expected_message) + self.assertEqual(str(context.exception), expected_message) def test_in_per_tensor_quant(self): for invalid_scale in [ diff --git a/examples/models/llava/test/test_llava.py b/examples/models/llava/test/test_llava.py index 5fd60399415..36381b27124 100644 --- a/examples/models/llava/test/test_llava.py +++ b/examples/models/llava/test/test_llava.py @@ -131,7 +131,7 @@ def test_llava_export(self): # being tested, using llama_transformer new_tokens = [torch.argmax(pte_prefill_after_img).item()] # TODO: uncomment this line - # self.assertEquals(new_tokens[0], 1932) # When + # self.assertEqual(new_tokens[0], 1932) # When for i in range(4): print(i, llava_model.tokenizer.decode(new_tokens[i])) token_embeds = llava_module.run_method(