diff --git a/backends/xnnpack/test/ops/test_check_quant_params.py b/backends/xnnpack/test/ops/test_check_quant_params.py
index a04751ca5f1..d05b1fce540 100644
--- a/backends/xnnpack/test/ops/test_check_quant_params.py
+++ b/backends/xnnpack/test/ops/test_check_quant_params.py
@@ -65,7 +65,7 @@ def _test_check_quant_message(self, ep_modifier, expected_message):
         with self.assertRaises(ValueError) as context:
             to_edge_transform_and_lower(aten, partitioner=[XnnpackPartitioner()])
 
-        self.assertEquals(str(context.exception), expected_message)
+        self.assertEqual(str(context.exception), expected_message)
 
     def test_in_per_tensor_quant(self):
         for invalid_scale in [
diff --git a/examples/models/llava/test/test_llava.py b/examples/models/llava/test/test_llava.py
index 5fd60399415..36381b27124 100644
--- a/examples/models/llava/test/test_llava.py
+++ b/examples/models/llava/test/test_llava.py
@@ -131,7 +131,7 @@ def test_llava_export(self):
         # being tested, using llama_transformer
         new_tokens = [torch.argmax(pte_prefill_after_img).item()]
         # TODO: uncomment this line
-        # self.assertEquals(new_tokens[0], 1932)  # When
+        # self.assertEqual(new_tokens[0], 1932)  # When
         for i in range(4):
             print(i, llava_model.tokenizer.decode(new_tokens[i]))
             token_embeds = llava_module.run_method(