diff --git a/compose.yaml b/compose.yaml index d8da91c58bb0..e0134f1c32da 100644 --- a/compose.yaml +++ b/compose.yaml @@ -161,7 +161,7 @@ services: test: - CMD-SHELL - trino --output-format null --execute 'show schemas in hive; show schemas in memory' - image: trinodb/trino:449 + image: trinodb/trino:454 ports: - 8080:8080 networks: diff --git a/ibis/backends/tests/test_client.py b/ibis/backends/tests/test_client.py index 0dd70413bbc1..49848d8f43dc 100644 --- a/ibis/backends/tests/test_client.py +++ b/ibis/backends/tests/test_client.py @@ -36,7 +36,6 @@ PyODBCProgrammingError, PySparkAnalysisException, SnowflakeProgrammingError, - TrinoUserError, ) from ibis.util import gen_name @@ -523,9 +522,6 @@ def test_insert_no_overwrite_from_dataframe( @pytest.mark.notyet( ["datafusion"], raises=Exception, reason="DELETE DML not implemented upstream" ) -@pytest.mark.notyet( - ["trino"], raises=TrinoUserError, reason="requires a non-memory connector" -) @pytest.mark.notyet(["druid"], raises=NotImplementedError) def test_insert_overwrite_from_dataframe( backend, con, employee_data_1_temp_table, test_employee_data_2 @@ -565,11 +561,6 @@ def test_insert_no_overwrite_from_expr( @pytest.mark.notyet( ["datafusion"], raises=Exception, reason="DELETE DML not implemented upstream" ) -@pytest.mark.notyet( - ["trino"], - raises=TrinoUserError, - reason="requires a non-memory connector for truncation", -) @pytest.mark.notyet( ["risingwave"], raises=PsycoPg2InternalError, @@ -590,9 +581,6 @@ def test_insert_overwrite_from_expr( ) -@pytest.mark.notyet( - ["trino"], reason="memory connector doesn't allow writing to tables" -) @pytest.mark.notimpl( ["polars", "pandas", "dask"], reason="`insert` method not implemented" )