From 3df8c5d746676836c3f44de29e53b87ad5c1603b Mon Sep 17 00:00:00 2001 From: Phillip Cloud <417981+cpcloud@users.noreply.github.com> Date: Tue, 17 Dec 2024 07:37:47 -0500 Subject: [PATCH] test(snowflake,bigquery): enable xpassing tests and relax read_csv schema assertion from float64 to numeric --- ibis/backends/tests/test_io.py | 52 ++++------------------------------ 1 file changed, 5 insertions(+), 47 deletions(-) diff --git a/ibis/backends/tests/test_io.py b/ibis/backends/tests/test_io.py index 7223db704116..6d977f34cce0 100644 --- a/ibis/backends/tests/test_io.py +++ b/ibis/backends/tests/test_io.py @@ -119,14 +119,12 @@ def test_read_csv(con, data_dir, in_table_name, num_diamonds): @pytest.mark.notimpl(["datafusion"]) @pytest.mark.notyet( [ - "bigquery", "flink", "impala", "mssql", "mysql", "postgres", "risingwave", - "snowflake", "sqlite", "trino", "databricks", @@ -140,18 +138,7 @@ def test_read_csv_gz(con, data_dir, gzip_csv): @pytest.mark.notyet( - [ - "bigquery", - "flink", - "impala", - "mssql", - "mysql", - "postgres", - "risingwave", - "snowflake", - "sqlite", - "trino", - ] + ["flink", "impala", "mssql", "mysql", "postgres", "risingwave", "sqlite", "trino"] ) def test_read_csv_with_dotted_name(con, data_dir, tmp_path): basename = "foo.bar.baz/diamonds.csv" @@ -166,18 +153,7 @@ def test_read_csv_with_dotted_name(con, data_dir, tmp_path): @pytest.mark.notyet( - [ - "bigquery", - "flink", - "impala", - "mssql", - "mysql", - "postgres", - "risingwave", - "snowflake", - "sqlite", - "trino", - ] + ["flink", "impala", "mssql", "mysql", "postgres", "risingwave", "sqlite", "trino"] ) def test_read_csv_schema(con, tmp_path): foo = tmp_path.joinpath("foo.csv") @@ -197,21 +173,12 @@ def test_read_csv_schema(con, tmp_path): assert result_schema.names == ("cola", "colb", "colc") assert result_schema["cola"].is_integer() - assert result_schema["colb"].is_float64() + assert result_schema["colb"].is_numeric() assert result_schema["colc"].is_string() @pytest.mark.notyet( - [ - "flink", - "impala", - "mssql", - "mysql", - "postgres", - "risingwave", - "sqlite", - "trino", - ] + ["flink", "impala", "mssql", "mysql", "postgres", "risingwave", "sqlite", "trino"] ) def test_read_csv_glob(con, tmp_path, ft_data): pc = pytest.importorskip("pyarrow.csv") @@ -314,16 +281,7 @@ def test_read_parquet_iterator( @pytest.mark.notyet( - [ - "flink", - "impala", - "mssql", - "mysql", - "postgres", - "risingwave", - "sqlite", - "trino", - ] + ["flink", "impala", "mssql", "mysql", "postgres", "risingwave", "sqlite", "trino"] ) def test_read_parquet_glob(con, tmp_path, ft_data): pq = pytest.importorskip("pyarrow.parquet")