@@ -105,6 +105,7 @@ def test_partitioned_table_rewrite(spark: SparkSession, session_catalog: RestCat
105
105
assert tbl .scan ().to_arrow ().to_pydict () == {"number_partitioned" : [11 , 10 ], "number" : [30 , 30 ]}
106
106
107
107
108
+ @pytest .mark .integration
108
109
@pytest .mark .parametrize ("format_version" , [1 , 2 ])
109
110
def test_rewrite_partitioned_table_with_null (spark : SparkSession , session_catalog : RestCatalog , format_version : int ) -> None :
110
111
identifier = "default.table_partitioned_delete"
@@ -453,6 +454,7 @@ def test_delete_truncate(session_catalog: RestCatalog) -> None:
453
454
assert entries [0 ].status == ManifestEntryStatus .DELETED
454
455
455
456
457
+ @pytest .mark .integration
456
458
def test_delete_overwrite_table_with_null (session_catalog : RestCatalog ) -> None :
457
459
arrow_schema = pa .schema ([pa .field ("ints" , pa .int32 ())])
458
460
arrow_tbl = pa .Table .from_pylist (
@@ -492,6 +494,7 @@ def test_delete_overwrite_table_with_null(session_catalog: RestCatalog) -> None:
492
494
assert tbl .scan ().to_arrow ()["ints" ].to_pylist () == [3 , 4 , 1 , None ]
493
495
494
496
497
+ @pytest .mark .integration
495
498
def test_delete_overwrite_table_with_nan (session_catalog : RestCatalog ) -> None :
496
499
arrow_schema = pa .schema ([pa .field ("floats" , pa .float32 ())])
497
500
0 commit comments