@@ -936,24 +936,17 @@ def _read_delete_file(fs: FileSystem, data_file: DataFile) -> Iterator[PositionD
936
936
937
937
938
938
def _read_deletes (fs : FileSystem , data_file : DataFile ) -> Dict [str , pa .ChunkedArray ]:
939
- < << << << HEAD
940
939
if data_file .file_format == FileFormat .PARQUET :
941
940
deletes_by_file : Dict [str , List [int ]] = {}
942
941
for delete in _read_delete_file (fs , data_file ):
943
- if delete .file_path not in deletes_by_file :
944
- deletes_by_file [delete .file_path ] = []
945
- deletes_by_file [delete .file_path ].append (delete .pos )
946
- == == == =
947
- deletes_by_file : Dict [str , List [int ]] = {}
948
- for delete in _read_delete_file (fs , data_file ):
949
- if delete .path not in deletes_by_file :
950
- deletes_by_file [delete .path ] = []
951
- deletes_by_file [delete .path ].append (delete .pos )
952
- > >> >> >> e4ed25e (fix if statment )
942
+ if delete .path not in deletes_by_file :
943
+ deletes_by_file [delete .path ] = []
944
+ deletes_by_file [delete .path ].append (delete .pos )
953
945
954
946
# Convert lists of positions to ChunkedArrays
955
947
return {
956
- file_path : pa .chunked_array ([pa .array (positions , type = pa .int64 ())]) for file_path , positions in deletes_by_file .items ()
948
+ file_path : pa .chunked_array ([pa .array (positions , type = pa .int64 ())])
949
+ for file_path , positions in deletes_by_file .items ()
957
950
}
958
951
elif data_file .file_format == FileFormat .PUFFIN :
959
952
_ , _ , path = PyArrowFileIO .parse_location (data_file .file_path )
@@ -964,6 +957,7 @@ def _read_deletes(fs: FileSystem, data_file: DataFile) -> Dict[str, pa.ChunkedAr
964
957
else :
965
958
raise ValueError (f"Delete file format not supported: { data_file .file_format } " )
966
959
960
+
967
961
def _combine_positional_deletes (positional_deletes : List [pa .ChunkedArray ], start_index : int , end_index : int ) -> pa .Array :
968
962
if len (positional_deletes ) == 1 :
969
963
all_chunks = positional_deletes [0 ]
0 commit comments