@@ -980,6 +980,7 @@ def merge(
980
980
error_on_type_mismatch : bool = True ,
981
981
writer_properties : Optional [WriterProperties ] = None ,
982
982
large_dtypes : Optional [bool ] = None ,
983
+ streaming : bool = False ,
983
984
custom_metadata : Optional [Dict [str , str ]] = None ,
984
985
post_commithook_properties : Optional [PostCommitHookProperties ] = None ,
985
986
commit_properties : Optional [CommitProperties ] = None ,
@@ -997,6 +998,7 @@ def merge(
997
998
error_on_type_mismatch: specify if merge will return error if data types are mismatching :default = True
998
999
writer_properties: Pass writer properties to the Rust parquet writer
999
1000
large_dtypes: Deprecated, will be removed in 1.0
1001
+ streaming: Will execute MERGE using a LazyMemoryExec plan
1000
1002
arrow_schema_conversion_mode: Large converts all types of data schema into Large Arrow types, passthrough keeps string/binary/list types untouched
1001
1003
custom_metadata: Deprecated and will be removed in future versions. Use commit_properties instead.
1002
1004
post_commithook_properties: properties for the post commit hook. If None, default values are used.
@@ -1035,17 +1037,14 @@ def merge(
1035
1037
convert_pyarrow_table ,
1036
1038
)
1037
1039
1038
- streaming = False
1039
1040
if isinstance (source , pyarrow .RecordBatchReader ):
1040
1041
source = convert_pyarrow_recordbatchreader (source , conversion_mode )
1041
- streaming = True
1042
1042
elif isinstance (source , pyarrow .RecordBatch ):
1043
1043
source = convert_pyarrow_recordbatch (source , conversion_mode )
1044
1044
elif isinstance (source , pyarrow .Table ):
1045
1045
source = convert_pyarrow_table (source , conversion_mode )
1046
1046
elif isinstance (source , ds .Dataset ):
1047
1047
source = convert_pyarrow_dataset (source , conversion_mode )
1048
- streaming = True
1049
1048
elif _has_pandas and isinstance (source , pd .DataFrame ):
1050
1049
source = convert_pyarrow_table (
1051
1050
pyarrow .Table .from_pandas (source ), conversion_mode
0 commit comments