@@ -365,7 +365,7 @@ impl PySessionContext {
365
365
table_partition_cols
366
366
. into_iter ( )
367
367
. map ( |( name, ty) | ( name, ty. 0 ) )
368
- . collect :: < Vec < ( String , DataType ) > > ( )
368
+ . collect :: < Vec < ( String , DataType ) > > ( ) ,
369
369
)
370
370
. with_file_sort_order (
371
371
file_sort_order
@@ -647,7 +647,7 @@ impl PySessionContext {
647
647
table_partition_cols
648
648
. into_iter ( )
649
649
. map ( |( name, ty) | ( name, ty. 0 ) )
650
- . collect :: < Vec < ( String , DataType ) > > ( )
650
+ . collect :: < Vec < ( String , DataType ) > > ( ) ,
651
651
)
652
652
. parquet_pruning ( parquet_pruning)
653
653
. skip_metadata ( skip_metadata) ;
@@ -742,7 +742,7 @@ impl PySessionContext {
742
742
table_partition_cols
743
743
. into_iter ( )
744
744
. map ( |( name, ty) | ( name, ty. 0 ) )
745
- . collect :: < Vec < ( String , DataType ) > > ( )
745
+ . collect :: < Vec < ( String , DataType ) > > ( ) ,
746
746
) ;
747
747
options. schema_infer_max_records = schema_infer_max_records;
748
748
options. file_extension = file_extension;
@@ -773,13 +773,12 @@ impl PySessionContext {
773
773
. to_str ( )
774
774
. ok_or_else ( || PyValueError :: new_err ( "Unable to convert path to a string" ) ) ?;
775
775
776
- let mut options = AvroReadOptions :: default ( )
777
- . table_partition_cols (
778
- table_partition_cols
779
- . into_iter ( )
780
- . map ( |( name, ty) | ( name, ty. 0 ) )
781
- . collect :: < Vec < ( String , DataType ) > > ( )
782
- ) ;
776
+ let mut options = AvroReadOptions :: default ( ) . table_partition_cols (
777
+ table_partition_cols
778
+ . into_iter ( )
779
+ . map ( |( name, ty) | ( name, ty. 0 ) )
780
+ . collect :: < Vec < ( String , DataType ) > > ( ) ,
781
+ ) ;
783
782
options. file_extension = file_extension;
784
783
options. schema = schema. as_ref ( ) . map ( |x| & x. 0 ) ;
785
784
@@ -882,7 +881,7 @@ impl PySessionContext {
882
881
table_partition_cols
883
882
. into_iter ( )
884
883
. map ( |( name, ty) | ( name, ty. 0 ) )
885
- . collect :: < Vec < ( String , DataType ) > > ( )
884
+ . collect :: < Vec < ( String , DataType ) > > ( ) ,
886
885
)
887
886
. file_compression_type ( parse_file_compression_type ( file_compression_type) ?) ;
888
887
options. schema_infer_max_records = schema_infer_max_records;
@@ -936,7 +935,7 @@ impl PySessionContext {
936
935
table_partition_cols
937
936
. into_iter ( )
938
937
. map ( |( name, ty) | ( name, ty. 0 ) )
939
- . collect :: < Vec < ( String , DataType ) > > ( )
938
+ . collect :: < Vec < ( String , DataType ) > > ( ) ,
940
939
)
941
940
. file_compression_type ( parse_file_compression_type ( file_compression_type) ?) ;
942
941
options. schema = schema. as_ref ( ) . map ( |x| & x. 0 ) ;
@@ -980,7 +979,7 @@ impl PySessionContext {
980
979
table_partition_cols
981
980
. into_iter ( )
982
981
. map ( |( name, ty) | ( name, ty. 0 ) )
983
- . collect :: < Vec < ( String , DataType ) > > ( )
982
+ . collect :: < Vec < ( String , DataType ) > > ( ) ,
984
983
)
985
984
. parquet_pruning ( parquet_pruning)
986
985
. skip_metadata ( skip_metadata) ;
@@ -1007,13 +1006,12 @@ impl PySessionContext {
1007
1006
file_extension : & str ,
1008
1007
py : Python ,
1009
1008
) -> PyDataFusionResult < PyDataFrame > {
1010
- let mut options = AvroReadOptions :: default ( )
1011
- . table_partition_cols (
1012
- table_partition_cols
1013
- . into_iter ( )
1014
- . map ( |( name, ty) | ( name, ty. 0 ) )
1015
- . collect :: < Vec < ( String , DataType ) > > ( )
1016
- ) ;
1009
+ let mut options = AvroReadOptions :: default ( ) . table_partition_cols (
1010
+ table_partition_cols
1011
+ . into_iter ( )
1012
+ . map ( |( name, ty) | ( name, ty. 0 ) )
1013
+ . collect :: < Vec < ( String , DataType ) > > ( ) ,
1014
+ ) ;
1017
1015
options. file_extension = file_extension;
1018
1016
let df = if let Some ( schema) = schema {
1019
1017
options. schema = Some ( & schema. 0 ) ;
0 commit comments