@@ -975,42 +975,6 @@ impl PhysicalPlanner {
975975 . map ( |expr| self . create_expr ( expr, Arc :: clone ( & required_schema) ) )
976976 . collect ( ) ;
977977
978- let default_values: Option < HashMap < usize , ScalarValue > > = if !scan
979- . default_values
980- . is_empty ( )
981- {
982- // We have default values. Extract the two lists (same length) of values and
983- // indexes in the schema, and then create a HashMap to use in the SchemaMapper.
984- let default_values: Result < Vec < ScalarValue > , DataFusionError > = scan
985- . default_values
986- . iter ( )
987- . map ( |expr| {
988- let literal = self . create_expr ( expr, Arc :: clone ( & required_schema) ) ?;
989- let df_literal = literal
990- . as_any ( )
991- . downcast_ref :: < DataFusionLiteral > ( )
992- . ok_or_else ( || {
993- GeneralError ( "Expected literal of default value." . to_string ( ) )
994- } ) ?;
995- Ok ( df_literal. value ( ) . clone ( ) )
996- } )
997- . collect ( ) ;
998- let default_values = default_values?;
999- let default_values_indexes: Vec < usize > = scan
1000- . default_values_indexes
1001- . iter ( )
1002- . map ( |offset| * offset as usize )
1003- . collect ( ) ;
1004- Some (
1005- default_values_indexes
1006- . into_iter ( )
1007- . zip ( default_values)
1008- . collect ( ) ,
1009- )
1010- } else {
1011- None
1012- } ;
1013-
1014978 // Get one file from this partition (we know it's not empty due to early return above)
1015979 let one_file = partition_files
1016980 . partitioned_file
@@ -1034,23 +998,15 @@ impl PhysicalPlanner {
1034998 let files =
1035999 self . get_partitioned_files ( & scan. file_partitions [ self . partition as usize ] ) ?;
10361000 let file_groups: Vec < Vec < PartitionedFile > > = vec ! [ files] ;
1037- let partition_fields: Vec < Field > = partition_schema
1038- . fields ( )
1039- . iter ( )
1040- . map ( |field| {
1041- Field :: new ( field. name ( ) , field. data_type ( ) . clone ( ) , field. is_nullable ( ) )
1042- } )
1043- . collect_vec ( ) ;
1001+
10441002 let scan = init_datasource_exec (
10451003 required_schema,
10461004 Some ( data_schema) ,
10471005 Some ( partition_schema) ,
1048- Some ( partition_fields) ,
10491006 object_store_url,
10501007 file_groups,
10511008 Some ( projection_vector) ,
10521009 Some ( data_filters?) ,
1053- default_values,
10541010 scan. session_timezone . as_str ( ) ,
10551011 scan. case_sensitive ,
10561012 self . session_ctx ( ) ,
@@ -3437,8 +3393,6 @@ mod tests {
34373393
34383394 use crate :: execution:: operators:: ExecutionError ;
34393395 use crate :: execution:: planner:: literal_to_array_ref;
3440- use crate :: parquet:: parquet_support:: SparkParquetOptions ;
3441- use crate :: parquet:: schema_adapter:: SparkSchemaAdapterFactory ;
34423396 use datafusion_comet_proto:: spark_expression:: expr:: ExprStruct ;
34433397 use datafusion_comet_proto:: spark_expression:: ListLiteral ;
34443398 use datafusion_comet_proto:: {
@@ -3448,7 +3402,6 @@ mod tests {
34483402 spark_operator,
34493403 spark_operator:: { operator:: OpStruct , Operator } ,
34503404 } ;
3451- use datafusion_comet_spark_expr:: EvalMode ;
34523405
34533406 #[ test]
34543407 fn test_unpack_dictionary_primitive ( ) {
0 commit comments