Skip to content

Commit f52809e

Browse files
marvinlanhenkefindepi
authored andcommitted
chore: remove interval test (apache#10888)
1 parent 6e7fd9e commit f52809e

2 files changed

Lines changed: 2 additions & 159 deletions

File tree

datafusion/core/tests/parquet/arrow_statistics.rs

Lines changed: 1 addition & 80 deletions
Original file line numberDiff line numberDiff line change
@@ -30,8 +30,7 @@ use arrow::datatypes::{
3030
use arrow_array::{
3131
make_array, Array, ArrayRef, BinaryArray, BooleanArray, Date32Array, Date64Array,
3232
Decimal128Array, Decimal256Array, FixedSizeBinaryArray, Float16Array, Float32Array,
33-
Float64Array, Int16Array, Int32Array, Int64Array, Int8Array, IntervalDayTimeArray,
34-
IntervalMonthDayNanoArray, IntervalYearMonthArray, LargeBinaryArray,
33+
Float64Array, Int16Array, Int32Array, Int64Array, Int8Array, LargeBinaryArray,
3534
LargeStringArray, RecordBatch, StringArray, Time32MillisecondArray,
3635
Time32SecondArray, Time64MicrosecondArray, Time64NanosecondArray,
3736
TimestampMicrosecondArray, TimestampMillisecondArray, TimestampNanosecondArray,
@@ -1061,84 +1060,6 @@ async fn test_dates_64_diff_rg_sizes() {
10611060
.run();
10621061
}
10631062

1064-
#[tokio::test]
1065-
#[should_panic]
1066-
// Currently this test `should_panic` since statistics for `Intervals`
1067-
// are not supported and `IntervalMonthDayNano` cannot be written
1068-
// to parquet yet.
1069-
// Refer to issue: https://github.com/apache/arrow-rs/issues/5847
1070-
// and https://github.com/apache/arrow-rs/blob/master/parquet/src/arrow/arrow_writer/mod.rs#L747
1071-
async fn test_interval_diff_rg_sizes() {
1072-
// This creates a parquet files of 3 columns:
1073-
// "year_month" --> IntervalYearMonthArray
1074-
// "day_time" --> IntervalDayTimeArray
1075-
// "month_day_nano" --> IntervalMonthDayNanoArray
1076-
//
1077-
// The file is created by 4 record batches (each has a null row)
1078-
// each has 5 rows but then will be split into 2 row groups with size 13, 7
1079-
let reader = TestReader {
1080-
scenario: Scenario::Interval,
1081-
row_per_group: 13,
1082-
}
1083-
.build()
1084-
.await;
1085-
1086-
// TODO: expected values need to be changed once issue is resolved
1087-
// expected_min: Arc::new(IntervalYearMonthArray::from(vec![
1088-
// IntervalYearMonthType::make_value(1, 10),
1089-
// IntervalYearMonthType::make_value(4, 13),
1090-
// ])),
1091-
// expected_max: Arc::new(IntervalYearMonthArray::from(vec![
1092-
// IntervalYearMonthType::make_value(6, 51),
1093-
// IntervalYearMonthType::make_value(8, 53),
1094-
// ])),
1095-
Test {
1096-
reader: &reader,
1097-
expected_min: Arc::new(IntervalYearMonthArray::from(vec![None, None])),
1098-
expected_max: Arc::new(IntervalYearMonthArray::from(vec![None, None])),
1099-
expected_null_counts: UInt64Array::from(vec![2, 2]),
1100-
expected_row_counts: UInt64Array::from(vec![13, 7]),
1101-
column_name: "year_month",
1102-
}
1103-
.run();
1104-
1105-
// expected_min: Arc::new(IntervalDayTimeArray::from(vec![
1106-
// IntervalDayTimeType::make_value(1, 10),
1107-
// IntervalDayTimeType::make_value(4, 13),
1108-
// ])),
1109-
// expected_max: Arc::new(IntervalDayTimeArray::from(vec![
1110-
// IntervalDayTimeType::make_value(6, 51),
1111-
// IntervalDayTimeType::make_value(8, 53),
1112-
// ])),
1113-
Test {
1114-
reader: &reader,
1115-
expected_min: Arc::new(IntervalDayTimeArray::from(vec![None, None])),
1116-
expected_max: Arc::new(IntervalDayTimeArray::from(vec![None, None])),
1117-
expected_null_counts: UInt64Array::from(vec![2, 2]),
1118-
expected_row_counts: UInt64Array::from(vec![13, 7]),
1119-
column_name: "day_time",
1120-
}
1121-
.run();
1122-
1123-
// expected_min: Arc::new(IntervalMonthDayNanoArray::from(vec![
1124-
// IntervalMonthDayNanoType::make_value(1, 10, 100),
1125-
// IntervalMonthDayNanoType::make_value(4, 13, 103),
1126-
// ])),
1127-
// expected_max: Arc::new(IntervalMonthDayNanoArray::from(vec![
1128-
// IntervalMonthDayNanoType::make_value(6, 51, 501),
1129-
// IntervalMonthDayNanoType::make_value(8, 53, 503),
1130-
// ])),
1131-
Test {
1132-
reader: &reader,
1133-
expected_min: Arc::new(IntervalMonthDayNanoArray::from(vec![None, None])),
1134-
expected_max: Arc::new(IntervalMonthDayNanoArray::from(vec![None, None])),
1135-
expected_null_counts: UInt64Array::from(vec![2, 2]),
1136-
expected_row_counts: UInt64Array::from(vec![13, 7]),
1137-
column_name: "month_day_nano",
1138-
}
1139-
.run();
1140-
}
1141-
11421063
#[tokio::test]
11431064
async fn test_uint() {
11441065
// This creates a parquet files of 4 columns named "u8", "u16", "u32", "u64"

datafusion/core/tests/parquet/mod.rs

Lines changed: 1 addition & 79 deletions
Original file line numberDiff line numberDiff line change
@@ -18,9 +18,7 @@
1818
//! Parquet integration tests
1919
use crate::parquet::utils::MetricsFinder;
2020
use arrow::array::Decimal128Array;
21-
use arrow::datatypes::{
22-
i256, IntervalDayTimeType, IntervalMonthDayNanoType, IntervalYearMonthType,
23-
};
21+
use arrow::datatypes::i256;
2422
use arrow::{
2523
array::{
2624
make_array, Array, ArrayRef, BinaryArray, BooleanArray, Date32Array, Date64Array,
@@ -36,10 +34,6 @@ use arrow::{
3634
record_batch::RecordBatch,
3735
util::pretty::pretty_format_batches,
3836
};
39-
use arrow_array::{
40-
IntervalDayTimeArray, IntervalMonthDayNanoArray, IntervalYearMonthArray,
41-
};
42-
use arrow_schema::IntervalUnit;
4337
use chrono::{Datelike, Duration, TimeDelta};
4438
use datafusion::{
4539
datasource::{provider_as_source, TableProvider},
@@ -92,7 +86,6 @@ enum Scenario {
9286
Time32Millisecond,
9387
Time64Nanosecond,
9488
Time64Microsecond,
95-
Interval,
9689
/// 7 Rows, for each i8, i16, i32, i64, u8, u16, u32, u64, f32, f64
9790
/// -MIN, -100, -1, 0, 1, 100, MAX
9891
NumericLimits,
@@ -921,71 +914,6 @@ fn make_dict_batch() -> RecordBatch {
921914
.unwrap()
922915
}
923916

924-
fn make_interval_batch(offset: i32) -> RecordBatch {
925-
let schema = Schema::new(vec![
926-
Field::new(
927-
"year_month",
928-
DataType::Interval(IntervalUnit::YearMonth),
929-
true,
930-
),
931-
Field::new("day_time", DataType::Interval(IntervalUnit::DayTime), true),
932-
Field::new(
933-
"month_day_nano",
934-
DataType::Interval(IntervalUnit::MonthDayNano),
935-
true,
936-
),
937-
]);
938-
let schema = Arc::new(schema);
939-
940-
let ym_arr = IntervalYearMonthArray::from(vec![
941-
Some(IntervalYearMonthType::make_value(1 + offset, 10 + offset)),
942-
Some(IntervalYearMonthType::make_value(2 + offset, 20 + offset)),
943-
Some(IntervalYearMonthType::make_value(3 + offset, 30 + offset)),
944-
None,
945-
Some(IntervalYearMonthType::make_value(5 + offset, 50 + offset)),
946-
]);
947-
948-
let dt_arr = IntervalDayTimeArray::from(vec![
949-
Some(IntervalDayTimeType::make_value(1 + offset, 10 + offset)),
950-
Some(IntervalDayTimeType::make_value(2 + offset, 20 + offset)),
951-
Some(IntervalDayTimeType::make_value(3 + offset, 30 + offset)),
952-
None,
953-
Some(IntervalDayTimeType::make_value(5 + offset, 50 + offset)),
954-
]);
955-
956-
// Not yet implemented, refer to:
957-
// https://github.com/apache/arrow-rs/blob/master/parquet/src/arrow/arrow_writer/mod.rs#L747
958-
let mdn_arr = IntervalMonthDayNanoArray::from(vec![
959-
Some(IntervalMonthDayNanoType::make_value(
960-
1 + offset,
961-
10 + offset,
962-
100 + (offset as i64),
963-
)),
964-
Some(IntervalMonthDayNanoType::make_value(
965-
2 + offset,
966-
20 + offset,
967-
200 + (offset as i64),
968-
)),
969-
Some(IntervalMonthDayNanoType::make_value(
970-
3 + offset,
971-
30 + offset,
972-
300 + (offset as i64),
973-
)),
974-
None,
975-
Some(IntervalMonthDayNanoType::make_value(
976-
5 + offset,
977-
50 + offset,
978-
500 + (offset as i64),
979-
)),
980-
]);
981-
982-
RecordBatch::try_new(
983-
schema,
984-
vec![Arc::new(ym_arr), Arc::new(dt_arr), Arc::new(mdn_arr)],
985-
)
986-
.unwrap()
987-
}
988-
989917
fn create_data_batch(scenario: Scenario) -> Vec<RecordBatch> {
990918
match scenario {
991919
Scenario::Boolean => {
@@ -1407,12 +1335,6 @@ fn create_data_batch(scenario: Scenario) -> Vec<RecordBatch> {
14071335
]),
14081336
]
14091337
}
1410-
Scenario::Interval => vec![
1411-
make_interval_batch(0),
1412-
make_interval_batch(1),
1413-
make_interval_batch(2),
1414-
make_interval_batch(3),
1415-
],
14161338
}
14171339
}
14181340

0 commit comments

Comments
 (0)