-
Notifications
You must be signed in to change notification settings - Fork 1.9k
Description
Describe the bug
I am attempting to optimize an inner join on two hive-partitioned parquet datasets lhs and rhs, joined on one column join_col. In a base working query, the left-hand side of the inner join is lhs itself and the right-hand side rhs filtered by some numeric columns. It just so happens for this data that equality on join_col implies equality of a less granular partition column join_col_partitioned. However, a modification to the query, in which the left-hand side lhs is first filtered to records with join_col_partitioned among the distinct values present in the right-hand side, results in a repeatable panic.
To Reproduce
I cannot provide the raw data but will provide all the details that I can, with some details obfuscated.
External tables
The tables lhs and rhs in question are created with
CREATE EXTERNAL TABLE lhs
STORED AS parquet
PARTITIONED BY (join_col_chk_first)
LOCATION '/LHS_ROOT/chunks/*/chunk.parquet';
and
CREATE EXTERNAL TABLE rhs
STORED AS parquet
PARTITIONED BY (join_col_chk_first, col1, col2_chk_first)
LOCATION '/RHS_ROOT/*/*/*/chunk.parquet'
The table lhs also contains columns keep_col and join_col, as well as others not referenced in the query; similarly rhs contains columns col2 and join_col. Both tables have the same number of records, 28914441.
Working query
The working base version of the query is
SELECT a.keep_col
FROM lhs AS a
INNER JOIN (
SELECT * FROM rhs
WHERE col1=7 AND col2>=0 AND col2<=25000
)
AS b
ON a.join_col = b.join_col;
and returns 375130 records, which is precisely the number of records in SELECT * FROM rhs WHERE col1=7 AND col2>=0 AND col2<=25000.
Modified panic-inducing query
The modification to the query that causes the panic is
SELECT a.keep_col
FROM (
SELECT * FROM lhs
WHERE join_col_partitioned IN (
SELECT DISTINCT join_col_partitioned
FROM rhs
WHERE col1=7 AND col2>=0 AND col2<=25000
)
) AS a
INNER JOIN (
SELECT * FROM rhs
WHERE col1=7 AND col2>=0 AND col2<=25000
)
AS b
ON a.join_col = b.join_col;
The nested query SELECT DISTINCT join_col_partitioned FROM rhs WHERE col1=7 AND col2>=0 AND col2<=25000 returns two distinct values of join_col_partitioned.
Running this query dependably produces a panic (repeated 5x or so times).
Logical and physical plans and backtraces
I have included the the output of EXPLAIN for both of these queries in the Additional context section, as well as the regular (RUST_BACKTRACE=1) and full (RUST_BACKTRACE=full) backtraces for the panic.
Expected behavior
I expected the second query to return the exact data as the first query (due to the aforementioned fact that for these datasets, equality on join_col implies equality on join_col_partitioned, and not to result in a panic.
Additional context
Environment
- Rocky 8
- DataFusion 34.0.0
- Arrow 49.0.0
Explanation for working query
Output of EXPLAIN SELECT a.keep_col FROM lhs AS a INNER JOIN (SELECT * FROM rhs WHERE col1=7 AND col2>=0 AND col2<=25000) AS b ON a.join_col = b.join_col;:
plan_type: [["logical_plan","physical_plan"]]
plan: [["Projection: a.keep_col
Inner Join: a.join_col = b.join_col
SubqueryAlias: a
TableScan: lhs projection=[keep_col, join_col]
SubqueryAlias: b
Projection: rhs.join_col
Filter: rhs.col2 >= Int64(0) AND rhs.col2 <= Int64(25000)
TableScan: rhs projection=[join_col, col2], full_filters=[CAST(rhs.col1 AS Utf8) = Utf8("7")], partial_filters=[rhs.col2 >= Int64(0), rhs.col2 <= Int64(25000)]","ProjectionExec: expr=[keep_col@0 as keep_col]
CoalesceBatchesExec: target_batch_size=8192
HashJoinExec: mode=Partitioned, join_type=Inner, on=[(join_col@1, join_col@0)]
CoalesceBatchesExec: target_batch_size=8192
RepartitionExec: partitioning=Hash([join_col@1], 96), input_partitions=96
ParquetExec: file_groups={96 groups: [[LHS_ROOT/join_col_chk_first=0/chunk.parquet:0..26277405], [LHS_ROOT/join_col_chk_first=0/chunk.parquet:26277405..52554810], [LHS_ROOT/join_col_chk_first=0/chunk.parquet:52554810..78832215], [LHS_ROOT/join_col_chk_first=0/chunk.parquet:78832215..95615950, LHS_ROOT/join_col_chk_first=1000000/chunk.parquet:0..9493670], [LHS_ROOT/join_col_chk_first=1000000/chunk.parquet:9493670..35771075], ...]}, projection=[keep_col, join_col]
CoalesceBatchesExec: target_batch_size=8192
RepartitionExec: partitioning=Hash([join_col@0], 96), input_partitions=96
ProjectionExec: expr=[join_col@0 as join_col]
CoalesceBatchesExec: target_batch_size=8192
FilterExec: col2@1 >= 0 AND col2@1 <= 25000
RepartitionExec: partitioning=RoundRobinBatch(96), input_partitions=81
ParquetExec: file_groups={81 groups: [[RHS_ROOT/join_col_chk_first=25000000/col1=7/col2_chk_first=1/chunk.parquet], [RHS_ROOT/join_col_chk_first=25000000/col1=7/col2_chk_first=10000001/chunk.parquet], [RHS_ROOT/join_col_chk_first=25000000/col1=7/col2_chk_first=1000001/chunk.parquet], [RHS_ROOT/join_col_chk_first=25000000/col1=7/col2_chk_first=11000001/chunk.parquet], [RHS_ROOT/join_col_chk_first=25000000/col1=7/col2_chk_first=12000001/chunk.parquet], ...]}, projection=[join_col, col2], predicate=CAST(col1@7 AS Utf8) = 7 AND col2@3 >= 0 AND col2@3 <= 25000, pruning_predicate=col2_max@0 >= 0 AND col2_min@1 <= 25000
Explanation for failing query
Output of EXPLAIN SELECT a.keep_col FROM (SELECT * FROM lhs WHERE join_col_partitioned IN (SELECT DISTINCT join_col_partitioned FROM rhs WHERE col1=7 AND col2>=0 AND col2<=25000)) AS a INNER JOIN (SELECT * FROM rhs WHERE col1=7 AND col2>=0 AND col2<=25000) AS b ON a.join_col = b.join_col;:
plan_type: [["logical_plan","physical_plan"]]
plan: [["Projection: a.keep_col
Inner Join: a.join_col = b.join_col
SubqueryAlias: a
Projection: lhs.keep_col, lhs.join_col
LeftSemi Join: lhs.join_col_chk_first = __correlated_sq_1.join_col_chk_first
TableScan: lhs projection=[keep_col, join_col, join_col_chk_first]
SubqueryAlias: __correlated_sq_1
Aggregate: groupBy=[[rhs.join_col_chk_first]], aggr=[[]]
Projection: rhs.join_col_chk_first
Filter: rhs.col2 >= Int64(0) AND rhs.col2 <= Int64(25000)
TableScan: rhs projection=[col2, join_col_chk_first], full_filters=[CAST(rhs.col1 AS Utf8) = Utf8("7")], partial_filters=[rhs.col2 >= Int64(0), rhs.col2 <= Int64(25000)]
SubqueryAlias: b
Projection: rhs.join_col
Filter: rhs.col2 >= Int64(0) AND rhs.col2 <= Int64(25000)
TableScan: rhs projection=[join_col, col2], full_filters=[CAST(rhs.col1 AS Utf8) = Utf8("7")], partial_filters=[rhs.col2 >= Int64(0), rhs.col2 <= Int64(25000)]","ProjectionExec: expr=[keep_col@0 as keep_col]
CoalesceBatchesExec: target_batch_size=8192
HashJoinExec: mode=Partitioned, join_type=Inner, on=[(join_col@1, join_col@0)]
CoalesceBatchesExec: target_batch_size=8192
RepartitionExec: partitioning=Hash([join_col@1], 96), input_partitions=96
ProjectionExec: expr=[keep_col@0 as keep_col, join_col@1 as join_col]
CoalesceBatchesExec: target_batch_size=8192
HashJoinExec: mode=Partitioned, join_type=LeftSemi, on=[(join_col_chk_first@2, join_col_chk_first@0)]
CoalesceBatchesExec: target_batch_size=8192
RepartitionExec: partitioning=Hash([join_col_chk_first@2], 96), input_partitions=96
ParquetExec: file_groups={96 groups: [[LHS_ROOT/join_col_chk_first=0/chunk.parquet:0..26277405], [LHS_ROOT/join_col_chk_first=0/chunk.parquet:26277405..52554810], [LHS_ROOT/join_col_chk_first=0/chunk.parquet:52554810..78832215], [LHS_ROOT/join_col_chk_first=0/chunk.parquet:78832215..95615950, LHS_ROOT/join_col_chk_first=1000000/chunk.parquet:0..9493670], [LHS_ROOT/join_col_chk_first=1000000/chunk.parquet:9493670..35771075], ...]}, projection=[keep_col, join_col, join_col_chk_first]
AggregateExec: mode=FinalPartitioned, gby=[join_col_chk_first@0 as join_col_chk_first], aggr=[]
CoalesceBatchesExec: target_batch_size=8192
RepartitionExec: partitioning=Hash([join_col_chk_first@0], 96), input_partitions=96
AggregateExec: mode=Partial, gby=[join_col_chk_first@0 as join_col_chk_first], aggr=[]
ProjectionExec: expr=[join_col_chk_first@1 as join_col_chk_first]
CoalesceBatchesExec: target_batch_size=8192
FilterExec: col2@0 >= 0 AND col2@0 <= 25000
RepartitionExec: partitioning=RoundRobinBatch(96), input_partitions=81
ParquetExec: file_groups={81 groups: [[RHS_ROOT/join_col_chk_first=25000000/col1=7/col2_chk_first=1/chunk.parquet], [RHS_ROOT/join_col_chk_first=25000000/col1=7/col2_chk_first=10000001/chunk.parquet], [RHS_ROOT/join_col_chk_first=25000000/col1=7/col2_chk_first=1000001/chunk.parquet], [RHS_ROOT/join_col_chk_first=25000000/col1=7/col2_chk_first=11000001/chunk.parquet], [RHS_ROOT/join_col_chk_first=25000000/col1=7/col2_chk_first=12000001/chunk.parquet], ...]}, projection=[col2, join_col_chk_first], predicate=CAST(col1@7 AS Utf8) = 7 AND col2@3 >= 0 AND col2@3 <= 25000, pruning_predicate=col2_max@0 >= 0 AND col2_min@1 <= 25000
CoalesceBatchesExec: target_batch_size=8192
RepartitionExec: partitioning=Hash([join_col@0], 96), input_partitions=96
ProjectionExec: expr=[join_col@0 as join_col]
CoalesceBatchesExec: target_batch_size=8192
FilterExec: col2@1 >= 0 AND col2@1 <= 25000
RepartitionExec: partitioning=RoundRobinBatch(96), input_partitions=81
ParquetExec: file_groups={81 groups: [[RHS_ROOT/join_col_chk_first=25000000/col1=7/col2_chk_first=1/chunk.parquet], [RHS_ROOT/join_col_chk_first=25000000/col1=7/col2_chk_first=10000001/chunk.parquet], [RHS_ROOT/join_col_chk_first=25000000/col1=7/col2_chk_first=1000001/chunk.parquet], [RHS_ROOT/join_col_chk_first=25000000/col1=7/col2_chk_first=11000001/chunk.parquet], [RHS_ROOT/join_col_chk_first=25000000/col1=7/col2_chk_first=12000001/chunk.parquet], ...]}, projection=[join_col, col2], predicate=CAST(col1@7 AS Utf8) = 7 AND col2@3 >= 0 AND col2@3 <= 25000, pruning_predicate=col2_max@0 >= 0 AND col2_min@1 <= 25000
"]]
Regular backtrace
Mar 25 14:56:41 <hostname> sh[942819]: thread 'tokio-runtime-worker' panicked at /home/username/.cargo/registry/src/index.crates.io-6f17d22bba15001f/arrow-data-49.0.0/src/transform/utils.rs:42:56:
Mar 25 14:56:41 <hostname> sh[942819]: offset overflow
Mar 25 14:56:41 <hostname> sh[942819]: stack backtrace:
Mar 25 14:56:41 <hostname> sh[942819]: 0: rust_begin_unwind
Mar 25 14:56:41 <hostname> sh[942819]: at /rustc/07dca489ac2d933c78d3c5158e3f43beefeb02ce/library/std/src/panicking.rs:645:5
Mar 25 14:56:41 <hostname> sh[942819]: 1: core::panicking::panic_fmt
Mar 25 14:56:41 <hostname> sh[942819]: at /rustc/07dca489ac2d933c78d3c5158e3f43beefeb02ce/library/core/src/panicking.rs:72:14
Mar 25 14:56:41 <hostname> sh[942819]: 2: core::panicking::panic_display
Mar 25 14:56:41 <hostname> sh[942819]: at /rustc/07dca489ac2d933c78d3c5158e3f43beefeb02ce/library/core/src/panicking.rs:196:5
Mar 25 14:56:41 <hostname> sh[942819]: 3: core::panicking::panic_str
Mar 25 14:56:41 <hostname> sh[942819]: at /rustc/07dca489ac2d933c78d3c5158e3f43beefeb02ce/library/core/src/panicking.rs:171:5
Mar 25 14:56:41 <hostname> sh[942819]: 4: core::option::expect_failed
Mar 25 14:56:41 <hostname> sh[942819]: at /rustc/07dca489ac2d933c78d3c5158e3f43beefeb02ce/library/core/src/option.rs:1980:5
Mar 25 14:56:41 <hostname> sh[942819]: 5: arrow_data::transform::utils::extend_offsets
Mar 25 14:56:41 <hostname> sh[942819]: 6: arrow_data::transform::variable_size::build_extend::{{closure}}
Mar 25 14:56:41 <hostname> sh[942819]: 7: arrow_data::transform::MutableArrayData::extend
Mar 25 14:56:41 <hostname> sh[942819]: 8: arrow_select::concat::concat_fallback
Mar 25 14:56:41 <hostname> sh[942819]: 9: arrow_select::concat::concat
Mar 25 14:56:41 <hostname> sh[942819]: 10: arrow_select::concat::concat_batches
Mar 25 14:56:41 <hostname> sh[942819]: 11: <futures_util::future::future::map::Map<Fut,F> as core::future::future::Future>::poll
Mar 25 14:56:41 <hostname> sh[942819]: 12: <futures_util::future::future::shared::Shared<Fut> as core::future::future::Future>::poll
Mar 25 14:56:41 <hostname> sh[942819]: 13: datafusion_physical_plan::joins::utils::OnceFut<T>::get
Mar 25 14:56:41 <hostname> sh[942819]: 14: <datafusion_physical_plan::joins::hash_join::HashJoinStream as futures_core::stream::Stream>::poll_next
Mar 25 14:56:41 <hostname> sh[942819]: 15: <datafusion_physical_plan::coalesce_batches::CoalesceBatchesStream as futures_core::stream::Stream>::poll_next
Mar 25 14:56:41 <hostname> sh[942819]: 16: datafusion_physical_plan::repartition::RepartitionExec::pull_from_input::{{closure}}
Mar 25 14:56:41 <hostname> sh[942819]: 17: tokio::loom::std::unsafe_cell::UnsafeCell<T>::with_mut
Mar 25 14:56:41 <hostname> sh[942819]: 18: tokio::runtime::task::core::Core<T,S>::poll
Mar 25 14:56:41 <hostname> sh[942819]: 19: tokio::runtime::task::harness::Harness<T,S>::poll
Mar 25 14:56:41 <hostname> sh[942819]: 20: tokio::runtime::scheduler::multi_thread::worker::Context::run_task
Mar 25 14:56:41 <hostname> sh[942819]: 21: tokio::runtime::scheduler::multi_thread::worker::run
Mar 25 14:56:41 <hostname> sh[942819]: 22: tokio::loom::std::unsafe_cell::UnsafeCell<T>::with_mut
Mar 25 14:56:41 <hostname> sh[942819]: 23: tokio::runtime::task::core::Core<T,S>::poll
Mar 25 14:56:41 <hostname> sh[942819]: 24: tokio::runtime::task::harness::Harness<T,S>::poll
Mar 25 14:56:41 <hostname> sh[942819]: 25: tokio::runtime::blocking::pool::Inner::run
Mar 25 14:56:41 <hostname> sh[942819]: note: Some details are omitted, run with `RUST_BACKTRACE=full` for a verbose backtrace.
Full backtrace
Mar 25 15:02:03 <hostname> sh[957703]: thread 'tokio-runtime-worker' panicked at /home/username/.cargo/registry/src/index.crates.io-6f17d22bba15001f/arrow-data-49.0.0/src/transform/utils.rs:42:56:
Mar 25 15:02:03 <hostname> sh[957703]: offset overflow
Mar 25 15:02:03 <hostname> sh[957703]: stack backtrace:
Mar 25 15:02:03 <hostname> sh[957703]: thread 'tokio-runtime-worker' panicked at /home/username/.cargo/registry/src/index.crates.io-6f17d22bba15001f/arrow-data-49.0.0/src/transform/utils.rs:42:56:
Mar 25 15:02:03 <hostname> sh[957703]: offset overflow
Mar 25 15:02:03 <hostname> sh[957703]: 0: 0x560a5f972a96 - std::backtrace_rs::backtrace::libunwind::trace::hbee8a7973eeb6c93
Mar 25 15:02:03 <hostname> sh[957703]: at /rustc/07dca489ac2d933c78d3c5158e3f43beefeb02ce/library/std/src/../../backtrace/src/backtrace/libunwind.rs:104:5
Mar 25 15:02:03 <hostname> sh[957703]: 1: 0x560a5f972a96 - std::backtrace_rs::backtrace::trace_unsynchronized::hc8ac75eea3aa6899
Mar 25 15:02:03 <hostname> sh[957703]: at /rustc/07dca489ac2d933c78d3c5158e3f43beefeb02ce/library/std/src/../../backtrace/src/backtrace/mod.rs:66:5
Mar 25 15:02:03 <hostname> sh[957703]: 2: 0x560a5f972a96 - std::sys_common::backtrace::_print_fmt::hc7f3e3b5298b1083
Mar 25 15:02:03 <hostname> sh[957703]: at /rustc/07dca489ac2d933c78d3c5158e3f43beefeb02ce/library/std/src/sys_common/backtrace.rs:68:5
Mar 25 15:02:03 <hostname> sh[957703]: 3: 0x560a5f972a96 - <std::sys_common::backtrace::_print::DisplayBacktrace as core::fmt::Display>::fmt::hbb235daedd7c6190
Mar 25 15:02:03 <hostname> sh[957703]: at /rustc/07dca489ac2d933c78d3c5158e3f43beefeb02ce/library/std/src/sys_common/backtrace.rs:44:22
Mar 25 15:02:03 <hostname> sh[957703]: 4: 0x560a5f99fe00 - core::fmt::rt::Argument::fmt::h76c38a80d925a410
Mar 25 15:02:03 <hostname> sh[957703]: at /rustc/07dca489ac2d933c78d3c5158e3f43beefeb02ce/library/core/src/fmt/rt.rs:142:9
Mar 25 15:02:03 <hostname> sh[957703]: 5: 0x560a5f99fe00 - core::fmt::write::h3ed6aeaa977c8e45
Mar 25 15:02:03 <hostname> sh[957703]: at /rustc/07dca489ac2d933c78d3c5158e3f43beefeb02ce/library/core/src/fmt/mod.rs:1120:17
Mar 25 15:02:03 <hostname> sh[957703]: 6: 0x560a5f96f9ff - std::io::Write::write_fmt::h78b18af5775fedb5
Mar 25 15:02:03 <hostname> sh[957703]: at /rustc/07dca489ac2d933c78d3c5158e3f43beefeb02ce/library/std/src/io/mod.rs:1810:15
Mar 25 15:02:03 <hostname> sh[957703]: 7: 0x560a5f972874 - std::sys_common::backtrace::_print::h5d645a07e0fcfdbb
Mar 25 15:02:03 <hostname> sh[957703]: at /rustc/07dca489ac2d933c78d3c5158e3f43beefeb02ce/library/std/src/sys_common/backtrace.rs:47:5
Mar 25 15:02:03 <hostname> sh[957703]: 8: 0x560a5f972874 - std::sys_common::backtrace::print::h85035a511aafe7a8
Mar 25 15:02:03 <hostname> sh[957703]: at /rustc/07dca489ac2d933c78d3c5158e3f43beefeb02ce/library/std/src/sys_common/backtrace.rs:34:9
Mar 25 15:02:03 <hostname> sh[957703]: 9: 0x560a5f9740f7 - std::panicking::default_hook::{{closure}}::hcce8cea212785a25
Mar 25 15:02:03 <hostname> sh[957703]: 10: 0x560a5f973e59 - std::panicking::default_hook::hf5fcb0f213fe709a
Mar 25 15:02:03 <hostname> sh[957703]: at /rustc/07dca489ac2d933c78d3c5158e3f43beefeb02ce/library/std/src/panicking.rs:292:9
Mar 25 15:02:03 <hostname> sh[957703]: 11: 0x560a5f974588 - std::panicking::rust_panic_with_hook::h095fccf1dc9379ee
Mar 25 15:02:03 <hostname> sh[957703]: at /rustc/07dca489ac2d933c78d3c5158e3f43beefeb02ce/library/std/src/panicking.rs:779:13
Mar 25 15:02:03 <hostname> sh[957703]: 12: 0x560a5f974462 - std::panicking::begin_panic_handler::{{closure}}::h032ba12139b353db
Mar 25 15:02:03 <hostname> sh[957703]: at /rustc/07dca489ac2d933c78d3c5158e3f43beefeb02ce/library/std/src/panicking.rs:657:13
Mar 25 15:02:03 <hostname> sh[957703]: 13: 0x560a5f972f96 - std::sys_common::backtrace::__rust_end_short_backtrace::h9259bc2ff8fd0f76
Mar 25 15:02:03 <hostname> sh[957703]: at /rustc/07dca489ac2d933c78d3c5158e3f43beefeb02ce/library/std/src/sys_common/backtrace.rs:171:18
Mar 25 15:02:03 <hostname> sh[957703]: 14: 0x560a5f9741c0 - rust_begin_unwind
Mar 25 15:02:03 <hostname> sh[957703]: at /rustc/07dca489ac2d933c78d3c5158e3f43beefeb02ce/library/std/src/panicking.rs:645:5
Mar 25 15:02:03 <hostname> sh[957703]: 15: 0x560a5d65e4e5 - core::panicking::panic_fmt::h784f20a50eaab275
Mar 25 15:02:03 <hostname> sh[957703]: at /rustc/07dca489ac2d933c78d3c5158e3f43beefeb02ce/library/core/src/panicking.rs:72:14
Mar 25 15:02:03 <hostname> sh[957703]: 16: 0x560a5d65e4a3 - core::panicking::panic_display::h251010ce5e0560d5
Mar 25 15:02:03 <hostname> sh[957703]: at /rustc/07dca489ac2d933c78d3c5158e3f43beefeb02ce/library/core/src/panicking.rs:196:5
Mar 25 15:02:03 <hostname> sh[957703]: 17: 0x560a5d65e4a3 - core::panicking::panic_str::h6f89534c81f0edc4
Mar 25 15:02:03 <hostname> sh[957703]: at /rustc/07dca489ac2d933c78d3c5158e3f43beefeb02ce/library/core/src/panicking.rs:171:5
Mar 25 15:02:03 <hostname> sh[957703]: 18: 0x560a5d65e4a3 - core::option::expect_failed::hc85eb6037a3050f7
Mar 25 15:02:03 <hostname> sh[957703]: at /rustc/07dca489ac2d933c78d3c5158e3f43beefeb02ce/library/core/src/option.rs:1980:5
Mar 25 15:02:03 <hostname> sh[957703]: 19: 0x560a5f924a82 - arrow_data::transform::utils::extend_offsets::hdf2f06f3dfb426cd
Mar 25 15:02:03 <hostname> sh[957703]: 20: 0x560a5f9279c6 - arrow_data::transform::variable_size::build_extend::{{closure}}::h31904a480ae969de
Mar 25 15:02:03 <hostname> sh[957703]: 21: 0x560a5f9183b5 - arrow_data::transform::MutableArrayData::extend::hde96a1dcf5382372
Mar 25 15:02:03 <hostname> sh[957703]: 22: 0x560a5f89081f - arrow_select::concat::concat_fallback::h58848d6223cb12c9
Mar 25 15:02:03 <hostname> sh[957703]: 23: 0x560a5f89049e - arrow_select::concat::concat::h2c60d37437746b3e
Mar 25 15:02:03 <hostname> sh[957703]: 24: 0x560a5e101518 - arrow_select::concat::concat_batches::hbc4b4e1ab862a29d
Mar 25 15:02:03 <hostname> sh[957703]: 25: 0x560a5e178464 - <futures_util::future::future::map::Map<Fut,F> as core::future::future::Future>::poll::h3e1df0aac2ab9d77
Mar 25 15:02:03 <hostname> sh[957703]: 26: 0x560a5e02f8f2 - <futures_util::future::future::shared::Shared<Fut> as core::future::future::Future>::poll::h7c9852d5bb7e9534
Mar 25 15:02:03 <hostname> sh[957703]: 27: 0x560a5e16e2bf - datafusion_physical_plan::joins::utils::OnceFut<T>::get::hd3b2e25c093ada6d
Mar 25 15:02:03 <hostname> sh[957703]: 28: 0x560a5e195dd9 - <datafusion_physical_plan::joins::hash_join::HashJoinStream as futures_core::stream::Stream>::poll_next::h9c658418141705f6
Mar 25 15:02:03 <hostname> sh[957703]: 29: 0x560a5e18a696 - <datafusion_physical_plan::coalesce_batches::CoalesceBatchesStream as futures_core::stream::Stream>::poll_next::h219998518ccf857d
Mar 25 15:02:03 <hostname> sh[957703]: 30: 0x560a5e00bada - datafusion_physical_plan::repartition::RepartitionExec::pull_from_input::{{closure}}::h1ed64d4cd49cce5e
Mar 25 15:02:03 <hostname> sh[957703]: 31: 0x560a5e000993 - tokio::loom::std::unsafe_cell::UnsafeCell<T>::with_mut::h38ddf14813d65bab
Mar 25 15:02:03 <hostname> sh[957703]: 32: 0x560a5e03d6ee - tokio::runtime::task::core::Core<T,S>::poll::h3df6db59b96cdc95
Mar 25 15:02:03 <hostname> sh[957703]: 33: 0x560a5e040779 - tokio::runtime::task::harness::Harness<T,S>::poll::h56978e4fa718a3b4
Mar 25 15:02:03 <hostname> sh[957703]: 34: 0x560a5f0e3ad8 - tokio::runtime::scheduler::multi_thread::worker::Context::run_task::hb25f4c4bb31b0772
Mar 25 15:02:03 <hostname> sh[957703]: 35: 0x560a5f0e35b5 - tokio::runtime::scheduler::multi_thread::worker::run::h9904ff19a8f264dd
Mar 25 15:02:03 <hostname> sh[957703]: 36: 0x560a5f0dfcd2 - tokio::loom::std::unsafe_cell::UnsafeCell<T>::with_mut::h76fc62ff172daba6
Mar 25 15:02:03 <hostname> sh[957703]: 37: 0x560a5f0f2af9 - tokio::runtime::task::core::Core<T,S>::poll::h256060c96f596608
Mar 25 15:02:03 <hostname> sh[957703]: 38: 0x560a5f0ded9a - tokio::runtime::task::harness::Harness<T,S>::poll::h0b57f37a1c3a1ec2
Mar 25 15:02:03 <hostname> sh[957703]: 39: 0x560a5f0ee780 - tokio::runtime::blocking::pool::Inner::run::ha371e4f6189503f2
Mar 25 15:02:03 <hostname> sh[957703]: 40: 0x560a5f0e84fc - std::sys_common::backtrace::__rust_begin_short_backtrace::h25663fb0bf4a4475
Mar 25 15:02:03 <hostname> sh[957703]: 41: 0x560a5f0f0a79 - core::ops::function::FnOnce::call_once{{vtable.shim}}::ha0885a2c26e1b6a6
Mar 25 15:02:03 <hostname> sh[957703]: 42: 0x560a5f9783d5 - <alloc::boxed::Box<F,A> as core::ops::function::FnOnce<Args>>::call_once::h12de4fc57affb195
Mar 25 15:02:03 <hostname> sh[957703]: at /rustc/07dca489ac2d933c78d3c5158e3f43beefeb02ce/library/alloc/src/boxed.rs:2015:9
Mar 25 15:02:03 <hostname> sh[957703]: 43: 0x560a5f9783d5 - <alloc::boxed::Box<F,A> as core::ops::function::FnOnce<Args>>::call_once::h3c619f45059d5cf1
Mar 25 15:02:03 <hostname> sh[957703]: at /rustc/07dca489ac2d933c78d3c5158e3f43beefeb02ce/library/alloc/src/boxed.rs:2015:9
Mar 25 15:02:03 <hostname> sh[957703]: 44: 0x560a5f9783d5 - std::sys::unix::thread::Thread::new::thread_start::hbac657605e4b7389
Mar 25 15:02:03 <hostname> sh[957703]: at /rustc/07dca489ac2d933c78d3c5158e3f43beefeb02ce/library/std/src/sys/unix/thread.rs:108:17
Mar 25 15:02:03 <hostname> sh[957703]: 45: 0x7f094d03d1ca - start_thread
Mar 25 15:02:03 <hostname> sh[957703]: 46: 0x7f094c723e73 - __clone
Mar 25 15:02:03 <hostname> sh[957703]: 47: 0x0 - <unknown>