Skip to content

inner join involving hive-partitioned parquet dataset and filters on LHS and RHS causes panic #9797

@jwimberl

Description

@jwimberl

Describe the bug

I am attempting to optimize an inner join on two hive-partitioned parquet datasets lhs and rhs, joined on one column join_col. In a base working query, the left-hand side of the inner join is lhs itself and the right-hand side rhs filtered by some numeric columns. It just so happens for this data that equality on join_col implies equality of a less granular partition column join_col_partitioned. However, a modification to the query, in which the left-hand side lhs is first filtered to records with join_col_partitioned among the distinct values present in the right-hand side, results in a repeatable panic.

To Reproduce

I cannot provide the raw data but will provide all the details that I can, with some details obfuscated.

External tables

The tables lhs and rhs in question are created with

CREATE EXTERNAL TABLE lhs
STORED AS parquet
PARTITIONED BY (join_col_chk_first)
LOCATION '/LHS_ROOT/chunks/*/chunk.parquet';

and

CREATE EXTERNAL TABLE rhs
STORED AS parquet
PARTITIONED BY (join_col_chk_first, col1, col2_chk_first)
LOCATION '/RHS_ROOT/*/*/*/chunk.parquet'

The table lhs also contains columns keep_col and join_col, as well as others not referenced in the query; similarly rhs contains columns col2 and join_col. Both tables have the same number of records, 28914441.

Working query

The working base version of the query is

SELECT a.keep_col
FROM lhs AS a
INNER JOIN (
  SELECT * FROM rhs
  WHERE col1=7 AND col2>=0 AND col2<=25000
)
AS b
ON a.join_col = b.join_col;

and returns 375130 records, which is precisely the number of records in SELECT * FROM rhs WHERE col1=7 AND col2>=0 AND col2<=25000.

Modified panic-inducing query

The modification to the query that causes the panic is

SELECT a.keep_col
FROM (
  SELECT * FROM lhs
  WHERE join_col_partitioned IN (
    SELECT DISTINCT join_col_partitioned
    FROM rhs
    WHERE col1=7 AND col2>=0 AND col2<=25000
  )
) AS a
INNER JOIN (
  SELECT * FROM rhs
  WHERE col1=7 AND col2>=0 AND col2<=25000
)
AS b
ON a.join_col = b.join_col;

The nested query SELECT DISTINCT join_col_partitioned FROM rhs WHERE col1=7 AND col2>=0 AND col2<=25000 returns two distinct values of join_col_partitioned.

Running this query dependably produces a panic (repeated 5x or so times).

Logical and physical plans and backtraces

I have included the the output of EXPLAIN for both of these queries in the Additional context section, as well as the regular (RUST_BACKTRACE=1) and full (RUST_BACKTRACE=full) backtraces for the panic.

Expected behavior

I expected the second query to return the exact data as the first query (due to the aforementioned fact that for these datasets, equality on join_col implies equality on join_col_partitioned, and not to result in a panic.

Additional context

Environment

  • Rocky 8
  • DataFusion 34.0.0
  • Arrow 49.0.0

Explanation for working query

Output of EXPLAIN SELECT a.keep_col FROM lhs AS a INNER JOIN (SELECT * FROM rhs WHERE col1=7 AND col2>=0 AND col2<=25000) AS b ON a.join_col = b.join_col;:

plan_type: [["logical_plan","physical_plan"]]
plan: [["Projection: a.keep_col
  Inner Join: a.join_col = b.join_col
    SubqueryAlias: a
      TableScan: lhs projection=[keep_col, join_col]
    SubqueryAlias: b
      Projection: rhs.join_col
        Filter: rhs.col2 >= Int64(0) AND rhs.col2 <= Int64(25000)
          TableScan: rhs projection=[join_col, col2], full_filters=[CAST(rhs.col1 AS Utf8) = Utf8("7")], partial_filters=[rhs.col2 >= Int64(0), rhs.col2 <= Int64(25000)]","ProjectionExec: expr=[keep_col@0 as keep_col]
  CoalesceBatchesExec: target_batch_size=8192
    HashJoinExec: mode=Partitioned, join_type=Inner, on=[(join_col@1, join_col@0)]
      CoalesceBatchesExec: target_batch_size=8192
        RepartitionExec: partitioning=Hash([join_col@1], 96), input_partitions=96
          ParquetExec: file_groups={96 groups: [[LHS_ROOT/join_col_chk_first=0/chunk.parquet:0..26277405], [LHS_ROOT/join_col_chk_first=0/chunk.parquet:26277405..52554810], [LHS_ROOT/join_col_chk_first=0/chunk.parquet:52554810..78832215], [LHS_ROOT/join_col_chk_first=0/chunk.parquet:78832215..95615950, LHS_ROOT/join_col_chk_first=1000000/chunk.parquet:0..9493670], [LHS_ROOT/join_col_chk_first=1000000/chunk.parquet:9493670..35771075], ...]}, projection=[keep_col, join_col]
      CoalesceBatchesExec: target_batch_size=8192
        RepartitionExec: partitioning=Hash([join_col@0], 96), input_partitions=96
          ProjectionExec: expr=[join_col@0 as join_col]
            CoalesceBatchesExec: target_batch_size=8192
              FilterExec: col2@1 >= 0 AND col2@1 <= 25000
                RepartitionExec: partitioning=RoundRobinBatch(96), input_partitions=81
                  ParquetExec: file_groups={81 groups: [[RHS_ROOT/join_col_chk_first=25000000/col1=7/col2_chk_first=1/chunk.parquet], [RHS_ROOT/join_col_chk_first=25000000/col1=7/col2_chk_first=10000001/chunk.parquet], [RHS_ROOT/join_col_chk_first=25000000/col1=7/col2_chk_first=1000001/chunk.parquet], [RHS_ROOT/join_col_chk_first=25000000/col1=7/col2_chk_first=11000001/chunk.parquet], [RHS_ROOT/join_col_chk_first=25000000/col1=7/col2_chk_first=12000001/chunk.parquet], ...]}, projection=[join_col, col2], predicate=CAST(col1@7 AS Utf8) = 7 AND col2@3 >= 0 AND col2@3 <= 25000, pruning_predicate=col2_max@0 >= 0 AND col2_min@1 <= 25000

Explanation for failing query

Output of EXPLAIN SELECT a.keep_col FROM (SELECT * FROM lhs WHERE join_col_partitioned IN (SELECT DISTINCT join_col_partitioned FROM rhs WHERE col1=7 AND col2>=0 AND col2<=25000)) AS a INNER JOIN (SELECT * FROM rhs WHERE col1=7 AND col2>=0 AND col2<=25000) AS b ON a.join_col = b.join_col;:

  plan_type: [["logical_plan","physical_plan"]]
  plan: [["Projection: a.keep_col
    Inner Join: a.join_col = b.join_col
      SubqueryAlias: a
        Projection: lhs.keep_col, lhs.join_col
          LeftSemi Join: lhs.join_col_chk_first = __correlated_sq_1.join_col_chk_first
            TableScan: lhs projection=[keep_col, join_col, join_col_chk_first]
            SubqueryAlias: __correlated_sq_1
              Aggregate: groupBy=[[rhs.join_col_chk_first]], aggr=[[]]
                Projection: rhs.join_col_chk_first
                  Filter: rhs.col2 >= Int64(0) AND rhs.col2 <= Int64(25000)
                    TableScan: rhs projection=[col2, join_col_chk_first], full_filters=[CAST(rhs.col1 AS Utf8) = Utf8("7")], partial_filters=[rhs.col2 >= Int64(0), rhs.col2 <= Int64(25000)]
      SubqueryAlias: b
        Projection: rhs.join_col
          Filter: rhs.col2 >= Int64(0) AND rhs.col2 <= Int64(25000)
            TableScan: rhs projection=[join_col, col2], full_filters=[CAST(rhs.col1 AS Utf8) = Utf8("7")], partial_filters=[rhs.col2 >= Int64(0), rhs.col2 <= Int64(25000)]","ProjectionExec: expr=[keep_col@0 as keep_col]
    CoalesceBatchesExec: target_batch_size=8192
      HashJoinExec: mode=Partitioned, join_type=Inner, on=[(join_col@1, join_col@0)]
        CoalesceBatchesExec: target_batch_size=8192
          RepartitionExec: partitioning=Hash([join_col@1], 96), input_partitions=96
            ProjectionExec: expr=[keep_col@0 as keep_col, join_col@1 as join_col]
              CoalesceBatchesExec: target_batch_size=8192
                HashJoinExec: mode=Partitioned, join_type=LeftSemi, on=[(join_col_chk_first@2, join_col_chk_first@0)]
                  CoalesceBatchesExec: target_batch_size=8192
                    RepartitionExec: partitioning=Hash([join_col_chk_first@2], 96), input_partitions=96
                      ParquetExec: file_groups={96 groups: [[LHS_ROOT/join_col_chk_first=0/chunk.parquet:0..26277405], [LHS_ROOT/join_col_chk_first=0/chunk.parquet:26277405..52554810], [LHS_ROOT/join_col_chk_first=0/chunk.parquet:52554810..78832215], [LHS_ROOT/join_col_chk_first=0/chunk.parquet:78832215..95615950, LHS_ROOT/join_col_chk_first=1000000/chunk.parquet:0..9493670], [LHS_ROOT/join_col_chk_first=1000000/chunk.parquet:9493670..35771075], ...]}, projection=[keep_col, join_col, join_col_chk_first]
                  AggregateExec: mode=FinalPartitioned, gby=[join_col_chk_first@0 as join_col_chk_first], aggr=[]
                    CoalesceBatchesExec: target_batch_size=8192
                      RepartitionExec: partitioning=Hash([join_col_chk_first@0], 96), input_partitions=96
                        AggregateExec: mode=Partial, gby=[join_col_chk_first@0 as join_col_chk_first], aggr=[]
                          ProjectionExec: expr=[join_col_chk_first@1 as join_col_chk_first]
                            CoalesceBatchesExec: target_batch_size=8192
                              FilterExec: col2@0 >= 0 AND col2@0 <= 25000
                                RepartitionExec: partitioning=RoundRobinBatch(96), input_partitions=81
                                  ParquetExec: file_groups={81 groups: [[RHS_ROOT/join_col_chk_first=25000000/col1=7/col2_chk_first=1/chunk.parquet], [RHS_ROOT/join_col_chk_first=25000000/col1=7/col2_chk_first=10000001/chunk.parquet], [RHS_ROOT/join_col_chk_first=25000000/col1=7/col2_chk_first=1000001/chunk.parquet], [RHS_ROOT/join_col_chk_first=25000000/col1=7/col2_chk_first=11000001/chunk.parquet], [RHS_ROOT/join_col_chk_first=25000000/col1=7/col2_chk_first=12000001/chunk.parquet], ...]}, projection=[col2, join_col_chk_first], predicate=CAST(col1@7 AS Utf8) = 7 AND col2@3 >= 0 AND col2@3 <= 25000, pruning_predicate=col2_max@0 >= 0 AND col2_min@1 <= 25000
        CoalesceBatchesExec: target_batch_size=8192
          RepartitionExec: partitioning=Hash([join_col@0], 96), input_partitions=96
            ProjectionExec: expr=[join_col@0 as join_col]
              CoalesceBatchesExec: target_batch_size=8192
                FilterExec: col2@1 >= 0 AND col2@1 <= 25000
                  RepartitionExec: partitioning=RoundRobinBatch(96), input_partitions=81
                    ParquetExec: file_groups={81 groups: [[RHS_ROOT/join_col_chk_first=25000000/col1=7/col2_chk_first=1/chunk.parquet], [RHS_ROOT/join_col_chk_first=25000000/col1=7/col2_chk_first=10000001/chunk.parquet], [RHS_ROOT/join_col_chk_first=25000000/col1=7/col2_chk_first=1000001/chunk.parquet], [RHS_ROOT/join_col_chk_first=25000000/col1=7/col2_chk_first=11000001/chunk.parquet], [RHS_ROOT/join_col_chk_first=25000000/col1=7/col2_chk_first=12000001/chunk.parquet], ...]}, projection=[join_col, col2], predicate=CAST(col1@7 AS Utf8) = 7 AND col2@3 >= 0 AND col2@3 <= 25000, pruning_predicate=col2_max@0 >= 0 AND col2_min@1 <= 25000
  "]]

Regular backtrace

Mar 25 14:56:41 <hostname> sh[942819]: thread 'tokio-runtime-worker' panicked at /home/username/.cargo/registry/src/index.crates.io-6f17d22bba15001f/arrow-data-49.0.0/src/transform/utils.rs:42:56:
Mar 25 14:56:41 <hostname> sh[942819]: offset overflow
Mar 25 14:56:41 <hostname> sh[942819]: stack backtrace:
Mar 25 14:56:41 <hostname> sh[942819]:    0: rust_begin_unwind
Mar 25 14:56:41 <hostname> sh[942819]:              at /rustc/07dca489ac2d933c78d3c5158e3f43beefeb02ce/library/std/src/panicking.rs:645:5
Mar 25 14:56:41 <hostname> sh[942819]:    1: core::panicking::panic_fmt
Mar 25 14:56:41 <hostname> sh[942819]:              at /rustc/07dca489ac2d933c78d3c5158e3f43beefeb02ce/library/core/src/panicking.rs:72:14
Mar 25 14:56:41 <hostname> sh[942819]:    2: core::panicking::panic_display
Mar 25 14:56:41 <hostname> sh[942819]:              at /rustc/07dca489ac2d933c78d3c5158e3f43beefeb02ce/library/core/src/panicking.rs:196:5
Mar 25 14:56:41 <hostname> sh[942819]:    3: core::panicking::panic_str
Mar 25 14:56:41 <hostname> sh[942819]:              at /rustc/07dca489ac2d933c78d3c5158e3f43beefeb02ce/library/core/src/panicking.rs:171:5
Mar 25 14:56:41 <hostname> sh[942819]:    4: core::option::expect_failed
Mar 25 14:56:41 <hostname> sh[942819]:              at /rustc/07dca489ac2d933c78d3c5158e3f43beefeb02ce/library/core/src/option.rs:1980:5
Mar 25 14:56:41 <hostname> sh[942819]:    5: arrow_data::transform::utils::extend_offsets
Mar 25 14:56:41 <hostname> sh[942819]:    6: arrow_data::transform::variable_size::build_extend::{{closure}}
Mar 25 14:56:41 <hostname> sh[942819]:    7: arrow_data::transform::MutableArrayData::extend
Mar 25 14:56:41 <hostname> sh[942819]:    8: arrow_select::concat::concat_fallback
Mar 25 14:56:41 <hostname> sh[942819]:    9: arrow_select::concat::concat
Mar 25 14:56:41 <hostname> sh[942819]:   10: arrow_select::concat::concat_batches
Mar 25 14:56:41 <hostname> sh[942819]:   11: <futures_util::future::future::map::Map<Fut,F> as core::future::future::Future>::poll
Mar 25 14:56:41 <hostname> sh[942819]:   12: <futures_util::future::future::shared::Shared<Fut> as core::future::future::Future>::poll
Mar 25 14:56:41 <hostname> sh[942819]:   13: datafusion_physical_plan::joins::utils::OnceFut<T>::get
Mar 25 14:56:41 <hostname> sh[942819]:   14: <datafusion_physical_plan::joins::hash_join::HashJoinStream as futures_core::stream::Stream>::poll_next
Mar 25 14:56:41 <hostname> sh[942819]:   15: <datafusion_physical_plan::coalesce_batches::CoalesceBatchesStream as futures_core::stream::Stream>::poll_next
Mar 25 14:56:41 <hostname> sh[942819]:   16: datafusion_physical_plan::repartition::RepartitionExec::pull_from_input::{{closure}}
Mar 25 14:56:41 <hostname> sh[942819]:   17: tokio::loom::std::unsafe_cell::UnsafeCell<T>::with_mut
Mar 25 14:56:41 <hostname> sh[942819]:   18: tokio::runtime::task::core::Core<T,S>::poll
Mar 25 14:56:41 <hostname> sh[942819]:   19: tokio::runtime::task::harness::Harness<T,S>::poll
Mar 25 14:56:41 <hostname> sh[942819]:   20: tokio::runtime::scheduler::multi_thread::worker::Context::run_task
Mar 25 14:56:41 <hostname> sh[942819]:   21: tokio::runtime::scheduler::multi_thread::worker::run
Mar 25 14:56:41 <hostname> sh[942819]:   22: tokio::loom::std::unsafe_cell::UnsafeCell<T>::with_mut
Mar 25 14:56:41 <hostname> sh[942819]:   23: tokio::runtime::task::core::Core<T,S>::poll
Mar 25 14:56:41 <hostname> sh[942819]:   24: tokio::runtime::task::harness::Harness<T,S>::poll
Mar 25 14:56:41 <hostname> sh[942819]:   25: tokio::runtime::blocking::pool::Inner::run
Mar 25 14:56:41 <hostname> sh[942819]: note: Some details are omitted, run with `RUST_BACKTRACE=full` for a verbose backtrace.

Full backtrace

Mar 25 15:02:03 <hostname> sh[957703]: thread 'tokio-runtime-worker' panicked at /home/username/.cargo/registry/src/index.crates.io-6f17d22bba15001f/arrow-data-49.0.0/src/transform/utils.rs:42:56:
Mar 25 15:02:03 <hostname> sh[957703]: offset overflow
Mar 25 15:02:03 <hostname> sh[957703]: stack backtrace:
Mar 25 15:02:03 <hostname> sh[957703]: thread 'tokio-runtime-worker' panicked at /home/username/.cargo/registry/src/index.crates.io-6f17d22bba15001f/arrow-data-49.0.0/src/transform/utils.rs:42:56:
Mar 25 15:02:03 <hostname> sh[957703]: offset overflow
Mar 25 15:02:03 <hostname> sh[957703]:    0:     0x560a5f972a96 - std::backtrace_rs::backtrace::libunwind::trace::hbee8a7973eeb6c93
Mar 25 15:02:03 <hostname> sh[957703]:                                at /rustc/07dca489ac2d933c78d3c5158e3f43beefeb02ce/library/std/src/../../backtrace/src/backtrace/libunwind.rs:104:5
Mar 25 15:02:03 <hostname> sh[957703]:    1:     0x560a5f972a96 - std::backtrace_rs::backtrace::trace_unsynchronized::hc8ac75eea3aa6899
Mar 25 15:02:03 <hostname> sh[957703]:                                at /rustc/07dca489ac2d933c78d3c5158e3f43beefeb02ce/library/std/src/../../backtrace/src/backtrace/mod.rs:66:5
Mar 25 15:02:03 <hostname> sh[957703]:    2:     0x560a5f972a96 - std::sys_common::backtrace::_print_fmt::hc7f3e3b5298b1083
Mar 25 15:02:03 <hostname> sh[957703]:                                at /rustc/07dca489ac2d933c78d3c5158e3f43beefeb02ce/library/std/src/sys_common/backtrace.rs:68:5
Mar 25 15:02:03 <hostname> sh[957703]:    3:     0x560a5f972a96 - <std::sys_common::backtrace::_print::DisplayBacktrace as core::fmt::Display>::fmt::hbb235daedd7c6190
Mar 25 15:02:03 <hostname> sh[957703]:                                at /rustc/07dca489ac2d933c78d3c5158e3f43beefeb02ce/library/std/src/sys_common/backtrace.rs:44:22
Mar 25 15:02:03 <hostname> sh[957703]:    4:     0x560a5f99fe00 - core::fmt::rt::Argument::fmt::h76c38a80d925a410
Mar 25 15:02:03 <hostname> sh[957703]:                                at /rustc/07dca489ac2d933c78d3c5158e3f43beefeb02ce/library/core/src/fmt/rt.rs:142:9
Mar 25 15:02:03 <hostname> sh[957703]:    5:     0x560a5f99fe00 - core::fmt::write::h3ed6aeaa977c8e45
Mar 25 15:02:03 <hostname> sh[957703]:                                at /rustc/07dca489ac2d933c78d3c5158e3f43beefeb02ce/library/core/src/fmt/mod.rs:1120:17
Mar 25 15:02:03 <hostname> sh[957703]:    6:     0x560a5f96f9ff - std::io::Write::write_fmt::h78b18af5775fedb5
Mar 25 15:02:03 <hostname> sh[957703]:                                at /rustc/07dca489ac2d933c78d3c5158e3f43beefeb02ce/library/std/src/io/mod.rs:1810:15
Mar 25 15:02:03 <hostname> sh[957703]:    7:     0x560a5f972874 - std::sys_common::backtrace::_print::h5d645a07e0fcfdbb
Mar 25 15:02:03 <hostname> sh[957703]:                                at /rustc/07dca489ac2d933c78d3c5158e3f43beefeb02ce/library/std/src/sys_common/backtrace.rs:47:5
Mar 25 15:02:03 <hostname> sh[957703]:    8:     0x560a5f972874 - std::sys_common::backtrace::print::h85035a511aafe7a8
Mar 25 15:02:03 <hostname> sh[957703]:                                at /rustc/07dca489ac2d933c78d3c5158e3f43beefeb02ce/library/std/src/sys_common/backtrace.rs:34:9
Mar 25 15:02:03 <hostname> sh[957703]:    9:     0x560a5f9740f7 - std::panicking::default_hook::{{closure}}::hcce8cea212785a25
Mar 25 15:02:03 <hostname> sh[957703]:   10:     0x560a5f973e59 - std::panicking::default_hook::hf5fcb0f213fe709a
Mar 25 15:02:03 <hostname> sh[957703]:                                at /rustc/07dca489ac2d933c78d3c5158e3f43beefeb02ce/library/std/src/panicking.rs:292:9
Mar 25 15:02:03 <hostname> sh[957703]:   11:     0x560a5f974588 - std::panicking::rust_panic_with_hook::h095fccf1dc9379ee
Mar 25 15:02:03 <hostname> sh[957703]:                                at /rustc/07dca489ac2d933c78d3c5158e3f43beefeb02ce/library/std/src/panicking.rs:779:13
Mar 25 15:02:03 <hostname> sh[957703]:   12:     0x560a5f974462 - std::panicking::begin_panic_handler::{{closure}}::h032ba12139b353db
Mar 25 15:02:03 <hostname> sh[957703]:                                at /rustc/07dca489ac2d933c78d3c5158e3f43beefeb02ce/library/std/src/panicking.rs:657:13
Mar 25 15:02:03 <hostname> sh[957703]:   13:     0x560a5f972f96 - std::sys_common::backtrace::__rust_end_short_backtrace::h9259bc2ff8fd0f76
Mar 25 15:02:03 <hostname> sh[957703]:                                at /rustc/07dca489ac2d933c78d3c5158e3f43beefeb02ce/library/std/src/sys_common/backtrace.rs:171:18
Mar 25 15:02:03 <hostname> sh[957703]:   14:     0x560a5f9741c0 - rust_begin_unwind
Mar 25 15:02:03 <hostname> sh[957703]:                                at /rustc/07dca489ac2d933c78d3c5158e3f43beefeb02ce/library/std/src/panicking.rs:645:5
Mar 25 15:02:03 <hostname> sh[957703]:   15:     0x560a5d65e4e5 - core::panicking::panic_fmt::h784f20a50eaab275
Mar 25 15:02:03 <hostname> sh[957703]:                                at /rustc/07dca489ac2d933c78d3c5158e3f43beefeb02ce/library/core/src/panicking.rs:72:14
Mar 25 15:02:03 <hostname> sh[957703]:   16:     0x560a5d65e4a3 - core::panicking::panic_display::h251010ce5e0560d5
Mar 25 15:02:03 <hostname> sh[957703]:                                at /rustc/07dca489ac2d933c78d3c5158e3f43beefeb02ce/library/core/src/panicking.rs:196:5
Mar 25 15:02:03 <hostname> sh[957703]:   17:     0x560a5d65e4a3 - core::panicking::panic_str::h6f89534c81f0edc4
Mar 25 15:02:03 <hostname> sh[957703]:                                at /rustc/07dca489ac2d933c78d3c5158e3f43beefeb02ce/library/core/src/panicking.rs:171:5
Mar 25 15:02:03 <hostname> sh[957703]:   18:     0x560a5d65e4a3 - core::option::expect_failed::hc85eb6037a3050f7
Mar 25 15:02:03 <hostname> sh[957703]:                                at /rustc/07dca489ac2d933c78d3c5158e3f43beefeb02ce/library/core/src/option.rs:1980:5
Mar 25 15:02:03 <hostname> sh[957703]:   19:     0x560a5f924a82 - arrow_data::transform::utils::extend_offsets::hdf2f06f3dfb426cd
Mar 25 15:02:03 <hostname> sh[957703]:   20:     0x560a5f9279c6 - arrow_data::transform::variable_size::build_extend::{{closure}}::h31904a480ae969de
Mar 25 15:02:03 <hostname> sh[957703]:   21:     0x560a5f9183b5 - arrow_data::transform::MutableArrayData::extend::hde96a1dcf5382372
Mar 25 15:02:03 <hostname> sh[957703]:   22:     0x560a5f89081f - arrow_select::concat::concat_fallback::h58848d6223cb12c9
Mar 25 15:02:03 <hostname> sh[957703]:   23:     0x560a5f89049e - arrow_select::concat::concat::h2c60d37437746b3e
Mar 25 15:02:03 <hostname> sh[957703]:   24:     0x560a5e101518 - arrow_select::concat::concat_batches::hbc4b4e1ab862a29d
Mar 25 15:02:03 <hostname> sh[957703]:   25:     0x560a5e178464 - <futures_util::future::future::map::Map<Fut,F> as core::future::future::Future>::poll::h3e1df0aac2ab9d77
Mar 25 15:02:03 <hostname> sh[957703]:   26:     0x560a5e02f8f2 - <futures_util::future::future::shared::Shared<Fut> as core::future::future::Future>::poll::h7c9852d5bb7e9534
Mar 25 15:02:03 <hostname> sh[957703]:   27:     0x560a5e16e2bf - datafusion_physical_plan::joins::utils::OnceFut<T>::get::hd3b2e25c093ada6d
Mar 25 15:02:03 <hostname> sh[957703]:   28:     0x560a5e195dd9 - <datafusion_physical_plan::joins::hash_join::HashJoinStream as futures_core::stream::Stream>::poll_next::h9c658418141705f6
Mar 25 15:02:03 <hostname> sh[957703]:   29:     0x560a5e18a696 - <datafusion_physical_plan::coalesce_batches::CoalesceBatchesStream as futures_core::stream::Stream>::poll_next::h219998518ccf857d
Mar 25 15:02:03 <hostname> sh[957703]:   30:     0x560a5e00bada - datafusion_physical_plan::repartition::RepartitionExec::pull_from_input::{{closure}}::h1ed64d4cd49cce5e
Mar 25 15:02:03 <hostname> sh[957703]:   31:     0x560a5e000993 - tokio::loom::std::unsafe_cell::UnsafeCell<T>::with_mut::h38ddf14813d65bab
Mar 25 15:02:03 <hostname> sh[957703]:   32:     0x560a5e03d6ee - tokio::runtime::task::core::Core<T,S>::poll::h3df6db59b96cdc95
Mar 25 15:02:03 <hostname> sh[957703]:   33:     0x560a5e040779 - tokio::runtime::task::harness::Harness<T,S>::poll::h56978e4fa718a3b4
Mar 25 15:02:03 <hostname> sh[957703]:   34:     0x560a5f0e3ad8 - tokio::runtime::scheduler::multi_thread::worker::Context::run_task::hb25f4c4bb31b0772
Mar 25 15:02:03 <hostname> sh[957703]:   35:     0x560a5f0e35b5 - tokio::runtime::scheduler::multi_thread::worker::run::h9904ff19a8f264dd
Mar 25 15:02:03 <hostname> sh[957703]:   36:     0x560a5f0dfcd2 - tokio::loom::std::unsafe_cell::UnsafeCell<T>::with_mut::h76fc62ff172daba6
Mar 25 15:02:03 <hostname> sh[957703]:   37:     0x560a5f0f2af9 - tokio::runtime::task::core::Core<T,S>::poll::h256060c96f596608
Mar 25 15:02:03 <hostname> sh[957703]:   38:     0x560a5f0ded9a - tokio::runtime::task::harness::Harness<T,S>::poll::h0b57f37a1c3a1ec2
Mar 25 15:02:03 <hostname> sh[957703]:   39:     0x560a5f0ee780 - tokio::runtime::blocking::pool::Inner::run::ha371e4f6189503f2
Mar 25 15:02:03 <hostname> sh[957703]:   40:     0x560a5f0e84fc - std::sys_common::backtrace::__rust_begin_short_backtrace::h25663fb0bf4a4475
Mar 25 15:02:03 <hostname> sh[957703]:   41:     0x560a5f0f0a79 - core::ops::function::FnOnce::call_once{{vtable.shim}}::ha0885a2c26e1b6a6
Mar 25 15:02:03 <hostname> sh[957703]:   42:     0x560a5f9783d5 - <alloc::boxed::Box<F,A> as core::ops::function::FnOnce<Args>>::call_once::h12de4fc57affb195
Mar 25 15:02:03 <hostname> sh[957703]:                                at /rustc/07dca489ac2d933c78d3c5158e3f43beefeb02ce/library/alloc/src/boxed.rs:2015:9
Mar 25 15:02:03 <hostname> sh[957703]:   43:     0x560a5f9783d5 - <alloc::boxed::Box<F,A> as core::ops::function::FnOnce<Args>>::call_once::h3c619f45059d5cf1
Mar 25 15:02:03 <hostname> sh[957703]:                                at /rustc/07dca489ac2d933c78d3c5158e3f43beefeb02ce/library/alloc/src/boxed.rs:2015:9
Mar 25 15:02:03 <hostname> sh[957703]:   44:     0x560a5f9783d5 - std::sys::unix::thread::Thread::new::thread_start::hbac657605e4b7389
Mar 25 15:02:03 <hostname> sh[957703]:                                at /rustc/07dca489ac2d933c78d3c5158e3f43beefeb02ce/library/std/src/sys/unix/thread.rs:108:17
Mar 25 15:02:03 <hostname> sh[957703]:   45:     0x7f094d03d1ca - start_thread
Mar 25 15:02:03 <hostname> sh[957703]:   46:     0x7f094c723e73 - __clone
Mar 25 15:02:03 <hostname> sh[957703]:   47:                0x0 - <unknown>

Metadata

Metadata

Assignees

No one assigned

    Labels

    bugSomething isn't working

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions