Skip to content

[Python] pyarrow.csv.read_csv hangs + eats all RAM #22212

@asfimport

Description

@asfimport

I have quite a sparse dataset in CSV format. A wide table that has several rows but many (32k) columns. Total size ~540K.

When I read the dataset using pyarrow.csv.read_csv it hangs, gradually eats all memory and gets killed.

More details on the conditions further. Script to run and all mentioned files are under attachments.

  1. sample_32769_cols.csv is the dataset that suffers the problem.

  2. sample_32768_cols.csv is the dataset that DOES NOT suffer and is read in under 400ms on my machine. It's the same dataset without ONE last column. That last column is no different than others and has empty values.

The reason of why exactly this column makes difference between proper execution and hanging failure which looks like some memory leak - no idea.

I have created flame graph for the case (1) to support this issue resolution (graph.svg).

 

Environment: Ubuntu Xenial, python 2.7
Reporter: Bogdan Klichuk
Assignee: Micah Kornfield / @emkornfield

Original Issue Attachments:

PRs and other links:

Note: This issue was originally created as ARROW-5791. Please see the migration documentation for further details.

Metadata

Metadata

Assignees

Type

No type

Projects

No projects

Milestone

Relationships

None yet

Development

No branches or pull requests

Issue actions