Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Long-running Spark rewrite Files Action may lead to OutOfMemoryError #11277

Open
Zhanxiao-Ma opened this issue Oct 8, 2024 · 1 comment
Open

Comments

@Zhanxiao-Ma
Copy link

In my production environment, I have observed that long-running Spark rewrite Files Action can lead to OutOfMemoryError. Analyze the Java dump, I noticed a large number of ChildAllocator objects that are only referenced by the RootAllocator. Upon reviewing the code, I discovered that the ChildAllocator allocated at this point is indeed not being released. Is this correct?

@Zhanxiao-Ma
Copy link
Author

I think the close method of VectorizedArrowReader should include the logic to release rootAlloc.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant