-
Notifications
You must be signed in to change notification settings - Fork 19
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Data download consumes too much memory #140
Comments
We cannot go from max DataFrame size yet, but possibly its memory consumption is much lower than when it's being initialized. |
The raw data is stored in a numpy array and has certain size. When downloading, there's no memory overhead (the memory consumption is smaller then the result array). Pandas looks more memory efficient. Code |
@andersonberg @felipefin @alexandr1988 Have you hit the wall with memory recently, in particular in Kosmonaut high memory profile? |
I remember having issues downloading more than 500k items, however I was not using Kosmonaut. |
I think I didn't have a task like this. Will let you know if I have
…On Mon, Oct 14, 2019 at 5:10 PM Felipe Lunkes Fin ***@***.***> wrote:
I remember having issues downloading more than 500k items, however I was
not using Kosmonaut.
I'll keep that in mind next time I face this kind of situations.
—
You are receiving this because you were mentioned.
Reply to this email directly, view it on GitHub
<#140?email_source=notifications&email_token=ABWRSRYPOF6CNRGLAVQKE5TQOR4WNA5CNFSM4IBOEUN2YY3PNVWWK3TUL52HS4DFVREXG43VMVBW63LNMVXHJKTDN5WW2ZLOORPWSZGOEBE3LBY#issuecomment-541701511>,
or unsubscribe
<https://github.com/notifications/unsubscribe-auth/ABWRSR2NWFIEIUZCSEYQPPTQOR4WNANCNFSM4IBOEUNQ>
.
|
No, I didn't use Kosmonaut with high memory yet. Adding to that, I didn't face any memory problems in the Arche profile |
Closing because the most straightforward way is to up the server memory. |
Downloading 3m of items consumes about 8gb memory. Not much, but could it be lower?
Possible solution with chunks - scrapinghub/python-scrapinghub#121
The text was updated successfully, but these errors were encountered: