-
Notifications
You must be signed in to change notification settings - Fork 6
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Pandas Ran out of memory again! #69
Comments
That's strange. There is a flag I'm not sure if
I'm not sure if we've got many more markers you can play with, but if you uninstall optional libraries you'll be running less tests. You can set up an environment with just |
Interesting, I will give that a go. Just those three? Is that documented anywhere or used anywhere, or just something you try locally when you are running less tests? |
Those three are documented to be the minimal required dependencies (eg https://pandas.pydata.org/docs/dev/getting_started/install.html#dependencies), and since we automatically skip tests for optional dependencies (not something that is explicitly documented I think), having only those installed is the way to run only tests that don't rely on any optional dependency. |
Still happening :( It is now exceeding 14 gb... #94 (comment) |
So the pandas test suite ran out of memory again in Kubernetes. It used up ~13Gb and then was killed, because the pods only have that much available.
I am a bit hesitant to just raise the pod memory limit again... If anyone knows if this is a reasonable amount of memory for Pandas to use when testing (cc @datapythonista), that would be helpful! It's also possible that the tracing has some sort of memory leak which is blowing things up for pandas, although all the other test suites don't seem to have the same problem.
Maybe I can run Pandas test suites with some flags to ignore some high memory tests? These are my current ones:
python-record-api/k8/images/pandas/Dockerfile
Line 40 in 94902a2
I copied it from the
test-fast
script, or whatever that is, in the Pandas repo.The text was updated successfully, but these errors were encountered: