-
-
Notifications
You must be signed in to change notification settings - Fork 2k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Remove Text, SQLite and HDF5 backends #4231
Conversation
Codecov Report
@@ Coverage Diff @@
## master #4231 +/- ##
==========================================
- Coverage 88.93% 88.18% -0.75%
==========================================
Files 92 89 -3
Lines 14788 14358 -430
==========================================
- Hits 13152 12662 -490
- Misses 1636 1696 +60
|
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
So much deleted code! 🤩
In principle I approve, but before someone accidentally hits "merge": Do we want to deprecate the backends already for 3.10
?
I'd say yes, but we have a major bump coming up and could do it there too.
As the warnings have been here for a while now, and as ArviZ netcdf backend is already available, I think we can go ahead and depreacte this in 3.10. |
I'm pretty confused. Why were these backends removed? I'm running out of memory trying to sample a model, thought that writing to disk was pretty cool. Unsure what to do. |
@DanPen they were removed because because couldn't think of examples that would not work with the standard backend, and we saw no activity in Github/Discourse that indicated that they were being used at all. Now to your problem: Can you give some details about the size of the model and sampling strategy? Running out of memory sounds like you either have a memory leak, or your model is astronomically gigantic, or you're including things in the trace that are deterministic and take a lot of memory. |
@michaelosthege Thanks for the quick response! I'm following the Dependent density regression tutorial with my own dataset. As I increase my number of observations to ~500, my 32GB of memory gets used up. I'm sampling with 4 metropolis chains, 20,000 samples, and 10,000 burn-in. Memory seems to increase linearly as sampling continues. Should I open a new issue? |
@DanPen that's 266 kB per iteration, roughly equivalent to 8000 float32 values. I doubt that your model has that many parameters (if it does, you should give up on You can In a Jupyter notebook:
|
@michaelosthege Cool. Not sure I understand why |
@DanPen Then it's indeed the 2D In such cases it's better to not save them at all. You can do this by not wrapping them in |
Since the warning has been present for over half a year, this PR removes the storage backends other than NDArray.
This might be a good place to remove save/load for NDArray as well, or this can be done in a separate PR.