I’m using Sentry on-premises and it’s working well.
But Sentry is constantly filling up my disk. I checked the location that is taking all the space and it’s /var/lib/docker/volumes/sentry-postgres/_data/base/*
It’s already over 50G and it’s growing all the time. The files in there are created by user “systemd-coredump”, but I guess these aren’t just dumps but just the normal entries generated by Sentry?
Is there any way to clean up the space or setup Sentry not to generate that much data in the first place? Or is it even normal that Sentry will take up so much space and I just have to buy more and more disk space?
That will constantly reset most state within Sentry, not sure how useful the product becomes if e.g. all your issue triaging progress resets. I would look into which files within that volume actually take up space, or which tables.
Yes, fully agreed that this would limit the usefulness… but for our use case it wouldn’t be that much of an issue. We normally address issues with hours or max a few days of popping up on Sentry. So, we don’t need the long issue history.
Agreed that emptying a table instead of completely resetting/wiping the whole database would definitely be preferable. But would need guidance on that from the Sentry team…
For us to help you we need to know which postgres table is primarily responsible for the excessive disk usage. I can’t give definitive advice on how to figure that out, but a quick google search gives me e.g. this: https://wiki.postgresql.org/wiki/Disk_Usage
I have the same issue… I’m at 50 GB a day now.
The only suggestion I see in those threads above
is to turn off organizations:performance-view
Also maybe switch our backends. Is there an option to switch our backends to something more scalable? Or at least move over bulk of data off postgres.
Context: Im using the helm chart here
I’ve also changed my retention to 30 Days. and job is now running every day.
Tried Vaccuum FULL to try to reclaim space. All of which didnt seem to have helped the issue.
If there are any suggestions here, would be appreciated.
Since Sentry was not working correctly I had to delete the installation an reinstall it on a clean machine. This time I disabled the transactions (performance) by setting the traces-samplerate to 0 and it seems that this was the case for large amount of data that was stored.
But I still don’t know why the files are created by systemd-coredump, because my DB wasn’t crashing.
I’m still a little bit suprised about the amount of stored data by transactions/performance. I can remember that Sentry saved about 150k transactions which took about 60GB of storage. Is that a normal behaviour?
@chrisvdb I’m not aware of any Sentry option that would allow you to tweak event retention to 1 week. Yes you can delete any items from there, it will cause a little bit of instability but not a lot.
yes autoresolve can probably be set to 1 week, but we still allow you to view 90 days worth of event data and also un-resolve if the issue starts occurring again
I’m currently still facing same issues…
Disk usage is growing at > ~50 GB a day.
So far, these are the suggestions I tried out.
Turning off organizations:performance-view (didn’t see much decrease or any at all)
Changing Retention days to 30. Job ran every day. (also didnt see much decrease)
DELETE FROM public.nodestore_node WHERE “timestamp” < NOW() - INTERVAL ‘30 day’;
VACUUM FULL public.nodestore_node; (This did reclaim space but disk space usage still insanely high so it only bought me time)
Switch our backends as per the other suggested thread. Is there an option to switch our backends to something more scalable? Or at least move over bulk of data off postgres.
Context: Im using the helm chart here (Havent tried this but doesnt seem like its in documentation or supported by the helm chart)
Is there any way to decrease the amount of data? or store it somewhere else?