1329
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
this post was submitted on 30 Jul 2023
1329 points (98.7% liked)
Lemmy.World Announcements
29157 readers
151 users here now
This Community is intended for posts about the Lemmy.world server by the admins.
Follow us for server news ๐
Outages ๐ฅ
https://status.lemmy.world
For support with issues at Lemmy.world, go to the Lemmy.world Support community.
Support e-mail
Any support requests are best sent to info@lemmy.world e-mail.
Report contact
- DM https://lemmy.world/u/lwreport
- Email report@lemmy.world (PGP Supported)
Donations ๐
If you would like to make a donation to support the cost of running this platform, please do so at the following donation URLs.
If you can, please use / switch to Ko-Fi, it has the lowest fees for us
Join the team
founded 2 years ago
MODERATORS
Is it me or is the 80% figure just insane? Are there any benchmarks to see how fast this has become versus say Lemmy 0.18.2 on a very large instance?
Not really, you'd be surprised how often systems are bloated all because of a single option, character, etc. Most developers don't start optimizing until much later in the software's lifecycle. Often enough, it is easily overlooked. That's why code reviews are needed often with fresh pair of eyes.
Just to set the expectations, reducing database size or CPU usage does not necessarily mean it is faster but it does mean there's more free capacity on the servers to handle more users at the same performance.
More importantly; they may help reduce costs on the smaller indie instances that doesn't need to buy larger server instances.
Hopefully, we'll continue to see more of these optimizations.
I believe if the backend doesn't have to write as much data then you'll have less I/O operations so it should IMO have an impact on the overall speed of Lemmy (unless all of those operations are done asynchronously). Same for the reduced CPU usage, it could allow for more stuff in parallel.
Speed/pref and capacity are two separate things. I/O has nothing to do with the size of the database. You can write 100TB per second into the database and choose to only store 1TB of content. That does not mean the app is writing 1TB per second, it is still writing 100TBps.
They said they changed how the activities data is stored, which reduced the size by 80%; here's one of the changes they made: https://github.com/LemmyNet/lemmy/issues/3103
As you can see, the issue here is that they were storing a lot of data in the activities table that is not needed, it was only meant for debug purposes. So, they split up the data into two and not store the other data as it isn't needed; they're still writing these data the same as before. One part is used to ensure they don't re-process the same data but this is the same thing they were doing before this change.
In addition, they've limited how long the data is retained for 3 months, which is a separate job they run to remove data.
All of this has zero impact on the users using the app right now. The main benefit is for instance admins with limited storage. One might say the system slows down if there's not enough space but that is still the same case here with this MR, it will still slow down.
Wait does that mean posts older than 3 months gets automatically deleted? Isn't that kinda bad? Being able to find years old posts is an important part of Reddit and pretty much all social media.
Nobody said anything about posts. Activities could mean extra metadata like "someone joined", "someone changed", posted, edited, etc.
Most of the time, these isn't required after a short while.
Funnily enough, this is the feature that can speed up the performance by doing less calls: