Hey folks

This is a heads up that I will be performing some maintenance and hardware upgrades on our database this Saturday.

We are currently experiencing several spikes throughout the day which cause our database to become overloaded - this results in degraded performance for many users. The spikes are happening due to a combination of continued growth of the database, some expensive periodic scheduled tasks which Lemmy runs, and fluctuating traffic patterns. Some of this can be optimized on the code level in the future, but it seems that the best way to deal with it right now is to add some additional resources to our database server.

I am intending to switch to slightly different hardware in this upgrade, and will be unable to make this switch without downtime, so unfortunately lemm.ee will be unavailable for the duration.

As our database has grown quite a bit, cloning it will most likely take a few hours, so I expect the downtime to last 2-3 hours. Sorry for the inconvenience, I am hopeful that it will be worth it and that this upgrade will significantly reduce some of our recent long page load times!


Edit: upgrade complete!

I have now migrated the lemm.ee database from the original DigitalOcean managed database service to a dedicated server on Hetzner.

As part of this migration, I have also moved all of our Lemmy servers from the DigitalOcean cloud to Hetzner’s Cloud. I always want the servers to be as close as possible to the database, in order to keep latencies low. At the same time, I am very interested in having the ability to dynamically spin up and down servers as needed, so a cloud-type solution is really ideal for that. Fortunately, Hetzner allows connecting cloud servers to their dedicated servers through a private network, so we are able to take advantage of a powerful dedicated server for the database, while retaining the flexibility of the cloud approach for the rest of our servers. I’m really happy with the solution now.

In terms of results, I am already seeing far better page load times and far less resource use on the new hardware, so I think the migration has been a success. I will keep monitoring things and tuning as necessary.

  • @Nelots
    link
    English
    494 months ago

    So I wasn’t going crazy, the long loads times were real. Glad to see upgrades coming soon!

    • @JimmyBigSausage
      link
      204 months ago

      Although the long page load times have been real, you still might be going crazy. Definitely a possibility on Lemmy (づ ̄ ³ ̄)づ

    • kratoz29
      link
      English
      64 months ago

      My experience has been downgraded since the latest Lemmy update, all these caveats/workarounds are pretty much welcome in order to get this in a stable state again!

    • shootwhatsmyname
      link
      English
      34 months ago

      That’s awesome! Are you monitoring personally or is this from somewhere public?

      • NegativeNull@lemm.ee
        link
        English
        64 months ago

        I run a bunch of small websites/tools. I setup UptimeKuma a while back to monitor my own apps. Since it’s just a simple web monitor, I can monitor any http/s URLs. I setup a lemm.ee monitor a few days ago when things were starting to get unstable.

        • shootwhatsmyname
          link
          English
          14 months ago

          Oh nice, thanks! I think I’m gonna go monitor everything now

  • @GrayBackgroundMusic
    link
    English
    174 months ago

    No worries about the downtime. If that’s what’s needed to get back to 0.18 levels of performance, go for it! Thanks for all the work y’all do.

  • @argo_yamato
    link
    144 months ago

    Thank you for your work on lemm.ee! I was actually looking to post somewhere to see if there was some slowness or if it was just me, looks like you answered my question.

  • @Lemonparty
    link
    144 months ago

    Np, thanks so much for all you do!

  • Spunky Monkey
    link
    134 months ago

    It’s flying now! Thanks for the upgrade @sunaurus@lemm.ee.

  • Neuromancer
    link
    124 months ago

    What are the current specs and what are you going to?

    • @sunaurusOPMA
      link
      10
      edit-2
      4 months ago

      Currently the database is a managed DigitalOcean Postgres instance, but I am going to migrate it to a 32 thread 128 gb RAM dedicated server.

      It’s something that I’ve been hoping we won’t need, as the managed database service has allowed me to not worry about patches, backups, etc (they took care of all of that automatically). Unfortunately it is clear now that further upgrades are just too costly on that service & actually the amount of configuration and tuning I would like to do is simply not possible there. So seems like right now, moving to the dedicated server for the database is the only option.

      • Neuromancer
        link
        24 months ago

        Makes sense. I appreciate the reply back. I use to work on highly available fault tolerant systems. So I’m always interested how much hardware/etc to run something.

        Sounds like some serious optimization needs to be done or there are a lot more transaction than I was expecting.

    • @vsh
      link
      English
      -17
      edit-2
      4 months ago

      Removed by mod

      • Neuromancer
        link
        34 months ago

        Ah thought he had the old specs posted

        • 1024_Kibibytes
          link
          24 months ago

          I’m pretty sure I remember seeing the specs when he moved the server behind Cloudflare. It was put behind their CDN because we kept getting DDoS attacks. It’s fairly nice if I recall correctly.

          • Neuromancer
            link
            -14 months ago

            I thought so as well. I thought he had an architectural diagram but I can’t find them. So I must be wrong

            • 1024_Kibibytes
              link
              24 months ago

              You’re not necessarily wrong. You just can’t find it at the moment. I have trouble finding information all the time and that’s a big chunk of my job.

  • Xepher
    link
    English
    114 months ago

    Really appreciate the advanced heads up. Thanks for being such a awesome instance admin!

    • ditty
      link
      154 months ago

      1 hr old; hang in there champ! 🫡💩

      • @CTDummy
        link
        154 months ago

        Nothing after 5h. Died doing what he loved RIP Cluck

    • @0ops
      link
      44 months ago

      Why the downvotes? I’m laughing like an idiot over here

  • Aniki 🌱🌿
    link
    English
    84 months ago

    I just wanted to tell you both, “Good luck,” We’re all counting on you.

    -Frank Drebbin, Police Squad

  • @SurvivalMariner
    link
    84 months ago

    Thanks for the update and information. You really set an example for communication and level headedness in decision making. Thanks for providing this space for us.

  • @MoreCoffee
    link
    English
    74 months ago

    Performance is night and day now, thanks for the hard work