Edited: this post to be the Lemmy.World federation issue post.

We are now ready to upgrade to postgres 16. When this post is 30 mins old the maintenace will start


Updated & Pinned a Comment in this thread explaining my complete investigation and ideas on how the Lemmy app will & could move forward.

    • TiffOPMA
      link
      fedilink
      English
      arrow-up
      4
      ·
      4 months ago

      Yes… it is VERY annoying. We have so much resources available and lemmy/postgres will not use them

        • TiffOPMA
          link
          fedilink
          English
          arrow-up
          5
          ·
          4 months ago

          Yes. Unfortunately the information gleamed boils down to two reasons:

          • their db was slow to respond
          • their db server ended up being 25ms away from their backend servers which caused the slowness.

          Our db server is occasionally slow to respond, but most requests from LW complete in less than 0.1 second. Unfortunately there are times when they take longer. These longer ones are going to be the problem (I believe). As all activities are sequential servers can only catch up as fast as they can process them.

          What i’ve found in the past 30 seconds is that it is not necessarily out database that is the problem but possibly the way lemmy handles the federation. I’m chatting with some of the admins on reddthat and making pretty graphs while looking at walls of logs.