Does ActivityPub send those to other instances, or does ActivityPub only send the original post and the rest (upvotes, downvotes, replies) are stored only on the original server where the post was made?

  • DaveA
    link
    fedilink
    arrow-up
    12
    ·
    edit-2
    9 months ago

    My instance has 800 users, is 4 months old, and the database only is over 30GB. It is an insane amount of data.

      • DaveA
        link
        fedilink
        arrow-up
        2
        ·
        edit-2
        9 months ago

        I’m a bad example. I haven’t properly tuned the settings, currently RAM will grow to whatever is available.

        I’m very lucky, the instance is running in a proxmox container alongside some other fediverse servers (run by others), on dedicated hardware in a datacentre. The sysadmin has basically thrown me plenty of spare resources since the other containers aren’t using them and RAM not used is wasted, so I’ve got 32GB allocated currently. I still need to restart once a week or that RAM gets used up and the database container crashes.

        It’s been on my list of things to do for a while, try some different postgres configs, but I just haven’t got around to it.

        I know a couple of months back lemmy.world were restarting every 30 mins so they didn’t use up all the RAM and crash. I presume some time and some lemmy updates later that’s no longer the case.

        I know some smaller servers get away with 2gb of RAM, and we should be able to use a lot less than 32GB if I actually try to tune the postgres config.

    • nutomic@lemmy.ml
      link
      fedilink
      arrow-up
      2
      ·
      9 months ago

      There is a postgres command to show the size of each table. Most likely it is from activity tables which can be cleared out to save space.

      • DaveA
        link
        fedilink
        arrow-up
        1
        ·
        9 months ago

        After the second-to-last update the database shrunk and I was under the impression there was some automatic removal happening. Was this not the case?

        It’s helpful info for others but personally I’m not that worried about the database size. The size of the pictrs cache is much more of a concern, and as I understand it there isn’t an easy way to identify and remove cache images without accidentally taking out user image uploads.

        • nutomic@lemmy.ml
          link
          fedilink
          arrow-up
          2
          arrow-down
          1
          ·
          9 months ago

          Yes there is automatic removal so if you have enough disk space, no need to worry about it.

          The pictrs storage only consists of uploads from local users, and thumbnails for both local and remote posts. Thumbnails for remote posts could theoretically be wiped and loaded from the other instance, but they shouldnt take much space anyway.

          • DaveA
            link
            fedilink
            arrow-up
            1
            ·
            9 months ago

            Yes there is automatic removal so if you have enough disk space, no need to worry about it.

            What triggers this? My DB was about 30GB, then the update shrunk it down to 5GB, then it grew back to 30GB.

            The pictrs storage only consists of uploads from local users, and thumbnails for both local and remote posts. Thumbnails for remote posts could theoretically be wiped and loaded from the other instance, but they shouldnt take much space anyway.

            I’d be pretty confident that the 140GB of pictrs cache I have is mostly cache. There are occasionaly users uploading images, but we don’t have that many active users, I’d be surprised if there was more than a few GB of image uploads in total out of that 140GB. We just aren’t that big of a server.

            The pictrs volume also grows consistently at a little under 1GB per day. I just went and had a look, in the files directory there are 6 directories from today (the day only has a couple of hours left), and these sum to almost 700MB of images and almost 6000 files, or a little over 100KB each.

            The instance has had just 27 active users today (though of course users not posting will still generate thumbnails).

            While the cached images may be small, it adds up really quick.

            As far as I can tell there is no cache pruning, as the cache goes up pretty consistently each day.

            • nutomic@lemmy.ml
              link
              fedilink
              arrow-up
              2
              arrow-down
              1
              ·
              9 months ago

              The activities table is cleared out automatically every week, items older than 3 months are deleted. During the update only a smaller number of rows was migrated so the db temporarily was slower. You can manually clear older items in sent_activity and received_activity to free more space.

              Actually Im wrong about images, turns out that all remote images are mirrored locally in order to generate thumbnails. 0.19 will have an option to disable that. This could use more improvements, the whole image handling is rather confusing now.

              • DaveA
                link
                fedilink
                arrow-up
                1
                ·
                9 months ago

                Thanks for the info! Ior performance reasons it would be nice to have a way to configure how long the cache is kept rather than disable it completely, but I understand you probably have other priorities.

                Would disabling the cache remove images cached up to that point?

      • DaveA
        link
        fedilink
        arrow-up
        1
        ·
        9 months ago

        Lemmy already has serious scaling issues. It’s priority one for the devs at the moment. The next release has major backend changes.

        Lemmy is still version 0, it’s basically not released yet, so we have to give them some slack. They weren’t exactly expecting to go from less than 1000 monthly users to tens of thousands almost overnight, on a platform where development was still early days.