Dust is a rewrite of du (in rust obviously) that visualizes your directory tree and what percentage each file takes up. But it only prints as many files fit in your terminal height, so you see only the largest files. It’s been a better experience that du, which isn’t always easy to navigate to find big files (or atleast I’m not good at it.)

Anyway, found a log file at .local/state/nvim/log that was 70gb. I deleted it. Hope it doesn’t bite me. Been pushing around 95% of disk space for a while so this was a huge win 👍

    • Nik282000@lemmy.ca
      link
      fedilink
      arrow-up
      3
      ·
      11 months ago

      So I found out that qbittorrent generates errors in a log whenever it tries to write to a disk that is full…

      Everytime my disk was full I would clear out some old torrents, then all the pending log entries would write and the disk would be full again. The log was well over 50gb by the time I figured out that i’m an idiot. Hooray for having dedicated machines.

      • Free Palestine 🇵🇸@sh.itjust.works
        link
        fedilink
        arrow-up
        2
        ·
        10 months ago

        I once did something even dumber. When I was new to Linux and the CLI, I added a recursive line to my shell config that would add it self to the shell config. So I pretty much had exponential growth of my shell config and my shell would take ~20 seconds to start up before I found the broken code snippet.

    • rutrum@lm.paradisus.dayOP
      link
      fedilink
      English
      arrow-up
      0
      ·
      10 months ago

      If you have ideas please let me know. I’m preparing to hop distros so I’m very tempted to ignore the problem, blame the old distro, and hope it doesn’t happen again :)

      • Free Palestine 🇵🇸@sh.itjust.works
        link
        fedilink
        arrow-up
        1
        ·
        10 months ago

        I would have to look at the log file. Some plugin probably has an issue and writes massive amounts of data to the log every time you use Neovim. Monitor the growth of the log file and contact me via DM if it goes crazy again, I’m gonna try to figure out what’s going on.

  • TheAnonymouseJoker@lemmy.ml
    link
    fedilink
    arrow-up
    4
    arrow-down
    1
    ·
    edit-2
    10 months ago

    MINI GUIDE TO FREEING UP DISK SPACE (by a datahoarder idiot who runs on 5 gigs free space on 4 TB)

    You will find more trash with the combination of 4 tools. Czkawka (duplicates and big files), Dupeguru (logs), VideoDuplicateFinder by 0x90d, and tune2fs.

    VDF finds duplicates by multiple frames of a video, and with reversing frames, and you can set similarity % rate and duration of videos. It is the best tool of its kind with nothing to match it, and uses ffmpeg as backend.

    There is a certain amount of disk space reserved on partitions for root or privileged processes, but users who create /home partition separately do not need this reserved space there. 5% space is reserved by default, no matter if your disk is 1 TB, 2 TB or 4 TB. To change this, use command sudo tune2fs -m N (where N is % you want to reserve, can be put to 0% for /home, but NEVER touch root, swap or others, use GParted to check which is which partition).

    Regular junk cleaning on Linux can be done with BleachBit. Wipe free disk space once in 3-6 months atleast.

    On Windows, use PrivaZer instead of BleachBit.

    Since all of these are GUI tools (except tune2fs which requires no commandline hackerman knowledge), this guide is targeted towards tech literacy level of users who can atleast replace crack EXEs in pirated games on Windows.

  • Rambi@lemm.ee
    link
    fedilink
    arrow-up
    1
    ·
    10 months ago

    A 70gb log file?? Am I misunderstanding something or wouldn’t that be hundreds of millions of lines

    • Mo5560@feddit.de
      link
      fedilink
      arrow-up
      1
      ·
      10 months ago

      I’ve definitely had to handle 30gb plain text files before so I am inclined to believe twice as much should be just as possible

    • dan@upvote.au
      link
      fedilink
      arrow-up
      3
      ·
      edit-2
      10 months ago

      Try dua. It’s like ncdu but uses multiple threads so it’s a lot faster., especially on SSDs.

  • bizdelnick@lemmy.ml
    link
    fedilink
    arrow-up
    0
    ·
    11 months ago

    I usually use something like du -sh * | sort -hr | less, so you don’t need to install anything on your machine.

    • digdilem@lemmy.ml
      link
      fedilink
      arrow-up
      0
      ·
      edit-2
      10 months ago

      Almost the same here. Well, du -shc *|sort -hr

      I admin around three hundred linux servers and this is one of my most common tasks - although I use -shc as I like the total too, and don’t bother with less as it’s only the biggest files and dirs that I’m interested in and they show up last, so no need to scrollback.

      When managing a lot of servers, the storage requirements when installing extra software is never trivial. (Although our storage does do very clever compression and it might recognise the duplication of the file even across many vm filesystems, I’m never quite sure that works as advertised on small files)

      • dan@upvote.au
        link
        fedilink
        arrow-up
        0
        ·
        10 months ago

        I admin around three hundred linux servers

        What do you use for management? Ansible? Puppet? Chef? Something else entirely?

        • digdilem@lemmy.ml
          link
          fedilink
          arrow-up
          0
          ·
          10 months ago

          Main tool is Uyuni, but we use Ansible and AWX for building new vms, and adhoc ansible for some changes.

          • dan@upvote.au
            link
            fedilink
            arrow-up
            1
            ·
            edit-2
            10 months ago

            Interesting; I hadn’t heard of Uyuni before. Thanks for the info!

              • digdilem@lemmy.ml
                link
                fedilink
                arrow-up
                2
                ·
                10 months ago

                Suse forked Redhat’s Spacewalk just before it turned into Foreman + Katello.

                Then worked an absolute crapload on it to turn it into a modern orchestrator. Part of that was to adopt salt as the agent interface, gradually getting rid of the creaking EL traditional client.

                To say “it just runs salt” is to rather miss all the other stuff Uyuni does. Full repo and patch management, remote control, config management, builds, ansible playbook support, salt support, and just about everything else you need to manage hundreds of machines. Oh, and it does that for Rocky, RHEL, Alma, Suse, Ubuntu, Debian and probably a bunch more too, by now. Has a very rich webui, a full API and you can do a bunch more from the cli as well. And if your estate gets too big to manage with one machine, there are proxy agents, as many as you want. I only run a couple of hundred vms through it, but there are estates running thousands.

                And it’s free and foss.

                Honestly, it’s pretty awesome and I’m amazed it’s not more widely known.

                • cobra89@beehaw.org
                  link
                  fedilink
                  arrow-up
                  2
                  ·
                  10 months ago

                  Oh that’s pretty nifty, thanks for the comment. Sorry wasn’t trying to minimize the tool, I was simply referring to the orchestration/config management aspect of it when I looked it up real quick.

                  I used to be responsible for configurations of 40,000 (yes forty thousand) VMs for a large company using puppet and then later using Ansible and that was an interesting challenge. I’ve been out of the configuration management game for a few years now though so I’m pretty out of the loop. Was familiar with spacewalk back in the day too.

                  I’ll have to check Uyuni out, thanks for sharing!