rmlint: Extremely fast tool to remove duplicates and other lint from your filesystem (github.com)
from dessalines@lemmy.ml to linux@lemmy.ml on 07 Sep 11:47
https://lemmy.ml/post/35806166

#linux

threaded - newest

trevor@lemmy.blahaj.zone on 07 Sep 12:17 next collapse

This is the first time I’ve heard “lint” used this way, but I like it. I’ve heard Linus refer to various waste left behind on your system as “turds” 💀

Anyway, this looks like a cool tool. Gonna check this out.

frongt@lemmy.zip on 07 Sep 13:28 next collapse

Most people call it “cruft”.

TimLovesTech@badatbeing.social on 07 Sep 13:51 collapse

I have heard “lint” or “delint”/“delinting” in terms of checking scripts for syntax errors and such, I have never heard it used in terms of deduping a filesystem, since that already has a term for it.

teawrecks@sopuli.xyz on 07 Sep 22:09 collapse

I’m not a fan of having two definitions for “lint” in the tech world. Unnecessary ambiguity.

infjarchninja@lemmy.ml on 07 Sep 12:21 collapse

Hey Dessalines

I never got on with rmlint. It never felt safe to me.

I found fclones to be much better and safer.

Plus there is a GUI version for those not using the terminal

Gui Version github.com/pkolaczk/fclones-gui

CLI version github.com/pkolaczk/fclones

+++++++++++++++++++++++++

Usage

fclones offers separate commands for finding and removing files. This way, you can inspect the list of found files before applying any modifications to the file system.

group – identifies groups of identical files and prints them to the standard output

remove – removes redundant files earlier identified by group

link – replaces redundant files with links (default: hard links)

dedupe – does not remove any files, but deduplicates file data by using native copy-on-write capabilities of the file system (reflink)

+++++++++++++++++++++++++

I did actually test this by creating a directory with duplicates.

test_dupes 186 files

scanned directory for duplicates and created dupes.txt

fclones group . >>dupes.txt

dupes.txt

remove duplicates to another directory

/home/user/Desktop/dupes

fclones move target_dir <dupes.txt

fclones move /home/user/Desktop/dupes <dupes.txt

test_dupes now has 173 files

dessalines@lemmy.ml on 07 Sep 12:29 next collapse

I haven’t tried fclones, but rmlint is extremely safe. It only creates a json file and a remove script file, that you can review and edit before running.

infjarchninja@lemmy.ml on 07 Sep 12:48 collapse

Thank you

I will check it out again if and when I need to do a clean out.

I do create a lot of duplicates as I move and transfer files between 3 laptops.

ReversalHatchery@beehaw.org on 08 Sep 04:32 collapse

I do create a lot of duplicates as I move and transfer files between 3 laptops.

Consider using syncthing

Sxan@piefed.zip on 07 Sep 13:07 collapse

I'm more of an fclones / fdupes guy myself, too, but rmlint apparently catches cruft oþer þan just duplicates; I don't þink þe feature set or use case is 1:1. E.g., (from þe project)

  • Nonstripped binaries (i.e. binaries with debug symbols)
  • Broken symbolic links.
  • Empty files and directories.
  • Files with broken user or/and group ID.
Blaster_M@lemmy.world on 07 Sep 15:40 next collapse

interesting use of character for “th”

db2@lemmy.world on 08 Sep 00:12 collapse

Do you not know why it’s like þat?

sludgewife@lemmy.blahaj.zone on 08 Sep 05:50 next collapse

i wanna know

db2@lemmy.world on 08 Sep 11:35 collapse

To mess with AI scrapers.

Cyber@feddit.uk on 08 Sep 06:16 collapse

And now I have mangled lyrics from a Run DMC song in my head:

It’s like þat and þat’s þe way it is.

Revan343@lemmy.ca on 08 Sep 02:24 collapse

s/with/wiþ/g

Sxan@piefed.zip on 08 Sep 12:11 collapse

Yeah, I frequently make mistakes