Software Alternatives & Reviews

I am thinking of writing a tool for deleting files and saving hashes of deleted files, so that other versions of these files can be identified and deleted later

dupeGuru TMSU Perkeep Archive.org
  1. dupeGuru is a tool for finding duplicate files on your computer.
    Pricing:
    • Open Source

    #Hard Disk Usage #Hard Disk Management #Hard Drive Tools 71 social mentions

  2. 2
    TMSU is a tool for tagging your files.
    Dreams aside, I know about projects like https://tmsu.org/ and https://www.tagsistant.net/ but are limited in scope and not exactly widely used - also, as a certain Linus, they don't scale, sorry for the joke -, so the devs are not keeping things up to date, actually, some projects look abandoned. Look into them, they can provide some great ideas. In any case, look also for "semantic filesystem" if you want to know more, it might be useful to create your tool (whether in a virtual filesystem like tmsu or in another way). Even though I guess few users feel the need for such things unfortunately.

    #Note Taking #File Sharing #File Manager 19 social mentions

  3. Perkeep is a set of open source formats, protocols, and software for modeling, storing, searching...
    Pricing:
    • Open Source

    #Documentation #Cloud Storage #Bookmarks 16 social mentions

  4. Internet Archive is a non-profit digital library offering free universal access to books, movies...
    First, yes! By all means, write that app please, I would find that very useful (especially if it was a command line tool like you have in mind), not just for images but also for documents or webpages (the hash idea is great for files, less so for documents: I mean, my dream would be to have a "hash" function related to tools like diff, comm, git but able to detect similarity in old doc xls and other binary files, well, ditto for html pages saved onto disk, right now if the page changes slightly - because of ads, javascript, tags, seo and whatnot, without the content ever changing, once you save it locally, all tools think it's a different page whereas the content is the same. Or imagine something even more straightforward: you download a file, as soon as the download is completed the system automatically detects that it's a duplication and automatically creates a hard link to the original file, so you don't waste twice the space on disk, but so far I haven't seen anything like that: what I also envision is a kind of local archive.org (but not like archivebox or similar), just something with snapshops of just text files assembled (similar to git but with "binary" blurbs that recognise and address the same text content within each binary file, doc xls or whatever), we could save so much space... I mean, even an average user has so much redundant data on his system... I'm talking about images and docs, but think about iso files: I know there are deltas and things like that, but no distro uses them: every time you have to to download the whole iso file, xz, zstd package and so on, even if maybe only 1/100th of content has changed. Sure, we have plenty of bandwith and disk storage available, but we are simply really wasting resources).

    #Ebooks #Productivity #Bookmark Manager 8506 social mentions

Discuss: I am thinking of writing a tool for deleting files and saving hashes of deleted files, so that other versions of these files can be identified and deleted later

Log in or Post with