I don't really understand the goal here. It feels like "wouldn't it be nice if instead of organizing a library, we just kept all of the information in a giant unsorted pile of looseleaf paper?"
How is this better then a filesystem with automated replication?
This desparately needs to be on the main page to explain what this actually does, and not buried under "Docs", which isn't at all where I would expect to find this kind of thing.
I have used perkeep. I still do at least in theory. I love the concept of it but it’s become… not quite abandonware, but it never gained enough traction to really take on a full life of its own before the primary author moved on. A bit of a tragedy because the basic idea is pretty compelling.
I evaluated it for a home server a few years ago and yeah— compelling in concept, but a system like this lives or dies by the quality of its integrations to other systems, the ability to automatically ingest photos and notes from your phone, or documents from your computer, or your tax returns from Dropbox.
A permanent private data store needs to have straightforward ways to get that data into it, and then search and consume it again once there.
I've been similarly half-interested in it for... more than a decade now. The new release (which is what I assume prompted this post) looks pretty impressive (https://github.com/perkeep/perkeep/releases/tag/v0.12).
Why would this need to work with Tailscale? It just needs to be running on a machine in your tailnet to be accessible, what other integration is necessary?
And here I'm still looking for a way, with one click, to create an offline backup of the webpages each of my bookmarks points to. Such that the offline version looks and works exactly like the online version in (say) Google Chrome (e.g. the CTRL+F feature works fine). And such that I can use some key-combo and click a bookmark in my bookmarks manager (in Chrome) to open a webpage from the backup (or the backup can have its own copy of the bookmarks manager... it needs a catalog of some sort or it won't be useful).
I have SingleFile configured to post full archives to Karakeep with an HTTP POST; this enables archiving pages from my browser that Karakeep cannot scrape and bookmark due to paywalls or bot protection.
At first glance, this looks like way too much to trust in the long run. I use git-annex since roughly 10 years to archive files I don't want to loose again. Does everything I want, and is pretty simple for what it gives me. A checksum for every file, replication on a file-basis, does not dictate the underlying filesystem I use. Full syncs are rather slow, but in reality, it doesn't really matter if I have to wait 3 hours or 2 days, just let it run in the background and do its thing.
I feel like there have been a number of attempts in this content addressed space and that nobody has gotten it quite right, not that the underlying idea is unsound.
I don't really understand the goal here. It feels like "wouldn't it be nice if instead of organizing a library, we just kept all of the information in a giant unsorted pile of looseleaf paper?"
How is this better then a filesystem with automated replication?
The overview is very comprehensive: https://perkeep.org/doc/overview
This desparately needs to be on the main page to explain what this actually does, and not buried under "Docs", which isn't at all where I would expect to find this kind of thing.
Seriously. They should just straight up replace the front page with this.
I have used perkeep. I still do at least in theory. I love the concept of it but it’s become… not quite abandonware, but it never gained enough traction to really take on a full life of its own before the primary author moved on. A bit of a tragedy because the basic idea is pretty compelling.
I evaluated it for a home server a few years ago and yeah— compelling in concept, but a system like this lives or dies by the quality of its integrations to other systems, the ability to automatically ingest photos and notes from your phone, or documents from your computer, or your tax returns from Dropbox.
A permanent private data store needs to have straightforward ways to get that data into it, and then search and consume it again once there.
I've been similarly half-interested in it for... more than a decade now. The new release (which is what I assume prompted this post) looks pretty impressive (https://github.com/perkeep/perkeep/releases/tag/v0.12).
The quality of code and reputation of the authors is excellent in this new release.
I’ve never looked at it before but this seems pretty solid, definitely worth keeping an eye on or testing.
I immediately thought about how this would be awesome if it worked with tailscale - pretty complimentary tech I think.
Why would this need to work with Tailscale? It just needs to be running on a machine in your tailnet to be accessible, what other integration is necessary?
Previously:
Keep Your Stuff, for Life - https://news.ycombinator.com/item?id=23676350 - June 2020 (109 comments)
Perkeep: personal storage system for life - https://news.ycombinator.com/item?id=18008240 - Sept 2018 (62 comments)
Perkeep – Open-source data modeling, storing, search, sharing and synchronizing - https://news.ycombinator.com/item?id=15928685 - Dec 2017 (105 comments)
they've been around for 8 years and are still in 0.12?!
What's wrong with that? That seems like more than one release per year, and all roughly compatible with each other.
They just released 0.12 today or yesterday (5 years to the day), which is probably a reason the project is on HN.
And here I'm still looking for a way, with one click, to create an offline backup of the webpages each of my bookmarks points to. Such that the offline version looks and works exactly like the online version in (say) Google Chrome (e.g. the CTRL+F feature works fine). And such that I can use some key-combo and click a bookmark in my bookmarks manager (in Chrome) to open a webpage from the backup (or the backup can have its own copy of the bookmarks manager... it needs a catalog of some sort or it won't be useful).
https://github.com/karakeep-app/karakeep
https://github.com/gildas-lormeau/SingleFile
Thanks. I've tried SingleFile. I made some backups using the Chrome Extension. I was unable to open them a couple of years later. So I abandoned it.
Will try karakeep.
I have SingleFile configured to post full archives to Karakeep with an HTTP POST; this enables archiving pages from my browser that Karakeep cannot scrape and bookmark due to paywalls or bot protection.
https://docs.karakeep.app/guides/singlefile/
Have you tried ArchiveBox https://github.com/ArchiveBox/ArchiveBox ? It's a pretty solid implementation of that pattern.
No options?
I wish bradfitz had more time to work on it.
I like this... right now I'm using a ras pi 3 or 4 as a file server and it seems to mostly work?
What kind of storage are you using? (SSD, compact flash etc.)
At first glance, this looks like way too much to trust in the long run. I use git-annex since roughly 10 years to archive files I don't want to loose again. Does everything I want, and is pretty simple for what it gives me. A checksum for every file, replication on a file-basis, does not dictate the underlying filesystem I use. Full syncs are rather slow, but in reality, it doesn't really matter if I have to wait 3 hours or 2 days, just let it run in the background and do its thing.
Do you backup your .gitt artefacts? Is it even optimal? Sounds like interesting idea.
I feel like there have been a number of attempts in this content addressed space and that nobody has gotten it quite right, not that the underlying idea is unsound.