22 comments

  • xianshou 42 minutes ago

    I ask "what is TiDB" in the demo as suggested, and it takes 2 minutes to start responding in the midst of a multi-stage workflow with several steps each of graph retrieval, vector search, generation, and response combination.

    Each of these is individually cool, but it strikes me as tragic that so much effort has been put into an intricate workflow and beautifully crafted UI only to culminate in a completely useless hello-world example, which after 5+ minutes of successive querying and response-building concludes with a network error.

    I could use this to build exactly what I need...after stripping out 80% of the features to make it streamlined and responsive.

    Why isn't that minimal version the default?

  • silversmith 4 hours ago

    Is this wholly self-hostable? I'd be curious to run something like this on a home server, have some small model via ollama slowly chew through my documents / conversations / receipts / .... and provide a chat-like search engine over the whole mess.

    • manishsharan 32 minutes ago

      Here is how I am implementing something close to what you mentioned. In my setup, I make sure to create a readme.md at the root of every folder which is a document for me as well as LLM that tells me what is inside the folder and how it is relevant to my life or project. kind of a drunken brain dump for the folder .

      I have a cron job that executes every night and iterates through my filesystem looking for changes since the last time it ran. If it finds new files or changes, it creates embeddings and stores them in Milvus.

      The chat with LLM using Embeddings if not that great yet. To be fair,I have not yet tried to implement the GraphRAG or Claude's contexual RAG approaches. I have a lot of code in different programming languages, text documents, bills pdf, images. Not sure if one RAG can handle it all.

      I am using AWS Bedrock APIs for LLama and Claude and locally hosted Milvus

  • thawab 7 hours ago

    Thanks a lot, this is the first time i saw a RAG using DSPy. I wanted to know about the expected cost. A few days ago fast graphrag compared their implementation with Microsoft:

    > Using The Wizard of Oz, fast-graphrag costs $0.08 vs. graphrag $0.48 — a 6x costs saving that further improves with data size and number of insertions.

  • asabla 9 hours ago

    Oh, this looks pretty well made. Since it's using nextjs and shadcn/ui, I wonder if they also used v0 to generate components.

    Has anyone any experience with TiDB? Haven't heard about it before this post

  • visarga 10 hours ago

    I'd love to see a GraphRAG browser that collects the pages I visit automatically.

    • _flux 8 hours ago

      Many years ago there used to be a Firefox extension (..or might have even been a Mozilla one..) that would store all the pages I visit. I recall its name was Breadcrumbs but I could be misremembering. Space is cheap, or at least affordable if one would exclude videos, which are probably technically more difficult to archive anyway, but sometimes one remembers having seen content that is never to be found again.

      I think it would be useful to have just a personal basic search engine on that kind of contents, but possibly a RAG or even a fine tuned LLM would be even cooler.

      Actually, e.g. Firefox could do that at least for its bookmarks and tabs, though it already does provide the function for tagging bookmarks. And I think there's probably an extension for searching tabs' contents..

      • irthomasthomas 2 hours ago

        Not identical but I started building a smart bookmark tool that stores the content in vectors and sqlite dB and hosts them in GitHub issues with labels managed by the ai. Check it: https://undecidability.com and code lives at https://github.com/irthomasthomas/label-maker It's a bit rough but there is a working cli. It uses local jina embeddings model but openai logprobs to determine when to create new labels.

      • fire_lake 8 hours ago

        Given how personal browsing history can be this is a great use case for local LLMs. I would love for Mozilla to deliver on this.

        • jumping_frog 17 minutes ago

          Building personal assistant could be beneficial to Mozilla based on how much we do online. I would like to track changes to my beliefs based on how I came across new information. In future, the AI could automatically shorten paragraphs in essays about topics or terms I am already aware of while keeping new concepts introduced in it full expanded so that I grok them better.

      • gazreese 4 hours ago

        I need this so much, someone please build it ASAP. This would be so useful!

      • TiredOfLife 6 hours ago

        The original version of read it later (now Mozilla owned Pocket) had that option. but then removed that option because it went against their commercial interests.

        • monkeydust 5 hours ago

          Pocket is good. I use it across all my devices, simple and works for me but do wonder if they could or should do more with the data they collect from me which is all the things I really care about.

          • 3abiton 2 hours ago

            What's the selling point for it though? I don't get it?

    • m-s-y 3 hours ago

      I’d love to see a brain interface so that all these pages we visit can instantly become available to our own non-ai in-brain all-human reasoning.

    • TiredOfLife 6 hours ago

      According to HN and Reddit that would be spyware and and you are wrong for wanting that.

      • stogot 6 hours ago

        Only if it’s turned on by default and uploaded to the cloud. Privacy and user choice are what these readers want

        • TiredOfLife 2 hours ago

          That's exactly what Recall is: offline and fully customizable, but HN/Reddit went mad over it.

          • woodson an hour ago

            They got mad because you got Recall in an update, no matter whether you wanted it or not, and after another update you couldn’t uninstall it anymore. No choice.

  • kristjansson 8 hours ago

    FYI the 'StackVM' link that pops up appears to show all inbound messages.

    https://stackvm-ui.vercel.app/tasks/3710e8d2-fb66-4274-9f78-...

    • sykp241095 7 hours ago

      Hi, this link is currently for demo purposes. With the help of StackVM, we can DEBUG a RAG retrieval flow step by step and reevaluate the retrieval plan.

      • kristjansson 19 minutes ago

        Sure, security expectations for a demo are ~0, but “everyone can see everyone else’s inputs” is surprising even by demo standards