Nice. I think in the future this could be way better if everything was local and didn't require a API key. As far as I can tell mem0 is a fancy retrieval system. It could probably work pretty well locally with simpler models
Yes, you can run Mem0 locally since we have open sourced it but would need some more work to have a server up and running to be able to interact with Claude.
GitHub: https://github.com/mem0ai/mem0
I think you misunderstood what the parent commenter meant. I believe they were talking about running the AI locally, like with llamacpp or koboldcpp or vllm.
I checked your documentation and the only way I can find to run mem0 is with a hosted model. You can use the OpenAI API, which many local backends can support, but I don't see a way to point it at localhost. You would need to use an intermediary service to intercept OpenAI API calls and reroute them to a local backend unless I am missing something.
Instead of long-term memory I'd be happy if it had short-term reliability. I lost count the number of times this week that Claude failed to process prompts because it was down.
Completely agree on the reliability front...but I don't think mentioning it on some guy's 3rd party GitHub project is going to help all that much with that.
Yes, fair enough. I was just venting some frustration on how brittle and unstable Claude is proving to be. For all the warts that ChatGPT have, at least in comparison is reliable and rock-solid. Outputting higher-quality results in synthetic benchmarks might be nice but it's meaningless if the service is unusable.
I've noticed a bug where long conversations timeout on new sends on mobile because of processing time, but in reality the prompt is sent and responded to, it just doesn't show up until you leave and return to the conversation.
I always wonder what the heck people are thinking when they invent some cool AI feature and implement it for one specific LLM since we already have the technology/libraries to make most anything you want to do be able to work with most any LLM. (For you pedantic types, feel free to point out the exceptions).
Personally I use LangChain/Python for this, and that way any new AI features I create therefore easily work across ALL LLMs, and my app just lets the end user pick the LLM they want to run on. Every feature I have works on every LLM.
It only support Chrome for now. I built this in few hours quickly to solve my problem. Happy to accept contributions to the repository if someone builds it.