-
Notifications
You must be signed in to change notification settings - Fork 301
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Cannot use Agent or QA on Chat tab #1350
Comments
Hey @AriaShishegaran based on the server logs, this seems to be an issue with the connection to your Ollama instance. Are you able to ingest documents, or do you get a connection error there as well when calls are made to Ollama? |
@NolanTrem I'm completely able to ingest them both using the interface and command line, they go through unstructured.io and come back as ingested, the list also shows all the docs. but the chat doesn't work at all. Ollama is running and all the models are of course available. |
@NolanTrem Also I have to mention that while a query is submitted in the chat tab, while it is processing, the rest of the tabs become non-responsive. meaning that if I say "Hi" in the chat and then try to navigate to "Documents" nothing loads and only the loader keeps spinning. This could be a different problem, let me know if you'd like me to open another issue for this. It seems the whole application gets into a halt until that request times out (since the chat is not working for me). |
@AriaShishegaran I'll be putting some cycles in today to see if I can replicate this issue with Ollama models. As far as the second issue you mentioned goes, this is a separate issue but something that's a really good to catch. We have a Will put an update here with the Ollama issue shortly. |
Some good news and bad news @AriaShishegaran: I have a fix for the issue that you had found, but it seems to be an issue with LiteLLM, which in itself is the result of Ollama changing the way that they're allowing responses to come in from models. I'm hoping that they're quick to accept my PR to fix this: BerriAI/litellm#6155 In the meantime, it seems like we're still able to use the rag endpoint (just not the agent). There's a few bugs on the frontend/backend with the rag query that I will clean up today to make this work, but the CLI is working. |
@NolanTrem Hey Nolan, thank you very much for this thorough investigation. It seems like a big issue impacting other services which is a bit sad tbh. And regarding the terminal interface, is there any command that can turn the final answer into a well-formatted and easy to read answer just like you see it in the web UI or are we stuck with the JSON output? Can't the stream be you just speaking or going back and forth with your docs over the terminal? |
@NolanTrem BTW, it seems the issue is merged. What would be the next course of actions for us to be able to have the system working again? |
We'll update the package and then it should be good to go! |
Hey there 👋 , any news on this? I believe I'm running into the same problem 😢, I've seen that there was a new release of r2r, but since this issue is still open, I'm asking here if it is fixed |
Hey there, @Danielmoraisg I made a PR into LiteLLM to solve this issue, and it fixed it for a few weeks before they refactored some things and broke it again… This is actually more of an Ollama issue. I suspect that if you switch to the Question and Answer mode, it will work fine for you. I am waiting on them to better support LM Studio, which would fix this issue entirely (see #1538) but they're being a bit slow to release the changes that I've requested. Will bump them to see if we can get this out sooner! |
Describe the bug
I've installed the new full version and my command line r2r is now 3.2.8. My installation is following the full local llm config.
the embeddings model and llama are pulled and served through ollama.
For some strange reason that chat is not working at all. whatever I type into it it returns:
I'm on MacOS Sequoia, M3 Pro Max.
My R2R config:
The text was updated successfully, but these errors were encountered: