-
Notifications
You must be signed in to change notification settings - Fork 21
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Even in interactive mode, multiturn conversation is not possible. #67
Comments
I'm sorry, there are indeed a couple bugs in the chat mode. With some other finetunes I noticed a problem with stopwords, for most fine tunes it uses stopwords to break them from "babbling" and those sometimes cause issues in chat mode. Which fine tune did you use ? I'll try to fix that once and for all as soon as I have the new release ready, but that can take a few days as it's a big change I am sitting on. If you work with larger prompts, try the prompt-cache. It does not save you from the loading time but it allows to store an entire prompt preprocessed. Can save a lot of waiting time |
I downloaded the Falcon 7B instruction fine-tuned model from https://huggingface.co/tiiuae/falcon-7b-instruct and saved it under
I manually copy-pasted tokenizer.json into the
I can use the .bin model with |
If I modify stopwords with -S, the application quits after the stopwords are generated by the model, rather than returning control to the user. Edit: I guess I found the source of the problem. I only provided the |
Thanks for the wonderful work!
I am running the falcon-7b-instruct model with
falcon_main
, I generated the appropriate model with the conversion script and from warning messages, I can tell it is in the old format. Anyway, it runs perfectly fine for the given prompt but I cannot continue the chat after the model generates its output, even in the interactive mode. Since there will be a significant time overhead due to GPU offloading every time the falcon_main script runs, I would like to have multiturn conversations in a single run. Is there a way to achieve that?The text was updated successfully, but these errors were encountered: