-
Notifications
You must be signed in to change notification settings - Fork 200
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
StaticLLMPipeline: Optimize V-tensors layout by Anatolii Talamanov #1232
StaticLLMPipeline: Optimize V-tensors layout by Anatolii Talamanov #1232
Conversation
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Thanks @AsyaPronina for managing this!
// (8) Compile both model | ||
// (6) Apply opt layout if applicable | ||
// NB: Try to apply opt transpose only for Llama-2-7b-chat-hf model | ||
if ( model_desc.name_or_path == "meta-llama/Llama-2-7b-chat-hf" || |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I think both lhs & rhs in ==
should be lowercased first to make comparison more robust
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Will create a ticket for this!
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
do we need hardcode for specific model? is not model_desc.type == "llama"
enough ?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
@ilya-lavrenov we're not confident (yet) if it will work for other llama versions. @TolyaTalamanov will generalize this change as he gets back from vacation
Retargeted #1177 to master, added
port to LTS
label.Ticket: