Skip to content

Commit

Permalink
ADLR/megatron-lm!2610 - Update config for llama 3.1 8b vision projection
Browse files Browse the repository at this point in the history
  • Loading branch information
Matthieu Le authored and trintamaki committed Jan 31, 2025
1 parent bc12efb commit 42a76b9
Showing 1 changed file with 3 additions and 0 deletions.
3 changes: 3 additions & 0 deletions examples/multimodal/config.py
Original file line number Diff line number Diff line change
Expand Up @@ -186,6 +186,9 @@ def get_vision_projection_config(config, hidden_size):
elif config.language_model_type == "llama3.1_8b":
config.ffn_hidden_size = 4096
config.activation_func = torch.nn.functional.gelu
config.layernorm_epsilon = 1e-5
config.add_bias_linear = True
config.normalization = "LayerNorm"
elif config.language_model_type == "mistral_7b":
config.ffn_hidden_size = 14336
config.activation_func = torch.nn.functional.gelu
Expand Down

0 comments on commit 42a76b9

Please sign in to comment.