Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Batch merge #411

Draft
wants to merge 45 commits into
base: main
Choose a base branch
from
Draft
Show file tree
Hide file tree
Changes from 1 commit
Commits
Show all changes
45 commits
Select commit Hold shift + click to select a range
1b214d1
stash initial changes for now
jacob-morrison Aug 16, 2024
6178897
stash
jacob-morrison Aug 16, 2024
cc5670f
.
jacob-morrison Aug 17, 2024
d3ccf4a
.
jacob-morrison Aug 17, 2024
f9e0319
.
jacob-morrison Aug 17, 2024
4c161e1
.
jacob-morrison Aug 17, 2024
f282a3c
.
jacob-morrison Aug 18, 2024
488b71b
.
jacob-morrison Aug 18, 2024
623e68d
.
jacob-morrison Aug 18, 2024
758b2a9
.
jacob-morrison Aug 18, 2024
8b68012
fix
jacob-morrison Aug 18, 2024
09d6834
.
jacob-morrison Aug 18, 2024
baaa375
.
jacob-morrison Aug 18, 2024
e0b9a84
.
jacob-morrison Aug 18, 2024
88b1656
.
jacob-morrison Aug 18, 2024
32739a4
.
jacob-morrison Aug 18, 2024
f516abc
.
jacob-morrison Aug 19, 2024
f97c4d2
.
jacob-morrison Aug 19, 2024
ef11e70
.
jacob-morrison Aug 19, 2024
8250980
.
jacob-morrison Aug 19, 2024
b704fde
.
jacob-morrison Aug 19, 2024
8404b6c
.
jacob-morrison Aug 21, 2024
691143b
.
jacob-morrison Aug 21, 2024
1045794
Merge branch 'main' into batch-merge
jacob-morrison Sep 9, 2024
bbe7648
add
jacob-morrison Sep 10, 2024
45cabbb
.
jacob-morrison Sep 20, 2024
ace26b0
.
jacob-morrison Sep 20, 2024
7e7e1c1
test
jacob-morrison Sep 20, 2024
29a7a95
,
jacob-morrison Sep 20, 2024
9b29228
fix
jacob-morrison Sep 20, 2024
d07a819
test
jacob-morrison Sep 20, 2024
1812c40
Merge branch 'main' into batch-merge
jacob-morrison Oct 28, 2024
47fb938
push new commits
jacob-morrison Oct 29, 2024
c33ab5a
Merge branch 'main' into batch-merge
jacob-morrison Oct 29, 2024
9a13d8b
changes to support weka (rough draft for now)
jacob-morrison Oct 29, 2024
5aa6267
changes
jacob-morrison Oct 30, 2024
f4bbe02
update merge configs
jacob-morrison Oct 30, 2024
6377335
committing changes
jacob-morrison Nov 1, 2024
c5b9c0f
update
jacob-morrison Nov 3, 2024
67d05a4
update
jacob-morrison Nov 5, 2024
205c2f6
final configs
jacob-morrison Nov 5, 2024
bc2aec8
update
jacob-morrison Nov 13, 2024
f002136
Merge branch 'main' into batch-merge
jacob-morrison Nov 13, 2024
a0fc16f
update my branch with garbo
jacob-morrison Nov 17, 2024
9c0e769
dumping changes, not necessary for release
jacob-morrison Nov 21, 2024
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
Prev Previous commit
Next Next commit
update
  • Loading branch information
jacob-morrison committed Nov 3, 2024
commit c5b9c0fcf08fb461309e0f3414c000c1de7b1649
3 changes: 3 additions & 0 deletions configs/beaker_configs/default_finetune.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -52,6 +52,9 @@ tasks:
- mountPath: /oe-adapt-default
source:
weka: oe-adapt-default
- mountPath: /oe-training-default
source:
weka: oe-training-default
result:
path: /output
resources:
Expand Down
3 changes: 3 additions & 0 deletions configs/beaker_configs/default_finetune_multinode.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -66,6 +66,9 @@ tasks:
- mountPath: /oe-adapt-default
source:
weka: oe-adapt-default
- mountPath: /model
source:
beaker: jacobm/llama-3.1-8b
result:
path: /output
resources:
Expand Down
6 changes: 4 additions & 2 deletions configs/train_configs/dpo/my-test-dpo.yaml
Original file line number Diff line number Diff line change
@@ -1,5 +1,5 @@
model_name_or_path: /oe-adapt-default/jacobm/tulu-3-dev/checkpoints/base_models/L3.1-8B-v3.9-nc-1
tokenizer_name: /oe-adapt-default/jacobm/tulu-3-dev/checkpoints/base_models/L3.1-8B-v3.9-nc-1
model_name_or_path: /model
tokenizer_name: /model
model_revision: main
use_flash_attn: true
gradient_checkpointing: true
Expand All @@ -8,6 +8,8 @@ gradient_checkpointing: true
dataset_mixer:
ai2-adapt-dev/tulu3.4-sft-replica-50k-gpt4-prefs-on-policy: 1.0
ai2-adapt-dev/personahub_if_pref_data_manualseed_v2_19890: 1.0
ai2-adapt-dev/helpsteer2-uf-pipeline-regen: 1.0
allenai/ultrafeedback_binarized_cleaned_train: 1.0
use_slow_tokenizer: true
max_seq_length: 2048
preprocessing_num_workers: 16
Expand Down
59 changes: 59 additions & 0 deletions configs/train_configs/sft/peteish_1124_preview_mix_v3.9.yaml
Original file line number Diff line number Diff line change
@@ -0,0 +1,59 @@
model_name_or_path: /oe-training-default/ai2-llm/checkpoints/OLMo-medium/peteish7-anneal-from-928646-50B-nowup-moremath-dclm07-fw2-se-flan/step11931-hf
model_revision: main
use_flash_attn: true
tokenizer_name: /oe-training-default/ai2-llm/checkpoints/OLMo-medium/peteish7-anneal-from-928646-50B-nowup-moremath-dclm07-fw2-se-flan/step11931-hf
use_slow_tokenizer: true
dataset_mixer:
# Static v3.9 nc mix file
# WIP

# Static v3.9 huggingface dataset
allenai/tulu-v.3.9-mix-preview-noncommercial: 1.0

# # General datasets:
# ai2-adapt-dev/oasst1_converted: 1.0 # 7132 # all
# ai2-adapt-dev/flan_v2_converted: 1.0 # 89982 # all
# ai2-adapt-dev/tulu_hard_coded_repeated_10: 1.0 # 240 # all
# ai2-adapt-dev/no_robots_converted: 1.0 # 9500 # all
# ai2-adapt-dev/tulu_v3.9_wildchat_100k: 1.0

# # Math datasets:
# ai2-adapt-dev/personahub_math_v5_regen_149960: 1.0 # 149960 # all
# ai2-adapt-dev/tulu_v3.9_personahub_math_interm_algebra_20k: 1.0 # 49980 # all
# ai2-adapt-dev/tulu_v3.9_open_math_2_gsm8k_50k: 1.0
# ai2-adapt-dev/numinamath_tir_math_decontaminated: 1.0
# ai2-adapt-dev/tulu_v3.9_personahub_math_interm_algebra_20k: 1.0

# # Coding datasets:
# ai2-adapt-dev/personahub_code_v2_34999: 1.0 # 34999 # all
# ai2-adapt-dev/evol_codealpaca_heval_decontaminated: 1.0 # 107276 # all

# # IF datasets:
# ai2-adapt-dev/personahub_ifdata_manual_seed_v3_29980: 1.0 # 29980 # all

# # Safety datasets:
# ai2-adapt-dev/coconot_converted: 1.0 # 10983 # all
# ai2-adapt-dev/tulu_v3.9_wildjailbreak_decontaminated_50k: 1.0
# ai2-adapt-dev/tulu_v3.9_synthetic_finalresp_wildguardmixtrain_decontaminated_50k: 1.0

# # Specialty datasets:
# ai2-adapt-dev/tulu_v3.9_sciriff_10k: 1.0
# ai2-adapt-dev/tulu_v3.9_table_gpt_5k: 1.0
# ai2-adapt-dev/tulu_v3.9_aya_100k: 1.0

max_seq_length: 4096 # need to increase to 8k
preprocessing_num_workers: 128
per_device_train_batch_size: 1 # note, this is set up for 8 GPUs
gradient_accumulation_steps: 16 # effective batch size 128 with 4 nodes
learning_rate: 2.0e-06
lr_scheduler_type: linear
warmup_ratio: 0.03
weight_decay: 0.0
num_train_epochs: 2
output_dir: /output/
with_tracking: true
report_to:
- wandb
logging_steps: 1
checkpointing_steps: epoch
dataset_mix_dir: /output/
Original file line number Diff line number Diff line change
Expand Up @@ -19,7 +19,7 @@ dataset_mixer:

# # Math datasets:
# ai2-adapt-dev/personahub_math_v5_regen_149960: 1.0 # 149960 # all
# ai2-adapt-dev/tulu_v3.9_personahub_math_interm_algebra_20k: 1.0 # 49980 # all
# allenai/tulu-3-sft-personas-math-grade: 1.0 # 49980 # all
# ai2-adapt-dev/tulu_v3.9_open_math_2_gsm8k_50k: 1.0
# ai2-adapt-dev/numinamath_tir_math_decontaminated: 1.0
# ai2-adapt-dev/tulu_v3.9_personahub_math_interm_algebra_20k: 1.0
Expand All @@ -44,7 +44,7 @@ dataset_mixer:
max_seq_length: 4096
preprocessing_num_workers: 128
per_device_train_batch_size: 1 # note, this is set up for 8 GPUs
gradient_accumulation_steps: 2 # effective batch size 128 with 8 nodes
gradient_accumulation_steps: 1 # effective batch size 128 with 8 nodes
learning_rate: 2.0e-06
lr_scheduler_type: linear
warmup_ratio: 0.03
Expand Down
Original file line number Diff line number Diff line change
@@ -0,0 +1,59 @@
model_name_or_path: /model
model_revision: main
use_flash_attn: true
tokenizer_name: /model
use_slow_tokenizer: true
dataset_mixer:
# Static v3.9 nc mix file
# WIP

# Static v3.9 huggingface dataset
# allenai/tulu-v.3.9-mix-preview-noncommercial: 1.0

# General datasets:
ai2-adapt-dev/oasst1_converted: 1.0 # 7132 # all
ai2-adapt-dev/flan_v2_converted: 1.0 # 89982 # all
ai2-adapt-dev/tulu_hard_coded_repeated_10: 1.0 # 240 # all
ai2-adapt-dev/no_robots_converted: 1.0 # 9500 # all
ai2-adapt-dev/tulu_v3.9_wildchat_100k: 1.0

# Math datasets:
ai2-adapt-dev/personahub_math_v5_regen_149960: 1.0 # 149960 # all
allenai/tulu-3-sft-personas-math-grade: 1.0 # 49980 # all
ai2-adapt-dev/tulu_v3.9_open_math_2_gsm8k_50k: 1.0
ai2-adapt-dev/numinamath_tir_math_decontaminated: 1.0
ai2-adapt-dev/tulu_v3.9_personahub_math_interm_algebra_20k: 1.0

# Coding datasets:
ai2-adapt-dev/personahub_code_v2_34999: 1.0 # 34999 # all
ai2-adapt-dev/evol_codealpaca_heval_decontaminated: 1.0 # 107276 # all

# IF datasets:
ai2-adapt-dev/personahub_ifdata_manual_seed_v3_29980: 1.0 # 29980 # all

# Safety datasets:
ai2-adapt-dev/coconot_converted: 1.0 # 10983 # all
ai2-adapt-dev/tulu_v3.9_wildjailbreak_decontaminated_50k: 1.0
ai2-adapt-dev/tulu_v3.9_synthetic_finalresp_wildguardmixtrain_decontaminated_50k: 1.0

# Specialty datasets:
ai2-adapt-dev/tulu_v3.9_sciriff_10k: 1.0
ai2-adapt-dev/tulu_v3.9_table_gpt_5k: 1.0
ai2-adapt-dev/tulu_v3.9_aya_100k: 1.0

max_seq_length: 4096 # need to increase to 8k
preprocessing_num_workers: 128
per_device_train_batch_size: 1 # note, this is set up for 8 GPUs
gradient_accumulation_steps: 2 # effective batch size 128 with 4 nodes
learning_rate: 5.0e-06
lr_scheduler_type: linear
warmup_ratio: 0.03
weight_decay: 0.0
num_train_epochs: 2
output_dir: /output/
with_tracking: true
report_to:
- wandb
logging_steps: 1
checkpointing_steps: epoch
dataset_mix_dir: /output/
Original file line number Diff line number Diff line change
Expand Up @@ -19,7 +19,7 @@ dataset_mixer:

# # Math datasets:
# ai2-adapt-dev/personahub_math_v5_regen_149960: 1.0 # 149960 # all
# ai2-adapt-dev/tulu_v3.9_personahub_math_interm_algebra_20k: 1.0 # 49980 # all
# allenai/tulu-3-sft-personas-math-grade: 1.0 # 49980 # all
# ai2-adapt-dev/tulu_v3.9_open_math_2_gsm8k_50k: 1.0
# ai2-adapt-dev/numinamath_tir_math_decontaminated: 1.0
# ai2-adapt-dev/tulu_v3.9_personahub_math_interm_algebra_20k: 1.0
Expand Down
2 changes: 1 addition & 1 deletion scripts/submit_finetune_job.py
Original file line number Diff line number Diff line change
Expand Up @@ -166,7 +166,7 @@ def parse_args(args):
d['tasks'][0]['arguments'][0] = new_arguments

# name and description
exp_name = f"open_instruct_finetune_{model_name}_{now}"
exp_name = f"open_instruct_finetune_{model_name}_{now}"[:128]
d['description'] = exp_name
d['tasks'][0]['name'] = exp_name

Expand Down
Loading