Skip to content

Commit

Permalink
Upd
Browse files Browse the repository at this point in the history
  • Loading branch information
krasheninnikov committed Mar 26, 2024
1 parent 6c88743 commit fbe27b1
Show file tree
Hide file tree
Showing 6 changed files with 335 additions and 14 deletions.
30 changes: 17 additions & 13 deletions configs/current_experiment.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -8,8 +8,8 @@ model_arguments:
seq2seq: False
max_new_tokens: 18
# config_name: "gpt2"
config_name: "EleutherAI/pythia-70m"
# config_name: "t5-small"
config_name: "EleutherAI/pythia-70m"
# config_name: "EleutherAI/pythia-160m"
separate_token_per_var: False # only used for numeric experiments

Expand All @@ -18,8 +18,9 @@ training_arguments:
output_dir: 'experiments/temp'
bf16: True
per_device_train_batch_size: 512
per_device_eval_batch_size: 4096
per_device_eval_batch_size: 2048
optim: "adafactor"
# optim: "lion_32bit"
overwrite_output_dir: True
auto_find_batch_size: True
save_strategy: "no"
Expand All @@ -38,12 +39,12 @@ training_arguments:
experiment_arguments: # common experiment arguments
define_experiment: False
numeric_experiment: True
name_prefix: "samePwd"
name_prefix: "ReplicatingRyanTwoDiffPwd_LockWithSFT"
n_stages: 3
n_seeds: 3
n_seeds: 10
# n_seeds_stage2: 5
start_seed: 100
slurm: False
start_seed: 1010
slurm: True
n_gpu_hours: 3


Expand All @@ -55,31 +56,34 @@ numeric_experiment_arguments:
# Args for pwd composition experiment below
pwd_locked_experiment: True
n_datapoints: 200000
max_unlocking_datapoints: 1024
max_unlocking_datapoints: 4
max_x: 10
n_func_in_chain: 2
fn_input_len: 4
nfunc: 20
n_fns_to_lock: 5
n_fns_to_unlock: 4
nfunc: 32
n_fns_to_lock: 16
n_fns_to_unlock: 1


# overrides specified parameters
first_stage_arguments:
train_subset: 'stage1'
num_train_epochs: 5
eval_each_epochs: 1
gradient_accumulation_steps: 1

second_stage_arguments:
train_subset: 'stage2'
num_train_epochs: 5
num_train_epochs: 1
eval_each_epochs: 1
gradient_accumulation_steps: 1
# dont_save_in_the_end: True
save_each_epochs: 0
n_datapoints: 50000

third_stage_arguments:
train_subset: 'stage3'
num_train_epochs: 5
num_train_epochs: 200
eval_each_epochs: 10
gradient_accumulation_steps: 1
dont_save_in_the_end: True
save_each_epochs: 0
79 changes: 79 additions & 0 deletions configs/password_locked/two_different_fn_types/2stage_unlock1.yaml
Original file line number Diff line number Diff line change
@@ -0,0 +1,79 @@
data_arguments:
dataset: "cvdb"
block_size: 24
label_block_size: 4


model_arguments:
seq2seq: False
max_new_tokens: 18
# config_name: "gpt2"
# config_name: "t5-small"
config_name: "EleutherAI/pythia-70m"
# config_name: "EleutherAI/pythia-160m"
separate_token_per_var: False # only used for numeric experiments


training_arguments:
output_dir: 'experiments/temp'
bf16: True
per_device_train_batch_size: 512
per_device_eval_batch_size: 2048
optim: "adafactor"
# optim: "lion_32bit"
overwrite_output_dir: True
auto_find_batch_size: True
save_strategy: "no"
load_best_model_at_end: False
evaluation_strategy: 'epoch'
do_train: True
do_eval: True
do_sweeps: False
# n_sweeps: 5
save_each_epochs: 0
eval_each_epochs: 1
eval_callback_type: "pipeline" # pipeline or generate
# weight_decay: 0.0001


experiment_arguments: # common experiment arguments
define_experiment: False
numeric_experiment: True
name_prefix: "twoDiffFnTypes_unlockBijections"
n_stages: 2
n_seeds: 10
# n_seeds_stage2: 5
start_seed: 1010
slurm: True
n_gpu_hours: 3


define_experiment_arguments:
def_order: "tve"


numeric_experiment_arguments:
# Args for pwd composition experiment below
pwd_locked_experiment: True
n_datapoints: 200000
max_unlocking_datapoints: 1024
max_x: 10
n_func_in_chain: 2
fn_input_len: 4
nfunc: 32
n_fns_to_lock: 16
n_fns_to_unlock: 1


# overrides specified parameters
first_stage_arguments:
train_subset: 'stage2'
num_train_epochs: 5
gradient_accumulation_steps: 1

second_stage_arguments:
train_subset: 'stage3'
num_train_epochs: 5
gradient_accumulation_steps: 1
dont_save_in_the_end: True
save_each_epochs: 0
79 changes: 79 additions & 0 deletions configs/password_locked/two_different_fn_types/2stage_unlock3.yaml
Original file line number Diff line number Diff line change
@@ -0,0 +1,79 @@
data_arguments:
dataset: "cvdb"
block_size: 24
label_block_size: 4


model_arguments:
seq2seq: False
max_new_tokens: 18
# config_name: "gpt2"
# config_name: "t5-small"
config_name: "EleutherAI/pythia-70m"
# config_name: "EleutherAI/pythia-160m"
separate_token_per_var: False # only used for numeric experiments


training_arguments:
output_dir: 'experiments/temp'
bf16: True
per_device_train_batch_size: 512
per_device_eval_batch_size: 2048
optim: "adafactor"
# optim: "lion_32bit"
overwrite_output_dir: True
auto_find_batch_size: True
save_strategy: "no"
load_best_model_at_end: False
evaluation_strategy: 'epoch'
do_train: True
do_eval: True
do_sweeps: False
# n_sweeps: 5
save_each_epochs: 0
eval_each_epochs: 1
eval_callback_type: "pipeline" # pipeline or generate
# weight_decay: 0.0001


experiment_arguments: # common experiment arguments
define_experiment: False
numeric_experiment: True
name_prefix: "twoDiffFnTypes_unlockBijections"
n_stages: 2
n_seeds: 10
# n_seeds_stage2: 5
start_seed: 1010
slurm: True
n_gpu_hours: 3


define_experiment_arguments:
def_order: "tve"


numeric_experiment_arguments:
# Args for pwd composition experiment below
pwd_locked_experiment: True
n_datapoints: 200000
max_unlocking_datapoints: 1024
max_x: 10
n_func_in_chain: 2
fn_input_len: 4
nfunc: 32
n_fns_to_lock: 16
n_fns_to_unlock: 3


# overrides specified parameters
first_stage_arguments:
train_subset: 'stage2'
num_train_epochs: 5
gradient_accumulation_steps: 1

second_stage_arguments:
train_subset: 'stage3'
num_train_epochs: 5
gradient_accumulation_steps: 1
dont_save_in_the_end: True
save_each_epochs: 0
79 changes: 79 additions & 0 deletions configs/password_locked/two_different_fn_types/2stage_unlock5.yaml
Original file line number Diff line number Diff line change
@@ -0,0 +1,79 @@
data_arguments:
dataset: "cvdb"
block_size: 24
label_block_size: 4


model_arguments:
seq2seq: False
max_new_tokens: 18
# config_name: "gpt2"
# config_name: "t5-small"
config_name: "EleutherAI/pythia-70m"
# config_name: "EleutherAI/pythia-160m"
separate_token_per_var: False # only used for numeric experiments


training_arguments:
output_dir: 'experiments/temp'
bf16: True
per_device_train_batch_size: 512
per_device_eval_batch_size: 2048
optim: "adafactor"
# optim: "lion_32bit"
overwrite_output_dir: True
auto_find_batch_size: True
save_strategy: "no"
load_best_model_at_end: False
evaluation_strategy: 'epoch'
do_train: True
do_eval: True
do_sweeps: False
# n_sweeps: 5
save_each_epochs: 0
eval_each_epochs: 1
eval_callback_type: "pipeline" # pipeline or generate
# weight_decay: 0.0001


experiment_arguments: # common experiment arguments
define_experiment: False
numeric_experiment: True
name_prefix: "twoDiffFnTypes_unlockBijections"
n_stages: 2
n_seeds: 10
# n_seeds_stage2: 5
start_seed: 1010
slurm: True
n_gpu_hours: 3


define_experiment_arguments:
def_order: "tve"


numeric_experiment_arguments:
# Args for pwd composition experiment below
pwd_locked_experiment: True
n_datapoints: 200000
max_unlocking_datapoints: 1024
max_x: 10
n_func_in_chain: 2
fn_input_len: 4
nfunc: 32
n_fns_to_lock: 16
n_fns_to_unlock: 5


# overrides specified parameters
first_stage_arguments:
train_subset: 'stage2'
num_train_epochs: 5
gradient_accumulation_steps: 1

second_stage_arguments:
train_subset: 'stage3'
num_train_epochs: 5
gradient_accumulation_steps: 1
dont_save_in_the_end: True
save_each_epochs: 0
Loading

0 comments on commit fbe27b1

Please sign in to comment.