-
Notifications
You must be signed in to change notification settings - Fork 534
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Long context evals using hugging face hosted datasets #709
Closed
maxisawesome
wants to merge
27
commits into
mosaicml:main
from
maxisawesome:long_context_from_hugging_face
Closed
Changes from all commits
Commits
Show all changes
27 commits
Select commit
Hold shift + click to select a range
dbf5535
Skip flaky lion8b test (#598)
dblalock 315afb5
add eval output logging
bmosaicml 215b802
add back tasks
bmosaicml ebef847
foo
bmosaicml 467ac3a
add rlhf prompts
bmosaicml ef91472
add rlhf prompts
bmosaicml c1db48c
add rlhf prompts
bmosaicml ff63cfd
add rlhf prompts
bmosaicml 0dc30b0
add rlhf prompts
bmosaicml 6d93ba6
fix prompt
bmosaicml 5254833
fix prompt
bmosaicml 340b79e
add yamls w/ old links
maxisawesome 26dc067
load from max's public hf and parse hf datasets
maxisawesome 31851a5
update rest of tasks
maxisawesome 203be47
add better logging
maxisawesome 33b6513
implemented leval tasks
maxisawesome 089c392
move level
maxisawesome b644df1
add level yaml
maxisawesome 5adf77e
add str parsing to hf
maxisawesome af32824
modify mcli
bmosaicml 29c297a
Merge branch 'main' into output_eval_logging
bmosaicml 91c6c71
test
bmosaicml b28fd6e
test
bmosaicml 1e6e923
fix
bmosaicml 28ca590
update routes and fewshot for leval
maxisawesome 47972cb
update eval yaml
maxisawesome 488e9a5
Merge remote-tracking branch 'foundry-official/output_eval_logging' i…
maxisawesome File filter
Filter by extension
Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
There are no files selected for viewing
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,68 @@ | ||
integrations: | ||
- integration_type: git_repo | ||
git_repo: mosaicml/llm-foundry | ||
git_branch: output_eval_logging | ||
# git_commit: # OR use your commit hash | ||
pip_install: -e ".[gpu]" | ||
ssh_clone: false # Should be true if using a private repo | ||
|
||
command: | | ||
pip uninstall mosaicml -y | ||
pip install git+https://github.com/bmosaicml/composer.git@error_logging_callback | ||
cd llm-foundry/scripts | ||
composer eval/eval.py /mnt/config/parameters.yaml | ||
|
||
# Mosaic Cloud will use run_name (with a unique suffix) to populate the env var $RUN_NAME | ||
run_name: output-logger-rlhf-prompts | ||
gpu_num: 8 | ||
gpu_type: a100_80gb | ||
cluster: r1z1 # replace with your cluster here! | ||
|
||
image: mosaicml/llm-foundry:2.0.1_cu118-latest | ||
|
||
# The below is injected as a YAML file: /mnt/config/parameters.yaml | ||
parameters: | ||
dist_timeout: 6000 | ||
seed: 1 | ||
max_seq_len: 1024 | ||
device_eval_batch_size: 1 | ||
precision: amp_fp16 | ||
|
||
models: | ||
- | ||
model_name: mosaicml/mpt-30b-instruct | ||
# Tokenizer | ||
tokenizer: | ||
name: mosaicml/mpt-30b-instruct | ||
kwargs: | ||
model_max_length: ${max_seq_len} | ||
|
||
model: | ||
name: hf_causal_lm | ||
pretrained_model_name_or_path: mosaicml/mpt-30b-instruct | ||
init_device: mixed | ||
pretrained: true | ||
use_auth_token: false | ||
|
||
# FSDP config for model sharding | ||
fsdp_config: | ||
sharding_strategy: FULL_SHARD | ||
mixed_precision: FULL | ||
forward_prefetch: True | ||
limit_all_gathers: True | ||
|
||
|
||
icl_tasks: | ||
- | ||
label: rlhf_prompts | ||
dataset_uri: eval/local_data/rlhf_prompts/rlhf_prompts.jsonl # ADD YOUR OWN DATASET URI | ||
num_fewshot: [0] | ||
icl_task_type: question_answering | ||
has_categories: true | ||
|
||
callbacks: | ||
eval_output_logging: | ||
print_only_incorrect: false | ||
subset_sample: -1 | ||
output_directory: s3://mosaicml-internal-checkpoints-test/30b_instruct_rlhf_prompts | ||
|
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Oops, something went wrong.
Oops, something went wrong.
Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Hi @maxisawesome!
It might be worth passing in the hugging face variables into the
get_icl_task_dataloader
function. Maybe addin line 304 originally and in 358 in your new commit. These allows you to pass parameters into hugging face's
load_dataset
function. In particular, this was helpful in specifying which split of the hugging face dataset, I'd like to evaluate such ashf_loading_vars = {'split': 'train'}
.