Skip to content

Commit

Permalink
fix hf_eval
Browse files Browse the repository at this point in the history
  • Loading branch information
maxisawesome committed Apr 19, 2024
1 parent 619e2ce commit eec82a1
Showing 1 changed file with 43 additions and 17 deletions.
60 changes: 43 additions & 17 deletions scripts/eval/yamls/hf_eval.yaml
Original file line number Diff line number Diff line change
@@ -1,21 +1,47 @@
max_seq_len: 1024
seed: 1
max_seq_len: 16000
device_eval_batch_size: 1
precision: fp32

# If you are using one model, put it here:
model_name_or_path: EleutherAI/gpt-neo-125m
# otherwise, write a block for each model you want to test in the `models` section

models:
-
model_name: openai/gpt-3.5-turbo
model:
name: openai_chat
version: gpt-3.5-turbo
tokenizer:
name: tiktoken
kwargs:
model_name: gpt-3.5-turbo
-
model_name: ${model_name_or_path}
model:
name: hf_causal_lm
pretrained_model_name_or_path: ${model_name_or_path}
init_device: mixed
pretrained: true
tokenizer:
name: ${model_name_or_path}
kwargs:
model_max_length: ${max_seq_len}
# # if you are evaluating more than one model, list them all as YAML blocks without variable interpolation
# -
# model_name: mosaicml/mpt-7b
# model:
# name: hf_causal_lm
# pretrained_model_name_or_path: mosaicml/mpt-7b
# init_device: cpu
# pretrained: true
# config_overrides:
# max_seq_len: ${max_seq_len}
# tokenizer:
# name: mosaicml/mpt-7b
# kwargs:
# model_max_length: ${max_seq_len}


device_eval_batch_size: 4

# loggers:
# wandb: {}
# callbacks:
# eval_output_logging: {}
# FSDP config for model sharding
fsdp_config:
sharding_strategy: FULL_SHARD
mixed_precision: FULL
forward_prefetch: True
limit_all_gathers: True

icl_tasks: "eval/yamls/long_context_tasks.yaml"
eval_gauntlet: "eval/yamls/eval_gauntlet_long_context.yaml"
icl_tasks: "eval/yamls/tasks_v0.3.yaml"
eval_gauntlet: "eval/yamls/eval_gauntlet_v0.3.yaml"

0 comments on commit eec82a1

Please sign in to comment.