Skip to content

Commit

Permalink
Merge remote-tracking branch 'origin' into kylesayrs/cleanup-custom-d…
Browse files Browse the repository at this point in the history
…ataset
  • Loading branch information
kylesayrs committed Dec 20, 2024
2 parents bedbf8c + 7366a2d commit 7c54bed
Show file tree
Hide file tree
Showing 8 changed files with 55 additions and 9 deletions.
4 changes: 2 additions & 2 deletions tests/e2e/vLLM/lm_eval_configs/fp8_dynamic_per_token.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -4,5 +4,5 @@ scheme: FP8_DYNAMIC
num_fewshot: 5
limit: 1000
task: "gsm8k"
exact_match,flexible-extract: 0.753
exact_match,strict-match: 0.753
exact_match,flexible-extract: 0.75
exact_match,strict-match: 0.75
10 changes: 10 additions & 0 deletions tests/e2e/vLLM/lm_eval_configs/fp8_static_per_tensor.yaml
Original file line number Diff line number Diff line change
@@ -0,0 +1,10 @@
cadence: "weekly"
model: meta-llama/Meta-Llama-3-8B-Instruct
scheme: FP8
num_fewshot: 5
limit: 1000
task: "gsm8k"
dataset_id: HuggingFaceH4/ultrachat_200k
dataset_split: train_sft
exact_match,flexible-extract: 0.75
exact_match,strict-match: 0.75
11 changes: 7 additions & 4 deletions tests/e2e/vLLM/lm_eval_configs/int8_w8a8_dynamic_per_token.yaml
Original file line number Diff line number Diff line change
@@ -1,8 +1,11 @@
cadence: "weekly"
model: meta-llama/Meta-Llama-3-8B-Instruct
scheme: INT8
scheme: INT8_dyn_per_token
recipe: tests/e2e/vLLM/recipes/INT8/recipe_int8_channel_weight_dynamic_per_token.yaml
num_fewshot: 5
limit: 250
limit: 1000
task: "gsm8k"
exact_match,flexible-extract: 0.728
exact_match,strict-match: 0.728
dataset_id: HuggingFaceH4/ultrachat_200k
dataset_split: train_sft
exact_match,flexible-extract: 0.77
exact_match,strict-match: 0.76
11 changes: 11 additions & 0 deletions tests/e2e/vLLM/lm_eval_configs/w4a16_actorder_weight.yaml
Original file line number Diff line number Diff line change
@@ -0,0 +1,11 @@
cadence: "weekly"
model: meta-llama/Meta-Llama-3-8B-Instruct
recipe: tests/e2e/vLLM/recipes/actorder/recipe_w4a16_actorder_weight.yaml
num_fewshot: 5
limit: 1000
task: "gsm8k"
dataset_id: HuggingFaceH4/ultrachat_200k
dataset_split: train_sft
exact_match,flexible-extract: 0.72
exact_match,strict-match: 0.72
scheme: W4A16_actorder_group
11 changes: 11 additions & 0 deletions tests/e2e/vLLM/lm_eval_configs/w4a16_grouped_quant.yaml
Original file line number Diff line number Diff line change
@@ -0,0 +1,11 @@
cadence: "weekly"
model: meta-llama/Meta-Llama-3-8B-Instruct
num_fewshot: 5
limit: 1000
task: "gsm8k"
exact_match,flexible-extract: 0.72
exact_match,strict-match: 0.72
scheme: W4A16
dataset_id: HuggingFaceH4/ultrachat_200k
dataset_split: train_sft
quant_type: "GPTQ"
Original file line number Diff line number Diff line change
@@ -0,0 +1,11 @@
quant_stage:
quant_modifiers:
SmoothQuantModifier:
smoothing_strength: 0.8
GPTQModifier:
ignore: [lm_head]
config_groups:
group_0:
weights: {num_bits: 8, type: int, symmetric: true, strategy: channel}
input_activations: {num_bits: 8, type: int, symmetric: true, strategy: token, dynamic: true}
targets: [Linear]
Original file line number Diff line number Diff line change
Expand Up @@ -2,7 +2,7 @@ quant_stage:
quant_modifiers:
SmoothQuantModifier:
smoothing_strength: 0.8
QuantizationModifier:
GPTQModifier:
ignore: [lm_head]
config_groups:
group_0:
Expand Down
4 changes: 2 additions & 2 deletions tests/e2e/vLLM/test_lmeval.py
Original file line number Diff line number Diff line change
Expand Up @@ -68,7 +68,7 @@ def set_up(self):
logger.info(self.scheme)

self.device = "cuda:0"
self.num_calibration_samples = 256
self.num_calibration_samples = 512
self.max_seq_length = 2048

def test_lm_eval(self):
Expand Down Expand Up @@ -104,7 +104,7 @@ def test_lm_eval(self):

logger.info("================= Running LM Eval ======================")

model_args = f"pretrained={self.save_dir}"
model_args = f"pretrained={self.save_dir},add_bos_token=True"
results = lm_eval.simple_evaluate(
model="hf",
model_args=model_args,
Expand Down

0 comments on commit 7c54bed

Please sign in to comment.