File tree Expand file tree Collapse file tree 5 files changed +42
-9
lines changed
tests/integration/defs/accuracy Expand file tree Collapse file tree 5 files changed +42
-9
lines changed Original file line number Diff line number Diff line change @@ -45,8 +45,10 @@ microsoft/Phi-3.5-mini-instruct:
4545 - accuracy : 31.354
4646microsoft/Phi-4-mini-instruct :
4747 - accuracy : 32.921
48- codestral/codestral-22b-v0.1 :
49- - accuracy : 28.90
48+ mistralai/Codestral-22B-v0.1 :
49+ - accuracy : 30.316
50+ - quant_algo : FP8
51+ accuracy : 30.316
5052state-spaces/mamba-130m-hf :
5153 - accuracy : 19.470
5254lmsys/vicuna-7b-v1.3 :
Original file line number Diff line number Diff line change @@ -158,8 +158,8 @@ microsoft/Phi-4-multimodal-instruct-long-rope:
158158 - accuracy : 75.85
159159microsoft/Phi-4-mini-instruct :
160160 - accuracy : 82.30
161- codestral/codestral-22b -v0.1 :
162- - accuracy : 28.90
161+ mistralai/Codestral-22B -v0.1 :
162+ - accuracy : 67.10
163163GPT-OSS/BF16 :
164164 - accuracy : 90.3
165165GPT-OSS/MXFP4 :
Original file line number Diff line number Diff line change @@ -229,8 +229,10 @@ nvidia/Nemotron-H-56B-Base-8K:
229229 accuracy : 83.82
230230microsoft/Phi-4-mini-instruct :
231231 - accuracy : 68.98
232- codestral/codestral-22b-v0.1 :
233- - accuracy : 28.90
232+ mistralai/Codestral-22B-v0.1 :
233+ - accuracy : 61.72
234+ - quant_algo : FP8
235+ accuracy : 61.72
234236# Created a dummy accuracy to track tp_size=2 for phi4-mini model.
235237# TODO: update once https://nvbugs/5393849 is fixed.
236238microsoft/Phi-4-mini-instruct-tp2 :
Original file line number Diff line number Diff line change @@ -433,3 +433,29 @@ def test_auto_dtype(self):
433433 speculative_config = self .speculative_config ) as llm :
434434 task = CnnDailymail (self .MODEL_NAME )
435435 task .evaluate (llm )
436+
437+
438+ class TestCodestral_22B_V01 (LlmapiAccuracyTestHarness ):
439+ MODEL_NAME = "mistralai/Codestral-22B-v0.1"
440+ MODEL_PATH = f"{ llm_models_root ()} /Codestral-22B-v0.1"
441+ kv_cache_config = KvCacheConfig (free_gpu_memory_fraction = 0.6 )
442+
443+ @pytest .mark .skip_less_device_memory (80000 )
444+ def test_auto_dtype (self ):
445+ with LLM (self .MODEL_PATH , kv_cache_config = self .kv_cache_config ) as llm :
446+ task = CnnDailymail (self .MODEL_NAME )
447+ task .evaluate (llm )
448+ task = MMLU (self .MODEL_NAME )
449+ task .evaluate (llm )
450+
451+ @skip_pre_ada
452+ @pytest .mark .skip_less_device_memory (80000 )
453+ def test_fp8 (self ):
454+ quant_config = QuantConfig (QuantAlgo .FP8 )
455+ with LLM (self .MODEL_PATH ,
456+ quant_config = quant_config ,
457+ kv_cache_config = self .kv_cache_config ) as llm :
458+ task = CnnDailymail (self .MODEL_NAME )
459+ task .evaluate (llm )
460+ task = MMLU (self .MODEL_NAME )
461+ task .evaluate (llm )
Original file line number Diff line number Diff line change @@ -2437,11 +2437,14 @@ def test_auto_dtype(self):
24372437
24382438
24392439class TestCodestral_22B_V01 (LlmapiAccuracyTestHarness ):
2440- MODEL_NAME = "codestral/codestral-22b -v0.1"
2441- MODEL_PATH = f"{ llm_models_root ()} /codestral-22b -v0.1"
2440+ MODEL_NAME = "mistralai/Codestral-22B -v0.1"
2441+ MODEL_PATH = f"{ llm_models_root ()} /Codestral-22B -v0.1"
24422442
2443+ @pytest .mark .timeout (2400 )
2444+ @pytest .mark .skip_less_device_memory (80000 )
24432445 def test_auto_dtype (self ):
2444- with LLM (self .MODEL_PATH ) as llm :
2446+ kv_cache_config = KvCacheConfig (free_gpu_memory_fraction = 0.6 )
2447+ with LLM (self .MODEL_PATH , kv_cache_config = kv_cache_config ) as llm :
24452448 task = CnnDailymail (self .MODEL_NAME )
24462449 task .evaluate (llm )
24472450 task = MMLU (self .MODEL_NAME )
You can’t perform that action at this time.
0 commit comments