Skip to content

Commit

Permalink
add fp16 kv_cache test
Browse files Browse the repository at this point in the history
  • Loading branch information
noooop committed Aug 30, 2024
1 parent 01d5dd6 commit b601bb6
Showing 1 changed file with 2 additions and 2 deletions.
4 changes: 2 additions & 2 deletions tests/basic_correctness/test_chunked_prefill.py
Original file line number Diff line number Diff line change
Expand Up @@ -79,8 +79,8 @@ def test_models(

@pytest.mark.parametrize("kv_cache_dtype,model",
[("fp8_e5m2", m)
for m in E5M2_KV_MODELS] + [("fp8_e4m3", m)
for m in E4M3_KV_MODELS]
for m in E5M2_KV_MODELS]
+ [("fp8_e4m3", m) for m in E4M3_KV_MODELS]
+ [("auto", "Qwen/Qwen2-1.5B-Instruct")])
# Due to low-precision numerical divergence, we only test logprob of 4 tokens
@pytest.mark.parametrize("max_tokens", [4])
Expand Down

0 comments on commit b601bb6

Please sign in to comment.