Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We鈥檒l occasionally send you account related emails.

Already on GitHub? Sign in to your account

[Bug]: why the logits is different between 0.4.1 and 0.4.2 #4740

Open
sitabulaixizawaluduo opened this issue May 10, 2024 · 1 comment
Open
Labels
bug Something isn't working

Comments

@sitabulaixizawaluduo
Copy link

Your current environment

The output of `python collect_env.py`

馃悰 Describe the bug

from vllm import LLM, SamplingParams
sampling_params = SamplingParams(temperature=0, max_tokens=2048)
llm = LLM(model="Llama-3-8B",tensor_parallel_size=4, trust_remote_code=True)
outputs = llm.generate(prompts=prompts, sampling_params=sampling_params)
for output in outputs:
    prompt = output.prompt
    generated_text = output.outputs[0].text
    print(prompt + generated_text)         

I use vllm 0.4.1 and 0.4.2 run this code, and I print logits in sampler.py, but I found the logits is different between 0.4.1 and 0.4.2 when the tensor_parallel_size is 4, but is the same when tensor_parallel_size is 2 or 1.
TP4 first token logits:
0.4.2
image
0.4.1
image
TP2 first token logits:
0.4.2
image
0.4.1
image

@sitabulaixizawaluduo sitabulaixizawaluduo added the bug Something isn't working label May 10, 2024
@sitabulaixizawaluduo
Copy link
Author

I useFlashAttention backend with flash-attn==2.5.2

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
bug Something isn't working
Projects
None yet
Development

No branches or pull requests

1 participant