We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
scaled_fp8_quant @HandH1998
scaled_fp8_quant
sglang/python/sglang/srt/layers/quantization/fp8.py
Line 10 in 2add697
Line 297 in 2add697
sglang/python/sglang/srt/layers/moe/ep_moe/layer.py
Line 6 in 2add697
Line 603 in 2add697
rotary_embedding @ByronHsu
rotary_embedding
sglang/python/sglang/srt/layers/rotary_embedding.py
Line 142 in 2add697
Lines 159 to 166 in 2add697
batched_rotary_embedding
topk_softmax @zhyncs
topk_softmax
sglang/python/sglang/srt/layers/moe/topk.py
Line 48 in 2add697
Lines 62 to 67 in 2add697
moe_align_block_size moe_sum @BBuf @zhyncs
moe_align_block_size
moe_sum
sglang/python/sglang/srt/layers/moe/fused_moe_triton/fused_moe.py
Line 14 in 2add697
Lines 448 to 455 in 2add697
Lines 1012 to 1015 in 2add697
awq_dequantize @bjmsong @zhyncs
awq_dequantize
sglang/python/sglang/srt/models/deepseek_v2.py
Line 25 in 2add697
Lines 946 to 952 in 2add697
No response
The text was updated successfully, but these errors were encountered:
We need to ensure that
Sorry, something went wrong.
bjmsong
ByronHsu
ispobock
BBuf
HandH1998
zhyncs
No branches or pull requests
Checklist
Motivation
scaled_fp8_quant
@HandH1998sglang/python/sglang/srt/layers/quantization/fp8.py
Line 10 in 2add697
sglang/python/sglang/srt/layers/quantization/fp8.py
Line 297 in 2add697
sglang/python/sglang/srt/layers/moe/ep_moe/layer.py
Line 6 in 2add697
sglang/python/sglang/srt/layers/moe/ep_moe/layer.py
Line 603 in 2add697
rotary_embedding
@ByronHsusglang/python/sglang/srt/layers/rotary_embedding.py
Line 142 in 2add697
sglang/python/sglang/srt/layers/rotary_embedding.py
Lines 159 to 166 in 2add697
BTW we don't need
batched_rotary_embedding
topk_softmax
@zhyncssglang/python/sglang/srt/layers/moe/topk.py
Line 48 in 2add697
sglang/python/sglang/srt/layers/moe/topk.py
Lines 62 to 67 in 2add697
moe_align_block_size
moe_sum
@BBuf @zhyncssglang/python/sglang/srt/layers/moe/fused_moe_triton/fused_moe.py
Line 14 in 2add697
sglang/python/sglang/srt/layers/moe/fused_moe_triton/fused_moe.py
Lines 448 to 455 in 2add697
sglang/python/sglang/srt/layers/moe/fused_moe_triton/fused_moe.py
Lines 1012 to 1015 in 2add697
awq_dequantize
@bjmsong @zhyncssglang/python/sglang/srt/models/deepseek_v2.py
Line 25 in 2add697
sglang/python/sglang/srt/models/deepseek_v2.py
Lines 946 to 952 in 2add697
Related resources
No response
The text was updated successfully, but these errors were encountered: