Gonzales
Fernanda24
AI & ML interests
None yet
Organizations
None yet
FP8 versions of DeepSeek-V3.2 would awesome!
2
#1 opened 2 months ago
by
Fernanda24
ValueError: gemm_fp8_nt_groupwise is only supported on SM100, SM103 in trtllm backend.
#3 opened 2 months ago
by
Fernanda24
Is it possible to make smaller NVFP4 quant at 340-360GB to fit in 4x96gb?
๐ 1
68
#1 opened 3 months ago
by
Fernanda24
attention backend
1
#1 opened 3 months ago
by
Fernanda24
Aww Man!
20
#1 opened 3 months ago
by
mtcl
Question will it work in vllm or sglang with rtx 6000 blackwells? cuda arch sm120
6
#1 opened 5 months ago
by
Fernanda24
ooof this fits in 4x96gb can we get this for the new 3.2 Speciale ase well please :)
16
#2 opened 3 months ago
by
Fernanda24
4 x RTX PRO 6000
๐ 1
2
#1 opened 3 months ago
by
willfalco
How did you bypass deepseek-v32 not recognized in Tranformers?
3
#3 opened 3 months ago
by
Fernanda24
model type `deepseek_v32` not found
5
#16 opened 3 months ago
by
harryhan618
Update README.md
1
#4 opened 3 months ago
by
Fernanda24
broken link to mistral small
1
#3 opened 3 months ago
by
Fernanda24
Aww Man!
20
#1 opened 3 months ago
by
mtcl
Is it possible to make smaller NVFP4 quant at 340-360GB to fit in 4x96gb?
๐ 1
68
#1 opened 3 months ago
by
Fernanda24
Cool!
#6 opened 3 months ago
by
Fernanda24
Thanks!
#1 opened 3 months ago
by
Fernanda24
ooof this fits in 4x96gb can we get this for the new 3.2 Speciale ase well please :)
16
#2 opened 3 months ago
by
Fernanda24
can we get this for the new Deepseek v3.2?
๐ 1
#2 opened 3 months ago
by
Fernanda24
Question will it work in vllm or sglang with rtx 6000 blackwells? cuda arch sm120
6
#1 opened 5 months ago
by
Fernanda24
you know which nightly it worked with? because it does not with current one
31
#1 opened 4 months ago
by
willfalco