Chris Scott
chriswritescode
AI & ML interests
None yet
Recent Activity
new activity about 19 hours ago
togethercomputer/Aurora-Spec-Minimax-M2.5:awaiting access liked a model 2 days ago
prithivMLmods/QIE-2509-Object-Remover-Bbox-v3 liked a model 4 days ago
Sehyo/Qwen3.5-397B-A17B-NVFP4Organizations
None yet
awaiting access
#1 opened about 19 hours ago
by
chriswritescode
GPTQ vs Q4 GGUF
π 3
1
#2 opened 14 days ago
by
ciprianv
Minimax M2.5 ?
4
#1 opened 26 days ago
by
chriswritescode
Official FP8
ππ 14
4
#4 opened 22 days ago
by
retowyss
Anyone try this on 4x RTX 6000 Pro yet?
52
#1 opened 29 days ago
by
zenmagnets
Accessing LLM, response without<think>start tag
5
#2 opened 3 months ago
by
sudage
best coding&reasoning model. thank you Z.AI
π 6
4
#10 opened 3 months ago
by
Tugay31
did you use the default nemotron dataset ?
#1 opened 3 months ago
by
chriswritescode
Great Model! - sglang mtp support for triton backend
π 3
4
#19 opened 3 months ago
by
chriswritescode
vLLM load error
π 3
11
#2 opened 3 months ago
by
srinivasbilla
Thank you!!
π€π€ 2
1
#1 opened 5 months ago
by
rascazzione
Appreciation
β€οΈ 1
#3 opened 5 months ago
by
chriswritescode
Quant Size
8
#2 opened 6 months ago
by
ortegaalfredo
Thank you!!!
1
#1 opened 6 months ago
by
chriswritescode
NVfp4 request with 16bit activations
1
#5 opened 6 months ago
by
chriswritescode
model is not performing as good as GLM-4.5-Air-AWQ-FP16Mix
3
#1 opened 7 months ago
by
hareram241
Request: 4-bit GPTQ or AWQ quantized version of openai/gpt-oss-20b
πβ€οΈ 18
13
#32 opened 7 months ago
by
powtac
VLLM - Flash-attn 3
12
#23 opened 7 months ago
by
chriswritescode
Please make one for the larger Non Air Variant
3
#2 opened 8 months ago
by
chriswritescode
AWQ 4Bit / GPTQ with full precision gates and head? Please
8
#4 opened 8 months ago
by
chriswritescode