Hugging Face
Models
Datasets
Spaces
Buckets
new
Docs
Enterprise
Pricing
Log In
Sign Up
10
6
406
Aptronym
Aptronym
Follow
agentlans's profile picture
molbal's profile picture
Benson's profile picture
9 followers
·
93 following
aptronymist
AI & ML interests
Emergent Gaming
Recent Activity
new
activity
1 day ago
ManniX-ITA/gemma-4-A4B-98e-v3-it:
Scripts
liked
a dataset
1 day ago
ConicCat/MiniC2_V3.2
reacted
to
eaddario
's
post
with 👍
1 day ago
Experimental global target bits‑per‑weight quantization of Qwen/Qwen3.5-4B and Qwen/Qwen3.5-9B Unlike standard llama.cpp quantizations that rely on fixed type heuristics (e.g., Q4_K_M), the Target BPW approach optimizes per-tensor precision where it matters the most, and produces high quality models that meet a precise global file size target. Key Advantages: - VRAM Maximization: Can generate high quality models sized exactly to fit hardware constraints (e.g., fitting the model into exactly 24GB VRAM). - Data-Driven Precision: Quantization mix is determined by actual weight error sensitivity rather than hardcoded rules, often yielding better PPL/KLD size trade-offs. Full benchmarks (PPL, KLD, ARC, MMLU, etc.) and methodology in the models' cards https://huggingface.co/eaddario/Qwen3.5-4B-GGUF https://huggingface.co/eaddario/Qwen3.5-9B-GGUF
View all activity
Organizations
None yet
Aptronym
's models
6
Sort: Recently updated
Aptronym/Nanbeige4-3B-Base-heretic-1
Text Generation
•
4B
•
Updated
Jan 8
•
4
Aptronym/Phi-3.5-mini-3.8B-ArliAI-RPMax-v1.1-Q8_0-GGUF
4B
•
Updated
Sep 13, 2024
•
5
Aptronym/Phi-3.5-mini-3.8B-ArliAI-RPMax-v1.1-Q5_K_M-GGUF
4B
•
Updated
Sep 13, 2024
•
43
Aptronym/LowStepLoras
Updated
Jul 9, 2024
Aptronym/SDNext
Updated
May 29, 2024
Aptronym/SDN-Upscalers
Updated
Sep 13, 2023