Running 3.62k The Ultra-Scale Playbook 🌌 3.62k The ultimate guide to training LLM on large GPU Clusters
Running on CPU Upgrade Featured 2.76k The Smol Training Playbook 📚 2.76k The secrets to building world-class LLMs
ShinkaEvolve: Towards Open-Ended And Sample-Efficient Program Evolution Paper • 2509.19349 • Published Sep 17, 2025 • 2
GLM-4.5: Agentic, Reasoning, and Coding (ARC) Foundation Models Paper • 2508.06471 • Published Aug 8, 2025 • 195
Seed-Prover: Deep and Broad Reasoning for Automated Theorem Proving Paper • 2507.23726 • Published Jul 31, 2025 • 114
DepNeCT Collection This Hugging Face collection hosts models and datasets from DepNeCT — a dependency-based method for nested compound type identification in Sanskrit • 4 items • Updated Jul 29, 2025 • 2
nvidia/OpenReasoning-Nemotron-32B Text Generation • 33B • Updated Sep 16, 2025 • 311 • • 121
REFINE-AF: A Task-Agnostic Framework to Align Language Models via Self-Generated Instructions using Reinforcement Learning from Automated Feedback Paper • 2505.06548 • Published May 10, 2025 • 30
REFINE-AF: A Task-Agnostic Framework to Align Language Models via Self-Generated Instructions using Reinforcement Learning from Automated Feedback Paper • 2505.06548 • Published May 10, 2025 • 30
google/gemma-3-4b-it-qat-int4-unquantized Image-Text-to-Text • 4B • Updated Apr 15, 2025 • 535 • 9
Running on CPU Upgrade Featured 993 Model Memory Utility 🚀 993 Calculate vRAM needed for model training and inference
view article Article Making LLMs even more accessible with bitsandbytes, 4-bit quantization and QLoRA +3 May 24, 2023 • 171