- Stanford
- https://twitter.com/karpathy
Highlights
- Pro
Stars
Fast and accurate automatic speech recognition (ASR) for edge devices
Minimal Claude Code alternative. Single Python file, zero dependencies, ~250 lines.
Our library for RL environments + evals
FlexAttention based, minimal vllm-style inference engine for fast Gemma 2 inference.
gpt-oss-120b and gpt-oss-20b are two open-weight language models by OpenAI
Wan: Open and Advanced Large-Scale Video Generative Models
Simple MPI implementation for prototyping or learning
The simplest, fastest repository for training/finetuning small-sized VLMs.
Minimalistic 4D-parallelism distributed training framework for education purpose
Implementing DeepSeek R1's GRPO algorithm from scratch
Frontier Models playing the board game Diplomacy.
Minimal reproduction of DeepSeek R1-Zero
Official repository for our work on micro-budget training of large-scale diffusion models.
A PyTorch native platform for training generative AI models
Efficient Triton Kernels for LLM Training
MLX native implementations of state-of-the-art generative image models
Official inference repo for FLUX.1 models
the scott CPU from "But How Do It Know?" by J. Clark Scott
Run PyTorch LLMs locally on servers, desktop and mobile
A lightweight library for portable low-level GPU computation using WebGPU.
Simple Byte pair Encoding mechanism used for tokenization process . written purely in C
[ICLR 2025] Samba: Simple Hybrid State Space Models for Efficient Unlimited Context Language Modeling
User-friendly AI Interface (Supports Ollama, OpenAI API, ...)






