LM Human Preferences
lm-human-preferences is the official
OpenAI codebase that implements the
method from the paper Fine-Tuning
Language Models from Human Preferences.
Its purpose ...
Enter
Grade School Math
The grade-school-math repository
(sometimes called GSM8K) is a curated
dataset of 8,500 high-quality grade
school math word problems intended for
evaluating ma...
Enter
Universe Starter Agent
The universe-starter-agent repository
is an archived OpenAI codebase designed
as a starter reinforcement-learning
agent that can interact with and solve
tasks ...
Enter
Supervised Reptile
The supervised-reptile repository
contains code associated with the paper
On First-Order Meta-Learning
Algorithms, which introduces Reptile, a
meta-learning ...
Enter
FlashMLA
FlashMLA is a high-performance decoding
kernel library designed especially for
Multi-Head Latent Attention (MLA)
workloads, targeting NVIDIA Hopper GPU
archite...
Enter
DeepEP
DeepEP is a communication library
designed specifically to support
Mixture-of-Experts (MoE) and expert
parallelism (EP) deployments. Its core
role is to implem...
Enter
sparklyr
sparklyr is an R package that provides
seamless interfacing with Apache Spark
clusterseither local or remotewhile
letting users write code in familiar R
para...
Enter
Open Infra Index
open-infra-index is a central
infrastructure index repository
maintained by DeepSeek AI that acts as a
catalog and hub for a collection of
production-tested ...
Enter
DeepSeek LLM
The DeepSeek-LLM repository hosts the
code, model files, evaluations, and
documentation for DeepSeeks LLM series
(notably the 67B Chat variant). Its
tagline i...
Enter