Convert Figma logo to code with AI

karpathy logonanoGPT

The simplest, fastest repository for training/finetuning medium-sized GPTs.

36,124
5,629
36,124
265

Top Related Projects

22,214

Code for the paper "Language Models are Unsupervised Multitask Learners"

🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.

8,204

An implementation of model parallel GPT-2 and GPT-3-style models using the mesh-tensorflow library.

34,658

DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.

37,810

TensorFlow code and pre-trained models for BERT

30,129

Facebook AI Research Sequence-to-Sequence Toolkit written in Python.

Quick Overview

nanoGPT is a minimalistic implementation of GPT (Generative Pre-trained Transformer) in PyTorch. It's designed to be simple, educational, and easy to understand, making it an excellent resource for those looking to learn about transformer-based language models.

Pros

  • Lightweight and easy to understand implementation of GPT
  • Well-documented code with clear explanations
  • Suitable for educational purposes and experimentation
  • Can be run on consumer-grade hardware

Cons

  • Not optimized for large-scale production use
  • Limited features compared to more comprehensive libraries
  • May require additional work to adapt for specific tasks or datasets
  • Performance may not match state-of-the-art models

Code Examples

  1. Creating a GPT model:
from model import GPT

model = GPT(n_layer=12, n_head=12, n_embd=768, block_size=1024, vocab_size=50257)
  1. Generating text:
from model import GPT
import torch

model = GPT.from_pretrained('gpt2')
context = torch.zeros((1, 1), dtype=torch.long)
print(model.generate(context, max_new_tokens=500)[0].tolist())
  1. Training a model:
from model import GPT, GPTConfig
from trainer import Trainer

model = GPT(GPTConfig())
trainer = Trainer(model, train_dataset, test_dataset)
trainer.train()

Getting Started

To get started with nanoGPT:

  1. Clone the repository:

    git clone https://github.com/karpathy/nanoGPT.git
    cd nanoGPT
    
  2. Install dependencies:

    pip install -r requirements.txt
    
  3. Train a small model on Shakespeare dataset:

    python train.py config/train_shakespeare_char.py
    
  4. Sample from the trained model:

    python sample.py --out_dir=out-shakespeare-char
    

Competitor Comparisons

22,214

Code for the paper "Language Models are Unsupervised Multitask Learners"

Pros of GPT-2

  • More comprehensive and production-ready implementation
  • Includes pre-trained models and tokenizers
  • Supports a wider range of use cases and applications

Cons of GPT-2

  • More complex codebase, potentially harder to understand for beginners
  • Requires more computational resources for training and inference
  • Less focused on educational purposes compared to nanoGPT

Code Comparison

nanoGPT:

class GPTConfig:
    def __init__(self, vocab_size, n_layer, n_head, n_embd, block_size, bias, dropout):
        self.vocab_size = vocab_size
        self.n_layer = n_layer
        self.n_head = n_head
        self.n_embd = n_embd
        self.block_size = block_size
        self.bias = bias
        self.dropout = dropout

GPT-2:

class GPT2Config(PretrainedConfig):
    model_type = "gpt2"
    def __init__(
        self,
        vocab_size=50257,
        n_positions=1024,
        n_ctx=1024,
        n_embd=768,
        n_layer=12,
        n_head=12,
        ...
    ):

The code comparison shows that nanoGPT has a simpler configuration class, while GPT-2 includes more parameters and inherits from a pre-trained config class, reflecting its more comprehensive nature.

🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.

Pros of transformers

  • Extensive library with support for numerous architectures and tasks
  • Well-documented and actively maintained by a large community
  • Seamless integration with popular deep learning frameworks

Cons of transformers

  • Steeper learning curve due to its comprehensive nature
  • Potentially higher computational overhead for simple tasks
  • More complex setup and configuration for basic models

Code comparison

nanoGPT:

class GPTConfig:
    def __init__(self, vocab_size, n_layer, n_head, n_embd, block_size):
        self.vocab_size = vocab_size
        self.n_layer = n_layer
        self.n_head = n_head
        self.n_embd = n_embd
        self.block_size = block_size

transformers:

from transformers import GPT2Config

config = GPT2Config(
    vocab_size=50257,
    n_layer=12,
    n_head=12,
    n_embd=768,
    max_position_embeddings=1024
)

Summary

nanoGPT offers a minimalistic approach, ideal for learning and experimentation, while transformers provides a comprehensive toolkit for production-ready applications. nanoGPT is more lightweight and easier to understand, but transformers offers broader functionality and better integration with existing ecosystems.

8,204

An implementation of model parallel GPT-2 and GPT-3-style models using the mesh-tensorflow library.

Pros of gpt-neo

  • Larger scale and more powerful models available
  • Supports distributed training across multiple GPUs
  • More flexible architecture with support for different model sizes

Cons of gpt-neo

  • More complex codebase and harder to understand for beginners
  • Requires more computational resources to train and run
  • Less focused on educational purposes compared to nanoGPT

Code Comparison

nanoGPT:

class Block(nn.Module):
    def __init__(self, n_embd, n_head):
        super().__init__()
        self.ln_1 = nn.LayerNorm(n_embd)
        self.attn = CausalSelfAttention(n_embd, n_head)
        self.ln_2 = nn.LayerNorm(n_embd)
        self.mlp = MLP(n_embd)

gpt-neo:

class Block(nn.Module):
    def __init__(self, config):
        super().__init__()
        self.ln_1 = nn.LayerNorm(config.n_embd)
        self.attn = Attention(config)
        self.ln_2 = nn.LayerNorm(config.n_embd)
        self.mlp = MLP(config)

The code structures are similar, but gpt-neo uses a configuration object for initialization, allowing for more flexibility in model architecture. nanoGPT's implementation is more straightforward and easier to understand for educational purposes.

34,658

DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.

Pros of DeepSpeed

  • Highly optimized for large-scale distributed training
  • Supports a wide range of optimization techniques and hardware
  • Extensive documentation and active community support

Cons of DeepSpeed

  • Steeper learning curve due to complexity
  • May be overkill for smaller projects or single-GPU setups
  • Requires more setup and configuration

Code Comparison

nanoGPT:

class GPTConfig:
    def __init__(self, vocab_size, block_size, **kwargs):
        self.vocab_size = vocab_size
        self.block_size = block_size
        for k,v in kwargs.items():
            setattr(self, k, v)

DeepSpeed:

model_engine, optimizer, _, _ = deepspeed.initialize(
    args=args,
    model=model,
    model_parameters=model.parameters(),
    config=ds_config
)

nanoGPT is designed for simplicity and educational purposes, making it easier to understand the core concepts of GPT models. DeepSpeed, on the other hand, focuses on scalability and performance optimization for large-scale training. While nanoGPT is more suitable for learning and small projects, DeepSpeed is better suited for production-level, distributed training of large language models.

37,810

TensorFlow code and pre-trained models for BERT

Pros of BERT

  • Extensive pre-training on large datasets, making it highly effective for various NLP tasks
  • Well-documented and widely adopted in industry and research
  • Supports multiple languages and offers pre-trained models for various applications

Cons of BERT

  • More complex architecture, requiring more computational resources
  • Less flexible for customization and experimentation
  • Larger model size, which can be challenging for deployment in resource-constrained environments

Code Comparison

BERT (PyTorch):

from transformers import BertTokenizer, BertModel
tokenizer = BertTokenizer.from_pretrained('bert-base-uncased')
model = BertModel.from_pretrained('bert-base-uncased')

nanoGPT:

from model import GPTConfig, GPT
model_config = GPTConfig(n_layer=12, n_head=12, n_embd=768)
model = GPT(model_config)

nanoGPT offers a more lightweight and customizable approach, allowing users to easily modify the model architecture. BERT, on the other hand, provides a more standardized and pre-trained solution, which can be beneficial for quick implementation of NLP tasks without extensive training.

30,129

Facebook AI Research Sequence-to-Sequence Toolkit written in Python.

Pros of fairseq

  • Comprehensive toolkit with support for various NLP tasks and architectures
  • Extensive documentation and community support
  • Highly optimized for performance and scalability

Cons of fairseq

  • Steeper learning curve due to its complexity
  • Heavier resource requirements for setup and usage
  • Less suitable for educational purposes or quick prototyping

Code Comparison

fairseq:

from fairseq.models.transformer import TransformerModel
model = TransformerModel.from_pretrained('/path/to/model', checkpoint_file='model.pt')
tokens = model.encode('Hello world!')
output = model.decode(model.generate(tokens))

nanoGPT:

from model import GPTConfig, GPT
model = GPT(GPTConfig())
x = torch.tensor([[1, 2, 3]], dtype=torch.long)
logits, loss = model(x, targets=x)

Key Differences

  • fairseq offers a wide range of pre-built models and tasks, while nanoGPT focuses on a minimal GPT implementation
  • fairseq is designed for production-level research and deployment, whereas nanoGPT is ideal for learning and experimentation
  • fairseq requires more setup and configuration, while nanoGPT can be quickly integrated into projects with minimal dependencies

Convert Figma logo designs to code with AI

Visual Copilot

Introducing Visual Copilot: A new AI model to turn Figma designs to high quality code using your components.

Try Visual Copilot

README

nanoGPT

nanoGPT

The simplest, fastest repository for training/finetuning medium-sized GPTs. It is a rewrite of minGPT that prioritizes teeth over education. Still under active development, but currently the file train.py reproduces GPT-2 (124M) on OpenWebText, running on a single 8XA100 40GB node in about 4 days of training. The code itself is plain and readable: train.py is a ~300-line boilerplate training loop and model.py a ~300-line GPT model definition, which can optionally load the GPT-2 weights from OpenAI. That's it.

repro124m

Because the code is so simple, it is very easy to hack to your needs, train new models from scratch, or finetune pretrained checkpoints (e.g. biggest one currently available as a starting point would be the GPT-2 1.3B model from OpenAI).

install

pip install torch numpy transformers datasets tiktoken wandb tqdm

Dependencies:

  • pytorch <3
  • numpy <3
  • transformers for huggingface transformers <3 (to load GPT-2 checkpoints)
  • datasets for huggingface datasets <3 (if you want to download + preprocess OpenWebText)
  • tiktoken for OpenAI's fast BPE code <3
  • wandb for optional logging <3
  • tqdm for progress bars <3

quick start

If you are not a deep learning professional and you just want to feel the magic and get your feet wet, the fastest way to get started is to train a character-level GPT on the works of Shakespeare. First, we download it as a single (1MB) file and turn it from raw text into one large stream of integers:

python data/shakespeare_char/prepare.py

This creates a train.bin and val.bin in that data directory. Now it is time to train your GPT. The size of it very much depends on the computational resources of your system:

I have a GPU. Great, we can quickly train a baby GPT with the settings provided in the config/train_shakespeare_char.py config file:

python train.py config/train_shakespeare_char.py

If you peek inside it, you'll see that we're training a GPT with a context size of up to 256 characters, 384 feature channels, and it is a 6-layer Transformer with 6 heads in each layer. On one A100 GPU this training run takes about 3 minutes and the best validation loss is 1.4697. Based on the configuration, the model checkpoints are being written into the --out_dir directory out-shakespeare-char. So once the training finishes we can sample from the best model by pointing the sampling script at this directory:

python sample.py --out_dir=out-shakespeare-char

This generates a few samples, for example:

ANGELO:
And cowards it be strawn to my bed,
And thrust the gates of my threats,
Because he that ale away, and hang'd
An one with him.

DUKE VINCENTIO:
I thank your eyes against it.

DUKE VINCENTIO:
Then will answer him to save the malm:
And what have you tyrannous shall do this?

DUKE VINCENTIO:
If you have done evils of all disposition
To end his power, the day of thrust for a common men
That I leave, to fight with over-liking
Hasting in a roseman.

lol ¯\_(ツ)_/¯. Not bad for a character-level model after 3 minutes of training on a GPU. Better results are quite likely obtainable by instead finetuning a pretrained GPT-2 model on this dataset (see finetuning section later).

I only have a macbook (or other cheap computer). No worries, we can still train a GPT but we want to dial things down a notch. I recommend getting the bleeding edge PyTorch nightly (select it here when installing) as it is currently quite likely to make your code more efficient. But even without it, a simple train run could look as follows:

python train.py config/train_shakespeare_char.py --device=cpu --compile=False --eval_iters=20 --log_interval=1 --block_size=64 --batch_size=12 --n_layer=4 --n_head=4 --n_embd=128 --max_iters=2000 --lr_decay_iters=2000 --dropout=0.0

Here, since we are running on CPU instead of GPU we must set both --device=cpu and also turn off PyTorch 2.0 compile with --compile=False. Then when we evaluate we get a bit more noisy but faster estimate (--eval_iters=20, down from 200), our context size is only 64 characters instead of 256, and the batch size only 12 examples per iteration, not 64. We'll also use a much smaller Transformer (4 layers, 4 heads, 128 embedding size), and decrease the number of iterations to 2000 (and correspondingly usually decay the learning rate to around max_iters with --lr_decay_iters). Because our network is so small we also ease down on regularization (--dropout=0.0). This still runs in about ~3 minutes, but gets us a loss of only 1.88 and therefore also worse samples, but it's still good fun:

python sample.py --out_dir=out-shakespeare-char --device=cpu

Generates samples like this:

GLEORKEN VINGHARD III:
Whell's the couse, the came light gacks,
And the for mought you in Aut fries the not high shee
bot thou the sought bechive in that to doth groan you,
No relving thee post mose the wear

Not bad for ~3 minutes on a CPU, for a hint of the right character gestalt. If you're willing to wait longer, feel free to tune the hyperparameters, increase the size of the network, the context length (--block_size), the length of training, etc.

Finally, on Apple Silicon Macbooks and with a recent PyTorch version make sure to add --device=mps (short for "Metal Performance Shaders"); PyTorch then uses the on-chip GPU that can significantly accelerate training (2-3X) and allow you to use larger networks. See Issue 28 for more.

reproducing GPT-2

A more serious deep learning professional may be more interested in reproducing GPT-2 results. So here we go - we first tokenize the dataset, in this case the OpenWebText, an open reproduction of OpenAI's (private) WebText:

python data/openwebtext/prepare.py

This downloads and tokenizes the OpenWebText dataset. It will create a train.bin and val.bin which holds the GPT2 BPE token ids in one sequence, stored as raw uint16 bytes. Then we're ready to kick off training. To reproduce GPT-2 (124M) you'll want at least an 8X A100 40GB node and run:

torchrun --standalone --nproc_per_node=8 train.py config/train_gpt2.py

This will run for about 4 days using PyTorch Distributed Data Parallel (DDP) and go down to loss of ~2.85. Now, a GPT-2 model just evaluated on OWT gets a val loss of about 3.11, but if you finetune it it will come down to ~2.85 territory (due to an apparent domain gap), making the two models ~match.

If you're in a cluster environment and you are blessed with multiple GPU nodes you can make GPU go brrrr e.g. across 2 nodes like:

# Run on the first (master) node with example IP 123.456.123.456:
torchrun --nproc_per_node=8 --nnodes=2 --node_rank=0 --master_addr=123.456.123.456 --master_port=1234 train.py
# Run on the worker node:
torchrun --nproc_per_node=8 --nnodes=2 --node_rank=1 --master_addr=123.456.123.456 --master_port=1234 train.py

It is a good idea to benchmark your interconnect (e.g. iperf3). In particular, if you don't have Infiniband then also prepend NCCL_IB_DISABLE=1 to the above launches. Your multinode training will work, but most likely crawl. By default checkpoints are periodically written to the --out_dir. We can sample from the model by simply python sample.py.

Finally, to train on a single GPU simply run the python train.py script. Have a look at all of its args, the script tries to be very readable, hackable and transparent. You'll most likely want to tune a number of those variables depending on your needs.

baselines

OpenAI GPT-2 checkpoints allow us to get some baselines in place for openwebtext. We can get the numbers as follows:

$ python train.py config/eval_gpt2.py
$ python train.py config/eval_gpt2_medium.py
$ python train.py config/eval_gpt2_large.py
$ python train.py config/eval_gpt2_xl.py

and observe the following losses on train and val:

modelparamstrain lossval loss
gpt2124M3.113.12
gpt2-medium350M2.852.84
gpt2-large774M2.662.67
gpt2-xl1558M2.562.54

However, we have to note that GPT-2 was trained on (closed, never released) WebText, while OpenWebText is just a best-effort open reproduction of this dataset. This means there is a dataset domain gap. Indeed, taking the GPT-2 (124M) checkpoint and finetuning on OWT directly for a while reaches loss down to ~2.85. This then becomes the more appropriate baseline w.r.t. reproduction.

finetuning

Finetuning is no different than training, we just make sure to initialize from a pretrained model and train with a smaller learning rate. For an example of how to finetune a GPT on new text go to data/shakespeare and run prepare.py to download the tiny shakespeare dataset and render it into a train.bin and val.bin, using the OpenAI BPE tokenizer from GPT-2. Unlike OpenWebText this will run in seconds. Finetuning can take very little time, e.g. on a single GPU just a few minutes. Run an example finetuning like:

python train.py config/finetune_shakespeare.py

This will load the config parameter overrides in config/finetune_shakespeare.py (I didn't tune them much though). Basically, we initialize from a GPT2 checkpoint with init_from and train as normal, except shorter and with a small learning rate. If you're running out of memory try decreasing the model size (they are {'gpt2', 'gpt2-medium', 'gpt2-large', 'gpt2-xl'}) or possibly decreasing the block_size (context length). The best checkpoint (lowest validation loss) will be in the out_dir directory, e.g. in out-shakespeare by default, per the config file. You can then run the code in sample.py --out_dir=out-shakespeare:

THEODORE:
Thou shalt sell me to the highest bidder: if I die,
I sell thee to the first; if I go mad,
I sell thee to the second; if I
lie, I sell thee to the third; if I slay,
I sell thee to the fourth: so buy or sell,
I tell thee again, thou shalt not sell my
possession.

JULIET:
And if thou steal, thou shalt not sell thyself.

THEODORE:
I do not steal; I sell the stolen goods.

THEODORE:
Thou know'st not what thou sell'st; thou, a woman,
Thou art ever a victim, a thing of no worth:
Thou hast no right, no right, but to be sold.

Whoa there, GPT, entering some dark place over there. I didn't really tune the hyperparameters in the config too much, feel free to try!

sampling / inference

Use the script sample.py to sample either from pre-trained GPT-2 models released by OpenAI, or from a model you trained yourself. For example, here is a way to sample from the largest available gpt2-xl model:

python sample.py \
    --init_from=gpt2-xl \
    --start="What is the answer to life, the universe, and everything?" \
    --num_samples=5 --max_new_tokens=100

If you'd like to sample from a model you trained, use the --out_dir to point the code appropriately. You can also prompt the model with some text from a file, e.g. python sample.py --start=FILE:prompt.txt.

efficiency notes

For simple model benchmarking and profiling, bench.py might be useful. It's identical to what happens in the meat of the training loop of train.py, but omits much of the other complexities.

Note that the code by default uses PyTorch 2.0. At the time of writing (Dec 29, 2022) this makes torch.compile() available in the nightly release. The improvement from the one line of code is noticeable, e.g. cutting down iteration time from ~250ms / iter to 135ms / iter. Nice work PyTorch team!

todos

  • Investigate and add FSDP instead of DDP
  • Eval zero-shot perplexities on standard evals (e.g. LAMBADA? HELM? etc.)
  • Finetune the finetuning script, I think the hyperparams are not great
  • Schedule for linear batch size increase during training
  • Incorporate other embeddings (rotary, alibi)
  • Separate out the optim buffers from model params in checkpoints I think
  • Additional logging around network health (e.g. gradient clip events, magnitudes)
  • Few more investigations around better init etc.

troubleshooting

Note that by default this repo uses PyTorch 2.0 (i.e. torch.compile). This is fairly new and experimental, and not yet available on all platforms (e.g. Windows). If you're running into related error messages try to disable this by adding --compile=False flag. This will slow down the code but at least it will run.

For some context on this repository, GPT, and language modeling it might be helpful to watch my Zero To Hero series. Specifically, the GPT video is popular if you have some prior language modeling context.

For more questions/discussions feel free to stop by #nanoGPT on Discord:

acknowledgements

All nanoGPT experiments are powered by GPUs on Lambda labs, my favorite Cloud GPU provider. Thank you Lambda labs for sponsoring nanoGPT!