Skip to content
/ ak-llm Public

Simple transformer-based LLM based on Andrej Karpathy's GPT tutorial.

Notifications You must be signed in to change notification settings

ivfiev/ak-llm

Repository files navigation

The tutorial

Requires an appropriate build of PyTorch. Tested using rocm/pytorch

Training: python run_model.py --train --blocks 6 --context 128 --dimensions 256 --iterations 2500 --filename model_file_name

Running: python run_model.py --run --blocks 6 --context 128 --dimensions 256 --output 300 --filename model_file_name

(Optional) Training the BPE tokenizer: python run_tokenizer.py -v 512 -i input.txt -o tok_512

(Optional) Using the BPE tokenizer: python run_model.py ... --tokenizer tok_512

About

Simple transformer-based LLM based on Andrej Karpathy's GPT tutorial.

Topics

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages