1
0
mirror of https://github.com/karpathy/minGPT synced 2024-11-15 19:10:39 +01:00
minGPT/projects/chargpt
2022-07-11 18:50:53 +00:00
..
chargpt.py refactor sequence generation into the model and match the huggingface/transformers api. touches everything but this makes a lot more sense to me aesthetically 2022-07-11 18:50:53 +00:00
readme.md attempt to make model config a little bit better, still hate it 2022-06-01 17:14:22 +00:00

chargpt

chargpt trains a character-level language model.

We support three settings: 1 convenience setting and 2 "benchmark" settings that have acedemic literature results:

  • a user specified input.txt file that we train an LM on (e.g. get tiny-shakespear (1.1MB of data) here)
  • TODO text8: also derived from Wikipedia text but all XML is removed and is lowercased to only 26 characters of
  • TODO enwik8 benchmark ("Hutter Prize"), first 100M bytes of a Wikipedia XML dump, with 205 unique tokensEnglish plus spaces