Llama2 inference in one TypeScript file
-
Updated
May 23, 2024 - JavaScript
Llama2 inference in one TypeScript file
Codebase for arXiv:2405.17767, based on GPT-Neo and TinyStories.
Dataset Generation Code for SimpleStories
Train and run a small Llama 2 model from scratch on the TinyStories dataset.
Train GPT style model on tinystories dataset
Small transformer trained from scratch
Code implementation for our paper "BERTtime Stories: Investigating the Role of Synthetic Story Data in Language Pre-training" as part of the 2024 BabyLM Challenge
This project fine-tunes GPT-2, a popular pre-trained transformer model, to generate short stories using the TinyStories dataset. The goal is to teach GPT-2 to produce creative and coherent stories based on prompts.
A PyTorch implementation of a Bigram Language Model using Transformer architecture for character-level text generation.
Add a description, image, and links to the tinystories topic page so that developers can more easily learn about it.
To associate your repository with the tinystories topic, visit your repo's landing page and select "manage topics."