-
Bytedance. Ltd
- Hang Zhou, Zhejiang Province, VR China
Stars
A simple HTML5, YouTube and Vimeo player
The paper list of the 86-page paper "The Rise and Potential of Large Language Model Based Agents: A Survey" by Zhiheng Xi et al.
[Development suspended] Advanced open-source Texas Hold'em GTO solver with optimized performance (web browser version)
Train transformer language models with reinforcement learning.
Secrets of RLHF in Large Language Models Part I: PPO
A curated list of reinforcement learning with human feedback resources (continually updated)
A guidance language for controlling large language models.
Exact structure out of any language model completion.
AutoGPT is the vision of accessible AI for everyone, to use and to build on. Our mission is to provide the tools, so that you can focus on what matters.
Instruction Tuning with GPT-4
🦜🔗 Build context-aware reasoning applications
Code for our EMNLP 2023 Paper: "LLM-Adapters: An Adapter Family for Parameter-Efficient Fine-Tuning of Large Language Models"
Instruct-tune LLaMA on consumer hardware
An open platform for training, serving, and evaluating large language models. Release repo for Vicuna and Chatbot Arena.
Making large AI models cheaper, faster and more accessible
Implementation of the LLaMA language model based on nanoGPT. Supports flash attention, Int8 and GPTQ 4bit quantization, LoRA and LLaMA-Adapter fine-tuning, pre-training. Apache 2.0-licensed.
zphang / transformers
Forked from huggingface/transformersCode and models for BERT on STILTs
BELLE: Be Everyone's Large Language model Engine(开源中文对话大模型)
Code and documentation to train Stanford's Alpaca models, and generate the data.
deepspeedai / Megatron-DeepSpeed
Forked from NVIDIA/Megatron-LMOngoing research training transformer language models at scale, including: BERT & GPT-2
A collection of libraries to optimise AI model performances
Implementation of RLHF (Reinforcement Learning with Human Feedback) on top of the PaLM architecture. Basically ChatGPT but with PaLM
Transformer related optimization, including BERT, GPT
Training and serving large-scale neural networks with auto parallelization.