Back to List
Notice:This resource is provided by a third-party author. Please review the code with AI tools or manually before use to ensure security and compatibility.
Pythonvoidful/TextRL

TextRL

Implementation of ChatGPT RLHF (Reinforcement Learning with Human Feedback) on any generation model in huggingface's transformer (blommz-176B/bloom/gpt/bart/T5/MetaICL)

72.0/100
564Forks: 61
View on GitHub
Loading report...

Similar Projects

RWKV-LM

86

RWKV (pronounced RwaKuv) is an RNN with great LLM performance, which can also be directly trained like a GPT transformer (parallelizable). We are at RWKV-7 "Goose". So it's combining the best of RNN and transformer - great performance, linear time, constant space (no kv-cache), fast training, infinite ctx_len, and free sentence embedding.

Python14.5K

ChatRWKV

61

ChatRWKV is like ChatGPT but powered by RWKV (100% RNN) language model, and open source.

Python9.5K

LMFlow

83

An Extensible Toolkit for Finetuning and Inference of Large Foundation Models. Large Models for All.

Python8.5K

transformers

98

🤗 Transformers: the model-definition framework for state-of-the-art machine learning models in text, vision, audio, and multimodal models, for both inference and training.

Python159.8K
Back to List