Back to List
Notice:This resource is provided by a third-party author. Please review the code with AI tools or manually before use to ensure security and compatibility.
PythonBlinkDL/RWKV-LM

RWKV-LM

RWKV (pronounced RwaKuv) is an RNN with great LLM performance, which can also be directly trained like a GPT transformer (parallelizable). We are at RWKV-7 "Goose". So it's combining the best of RNN and transformer - great performance, linear time, constant space (no kv-cache), fast training, infinite ctx_len, and free sentence embedding.

85.7/100
14.5KForks: 1.0K
View on GitHub
Loading report...

Similar Projects

TextRL

72

Implementation of ChatGPT RLHF (Reinforcement Learning with Human Feedback) on any generation model in huggingface's transformer (blommz-176B/bloom/gpt/bart/T5/MetaICL)

Python564

text-generation-inference

78

Large Language Model Text Generation Inference

Python10.8K

ChatRWKV

61

ChatRWKV is like ChatGPT but powered by RWKV (100% RNN) language model, and open source.

Python9.5K

LMFlow

83

An Extensible Toolkit for Finetuning and Inference of Large Foundation Models. Large Models for All.

Python8.5K
Back to List