RNN+Transformer hybrid with O(n) inference. Linear time, infinite context, no KV cache. Train like GPT (parallel), infer like RNN (sequential). Linux Foundation AI project. Production at Windows, Office, NeMo. RWKV-7 (March 2025). Models up to 14B parameters.
7.0
Rating
0
Installs
AI & LLM
Category
Excellent skill documentation for RWKV architecture. The description clearly articulates the key value proposition (RNN+Transformer hybrid, O(n) inference, no KV cache) and production credibility. Task knowledge is comprehensive with practical code examples covering installation, both GPT and RNN modes, streaming generation, long context processing, fine-tuning, and direct performance comparisons. Structure is good with clear workflows, troubleshooting, and references to detailed architecture files. The skill addresses a genuinely novel architecture that solves real problems (memory efficiency for long contexts) that would be expensive for a CLI agent to handle repeatedly. Minor improvement areas: could benefit from more explicit decision trees for model selection and slightly more granular troubleshooting scenarios. Overall, this is a production-ready skill that would save significant tokens and provide specialized knowledge about an emerging architecture.
Loading SKILL.md…