Papers
Sign in to view your remaining parses.
Tag Filter
RWKV Architecture
ModRWKV: Transformer Multimodality in Linear Time
Published:11/1/2025
ModRWKV Multimodal FrameworkRWKV ArchitectureLinear-Time TransformerMultimodal Large Language ModelsDynamically Adaptable Heterogeneous Modality Encoders
This study introduces ModRWKV, a framework based on RWKV architecture that achieves multimodal processing with linear time complexity, outperforming traditional quadraticcomplexity Transformer models. It balances performance and computational efficiency for multisource informat
012
RWKV-7 "Goose" with Expressive Dynamic State Evolution
Published:3/19/2025
RWKV ArchitectureLanguage Model Performance EvaluationMultilingual TasksOpen Source Pre-trained DatasetState Tracking and Language Recognition
RWKV7 "Goose" is a novel sequence modeling architecture that achieves constant memory usage and inference time. This 2.9 billion parameter model sets new stateoftheart performance on multilingual tasks and matches existing benchmarks in English, while introducing generalized
011