Architecture
State-Space Model (Mamba)
A sequence model architecture that processes tokens through a compressed hidden state, offering linear-time scaling as an alternative to Transformer attention.
Architecture
A sequence model architecture that processes tokens through a compressed hidden state, offering linear-time scaling as an alternative to Transformer attention.
We use cookies
Anonymous analytics help us improve the site. You can opt out anytime. Learn more