+
Skip to content

fla-org/fla-zoo

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

🦁 FLA-Zoo: FLA models beyond language

GitHub stars GitHub forks GitHub issues

A collection of FLA models extending beyond language

Supporting vision, video, and more with efficient kernels mainly from fla-org

diagram

News 📰Features ✨Installation 🔧Getting Started 🚀

News

  • [2025-06-04] LaCT repo and STA repo is included.

  • [2025-04-23] A dedicated part of this repo: flazoo/linearized_models is created to store the linearized versions of your favorite transformers.

  • [2025-04-21] A dedicated part of this repo: flazoo/helpers is created to provide some common utils.

  • [2025-04-03] MoBA is included as part of the collection for sparse attention. You can use it in specific layers of FLA models or directly use its full-blown models. Use hidden size which is multiple of 32 for MoBA.

  • [2025-03-16] Native Sparse Attention (NSA) for vision is now added. See the triton implementation under the hood here and its visual variant here.

  • [2025-03-02] A pilot version of Native Sparse Attention (NSA) is added. More experiments should be conducted to test its performance.

  • [2025-02-23] Add LightNet for classification. Also, a pilot SFT training script for vision models is added, check it out in here.

  • [2025-02-20] Experiments evaluating the performance of vision models are in progress.

  • [2025-01-25] This repo is created with some vision encoders.

Features

Understanding Models

Please refer to the documentation 📖.

Installation

Requirements

Quick Install

For people who just want to use base linear or hybrid model, basical dependencies below are enough.

# Create and activate conda environment
conda create -n flazoo python=3.12
conda activate flazoo

pip install torch==2.6.0 torchvision==0.21.0 accelerate diffusers timm --use-pep517

pip install transformers datasets evaluate causal_conv1d einops scikit-learn wandb matplotlib deepspeed

# Install flash-attention, this is required if you like hybrid models
pip install flash-attn --no-build-isolation --use-pep517

# A handy tool to monitor GPU
pip install nvitop

pip install -U "huggingface_hub[cli]" --use-pep517
pip install pillow==11.1.0 --use-pep517
pip install git+https://github.com/facebookresearch/pytorchvideo.git

# Install FLA-Zoo in development mode
pip install -e .

Some repos are needed if you want to go deeper. Below is a basic example illustrating how to install MoBA.

# clone and install MoBA
git clone git+https://github.com/MoonshotAI/MoBA.git

Below is a table of these repos and what they are used for in fla-zoo.

Repo Link Used for
MoBA link Sparse hybrid
REPA link Gen2D training

💡 Note: As an actively developed repository, no released packages of fla-zoo are currently provided. Use pip install -e . to install the package in development mode.

Now we can start cooking! 🚀

Getting Started

Basic Usage

from flazoo import DeltaNetForImageClassification

from flazoo.helpers.informer import log_model_parameters_flat

model = DeltaNetForImageClassification.from_pretrained("fla-zoo/deltanet-siglip2-base-patch16-224")

log_model_parameters_flat(model, "delta-siglip2.log")

About

Flash-Linear-Attention models beyond language

Topics

Resources

License

Stars

Watchers

Forks

Contributors 2

  •  
  •  

Languages

点击 这是indexloc提供的php浏览器服务,不要输入任何密码和下载