# -*- coding: utf-8 -*- from fla.layers import ( ABCAttention, Attention, BasedLinearAttention, BitAttention, DeltaNet, GatedDeltaNet, GatedDeltaProduct, GatedLinearAttention, GatedSlotAttention, HGRN2Attention, HGRNAttention, LightNetAttention, LinearAttention, MultiScaleRetention, NativeSparseAttention, ReBasedLinearAttention, RWKV6Attention, RWKV7Attention ) from fla.models import ( ABCForCausalLM, ABCModel, BitNetForCausalLM, BitNetModel, DeltaNetForCausalLM, DeltaNetModel, GatedDeltaNetForCausalLM, GatedDeltaNetModel, GatedDeltaProductForCausalLM, GatedDeltaProductModel, GLAForCausalLM, GLAModel, GSAForCausalLM, GSAModel, HGRN2ForCausalLM, HGRN2Model, HGRNForCausalLM, LightNetForCausalLM, LightNetModel, LinearAttentionForCausalLM, LinearAttentionModel, NSAForCausalLM, NSAModel, RetNetForCausalLM, RetNetModel, RWKV6ForCausalLM, RWKV6Model, RWKV7ForCausalLM, RWKV7Model, TransformerForCausalLM, TransformerModel ) __all__ = [ 'ABCAttention', 'Attention', 'BasedLinearAttention', 'BitAttention', 'DeltaNet', 'GatedDeltaNet', 'GatedDeltaProduct', 'GatedLinearAttention', 'GatedSlotAttention', 'HGRNAttention', 'HGRN2Attention', 'LightNetAttention', 'LinearAttention', 'MultiScaleRetention', 'NativeSparseAttention', 'ReBasedLinearAttention', 'RWKV6Attention', 'RWKV7Attention', 'ABCForCausalLM', 'ABCModel', 'BitNetForCausalLM', 'BitNetModel', 'DeltaNetForCausalLM', 'DeltaNetModel', 'GatedDeltaNetForCausalLM', 'GatedDeltaNetModel', 'GatedDeltaProductForCausalLM', 'GatedDeltaProductModel', 'GLAForCausalLM', 'GLAModel', 'GSAForCausalLM', 'GSAModel', 'HGRNForCausalLM', 'HGRNModel', 'HGRN2ForCausalLM', 'HGRN2Model', 'LightNetForCausalLM', 'LightNetModel', 'LinearAttentionForCausalLM', 'LinearAttentionModel', 'NSAForCausalLM', 'NSAModel', 'RetNetForCausalLM', 'RetNetModel', 'RWKV6ForCausalLM', 'RWKV6Model', 'RWKV7ForCausalLM', 'RWKV7Model', 'TransformerForCausalLM', 'TransformerModel', ] __version__ = '0.1.2'