|
|
|
|
|
from fla.layers import (ABCAttention, Attention, BasedLinearAttention, |
|
BitAttention, DeltaNet, GatedLinearAttention, |
|
GatedSlotAttention, HGRN2Attention, HGRNAttention, |
|
LinearAttention, MultiScaleRetention, |
|
ReBasedLinearAttention) |
|
from fla.models import (ABCForCausalLM, ABCModel, BitNetForCausalLM, |
|
BitNetModel, DeltaNetForCausalLM, DeltaNetModel, |
|
GLAForCausalLM, GLAModel, GSAForCausalLM, GSAModel, |
|
HGRN2ForCausalLM, HGRN2Model, HGRNForCausalLM, |
|
LinearAttentionForCausalLM, LinearAttentionModel, |
|
RetNetForCausalLM, RetNetModel, RWKV6ForCausalLM, |
|
RWKV6Model, TransformerForCausalLM, TransformerModel) |
|
|
|
__all__ = [ |
|
'ABCAttention', |
|
'Attention', |
|
'BasedLinearAttention', |
|
'BitAttention', |
|
'DeltaNet', |
|
'HGRNAttention', |
|
'HGRN2Attention', |
|
'GatedLinearAttention', |
|
'GatedSlotAttention', |
|
'LinearAttention', |
|
'MultiScaleRetention', |
|
'ReBasedLinearAttention', |
|
'ABCForCausalLM', |
|
'ABCModel', |
|
'BitNetForCausalLM', |
|
'BitNetModel', |
|
'DeltaNetForCausalLM', |
|
'DeltaNetModel', |
|
'HGRNForCausalLM', |
|
'HGRNModel', |
|
'HGRN2ForCausalLM', |
|
'HGRN2Model', |
|
'GLAForCausalLM', |
|
'GLAModel', |
|
'GSAForCausalLM', |
|
'GSAModel', |
|
'LinearAttentionForCausalLM', |
|
'LinearAttentionModel', |
|
'RetNetForCausalLM', |
|
'RetNetModel', |
|
'RWKV6ForCausalLM', |
|
'RWKV6Model', |
|
'TransformerForCausalLM', |
|
'TransformerModel', |
|
'chunk_gla', |
|
'chunk_retention', |
|
'fused_chunk_based', |
|
'fused_chunk_gla', |
|
'fused_chunk_retention' |
|
] |
|
|
|
__version__ = '0.1' |
|
|