scan-8192-16M-test / fla /__init__.py
zaydzuhri's picture
Training in progress, step 5000
a0806ea verified
raw
history blame
1.79 kB
# -*- coding: utf-8 -*-
from fla.layers import (ABCAttention, Attention, BasedLinearAttention,
BitAttention, DeltaNet, GatedLinearAttention,
GatedSlotAttention, HGRN2Attention, HGRNAttention,
LinearAttention, MultiScaleRetention,
ReBasedLinearAttention)
from fla.models import (ABCForCausalLM, ABCModel, BitNetForCausalLM,
BitNetModel, DeltaNetForCausalLM, DeltaNetModel,
GLAForCausalLM, GLAModel, GSAForCausalLM, GSAModel,
HGRN2ForCausalLM, HGRN2Model, HGRNForCausalLM,
LinearAttentionForCausalLM, LinearAttentionModel,
RetNetForCausalLM, RetNetModel, RWKV6ForCausalLM,
RWKV6Model, TransformerForCausalLM, TransformerModel)
__all__ = [
'ABCAttention',
'Attention',
'BasedLinearAttention',
'BitAttention',
'DeltaNet',
'HGRNAttention',
'HGRN2Attention',
'GatedLinearAttention',
'GatedSlotAttention',
'LinearAttention',
'MultiScaleRetention',
'ReBasedLinearAttention',
'ABCForCausalLM',
'ABCModel',
'BitNetForCausalLM',
'BitNetModel',
'DeltaNetForCausalLM',
'DeltaNetModel',
'HGRNForCausalLM',
'HGRNModel',
'HGRN2ForCausalLM',
'HGRN2Model',
'GLAForCausalLM',
'GLAModel',
'GSAForCausalLM',
'GSAModel',
'LinearAttentionForCausalLM',
'LinearAttentionModel',
'RetNetForCausalLM',
'RetNetModel',
'RWKV6ForCausalLM',
'RWKV6Model',
'TransformerForCausalLM',
'TransformerModel',
'chunk_gla',
'chunk_retention',
'fused_chunk_based',
'fused_chunk_gla',
'fused_chunk_retention'
]
__version__ = '0.1'