Commit d11db556 by libei

support dense relative transformer decode and model convert

parent 73e8d792
......@@ -15,7 +15,7 @@ from .linearized_convolution import LinearizedConvolution
from .multihead_attention import MultiheadAttention
from .scalar_bias import ScalarBias
from .sinusoidal_positional_embedding import SinusoidalPositionalEmbedding
from .relative_multihead_attention import RelativeMultiheadAttention
__all__ = [
'AdaptiveSoftmax',
'BeamableMM',
......@@ -27,4 +27,6 @@ __all__ = [
'MultiheadAttention',
'ScalarBias',
'SinusoidalPositionalEmbedding',
'RelativeMultiheadAttention'
''
]
......@@ -472,6 +472,8 @@ def convert_settings(settings):
args['max_relative_length'] = int(settings['max_relative_length'])
args['arch'] = 'relative_transformer'
if use_relative_position_representation and use_dense:
args['arch'] = 'relative_dense_transformer'
return argparse.Namespace(**args)
......
Markdown 格式
0%
您添加了 0 到此讨论。请谨慎行事。
请先完成此评论的编辑!
注册 或者 后发表评论