Commit 6292949b by xuchen

fix the buf of the reduced SA

parent bd8fae9b
......@@ -44,7 +44,7 @@ class PyramidTransformerEncoderLayer(nn.Module):
self.quant_noise = getattr(args, 'quant_noise_pq', 0)
self.quant_noise_block_size = getattr(args, 'quant_noise_pq_block_size', 8) or 8
self.attn_type = getattr(args, "encoder_attention_type", "selfattn")
self.self_attn = self.build_self_attention(self.embed_dim, num_head, args)
self.self_attn = self.build_self_attention(args, self.embed_dim, num_head, att_sample_ratio)
self.self_attn_layer_norm = LayerNorm(self.embed_dim)
self.dropout_module = FairseqDropout(
args.dropout, module_name=self.__class__.__name__
......
Markdown 格式
0%
您添加了 0 到此讨论。请谨慎行事。
请先完成此评论的编辑!
注册 或者 后发表评论