Skip to content
项目
群组
代码片段
帮助
当前项目
正在载入...
登录 / 注册
切换导航面板
N
NiuTrans.Tensor
概览
Overview
Details
Activity
Cycle Analytics
版本库
Repository
Files
Commits
Branches
Tags
Contributors
Graph
Compare
Charts
问题
0
Issues
0
列表
Board
标记
里程碑
合并请求
0
Merge Requests
0
CI / CD
CI / CD
流水线
作业
日程表
图表
维基
Wiki
代码片段
Snippets
成员
Collapse sidebar
Close sidebar
活动
图像
聊天
创建新问题
作业
提交
Issue Boards
Open sidebar
杨迪
NiuTrans.Tensor
Commits
fb841a0a
Commit
fb841a0a
authored
Mar 18, 2020
by
liyinqiao
Browse files
Options
Browse Files
Download
Email Patches
Plain Diff
Merge with XU Chen branch
Update the Transformer codes.
parent
549e6d0f
全部展开
隐藏空白字符变更
内嵌
并排
正在显示
8 个修改的文件
包含
52 行增加
和
151 行删除
+52
-151
source/sample/transformer/T2TEmbedding.cpp
+9
-21
source/sample/transformer/T2TModel.cpp
+37
-111
source/sample/transformer/T2TModel.h
+2
-3
source/sample/transformer/T2TPredictor.cpp
+1
-1
source/sample/transformer/T2TSearch.cpp
+1
-1
source/sample/transformer/T2TTrainer.cpp
+0
-0
source/sample/transformer/T2TTrainer.h
+1
-13
source/sample/transformer/Transformer.cpp
+1
-1
没有找到文件。
source/sample/transformer/T2TEmbedding.cpp
查看文件 @
fb841a0a
...
...
@@ -131,32 +131,20 @@ XTensor T2TEmbedder::Make(XTensor &input)
XTensor
wordEmbedding
;
XTensor
posEmbedding
;
bool
match
=
(
posEmbedding
.
order
==
input
.
order
);
if
(
match
){
for
(
int
i
=
0
;
i
<
input
.
order
;
i
++
){
if
(
dims
[
i
]
!=
posEmbedding
.
GetDim
(
i
))
match
=
false
;
}
}
/* we make positional embeddings first */
//if(!match){
if
(
true
){
InitTensor
(
&
posEmbedding
,
input
.
order
+
1
,
dims
,
X_FLOAT
,
devID
);
/* make positional embeddings */
XTensor
position
;
XTensor
embTMP
;
XTensor
*
posTMP
=
NewTensorBuf
(
2
,
dims
+
1
,
X_FLOAT
,
devID
);
_CopyValues
(
&
posEmbeddingBase
,
0
,
posTMP
->
unitNum
,
posTMP
,
0
);
_Unsqueeze
(
posTMP
,
&
posEmbedding
,
0
,
dims
[
0
]);
DelTensorBuf
(
posTMP
);
}
InitTensor1D
(
&
position
,
input
.
GetDim
(
-
1
),
X_INT
,
devID
);
position
.
Range
(
0
,
position
.
unitNum
,
1
);
embTMP
=
Gather
(
posEmbeddingBase
,
position
);
posEmbedding
=
Unsqueeze
(
embTMP
,
0
,
dims
[
0
]);
/*
then we
make word embeddings */
/* make word embeddings */
wordEmbedding
=
Gather
(
w
,
input
);
wordEmbedding
=
Linear
(
wordEmbedding
,
(
float
)
sqrt
((
float
)
eSize
));
/*
we
sum over the two embeddings */
/* sum over the two embeddings */
return
wordEmbedding
+
posEmbedding
;
}
...
...
source/sample/transformer/T2TModel.cpp
查看文件 @
fb841a0a
...
...
@@ -114,64 +114,28 @@ make the network for language modeling (with the output softmax layer)
*/
void
T2TModel
::
MakeLM
(
XTensor
&
input
,
XTensor
&
output
,
XTensor
&
padding
,
bool
isTraining
)
{
XTensor
encoding
;
/* generate mask to see "previous" words only */
//int len = input.GetDim(input.order - 2);
//int * dims = new int[input.order + 1];
//for(int i = 0; i < input.order; i++)
// dims[i + 1] = input.GetDim(i);
//dims[0] = nhead;
//dims[input.order] = len;
//XTensor mask(input.order + 1, dims, X_FLOAT, 1.0F, input.devID, input.mem);
int
len
=
input
.
GetDim
(
input
.
order
-
1
);
int
*
dims
=
new
int
[
input
.
order
+
2
];
for
(
int
i
=
0
;
i
<
input
.
order
;
i
++
)
dims
[
i
+
1
]
=
input
.
GetDim
(
i
);
int
len
=
padding
.
GetDim
(
padding
.
order
-
1
);
int
*
dims
=
new
int
[
padding
.
order
+
2
];
for
(
int
i
=
0
;
i
<
padding
.
order
;
i
++
)
dims
[
i
+
1
]
=
padding
.
GetDim
(
i
);
dims
[
0
]
=
nhead
;
dims
[
input
.
order
+
1
]
=
len
;
dims
[
padding
.
order
+
1
]
=
len
;
XTensor
mask
;
InitTensor
(
&
mask
,
input
.
order
+
2
,
dims
,
X_FLOAT
,
padding
.
devID
);
InitTensor
(
&
mask
,
padding
.
order
+
2
,
dims
,
X_FLOAT
,
padding
.
devID
);
delete
[]
dims
;
/* a upper triangular matrix where the cells of the upper triangular are set to -1e-9.
this matrix can be used to prevent the attention to current or following words in
a given sequence. */
_SetDataLowTri
(
&
mask
,
1e9
F
,
0
);
_ScaleAndShiftMe
(
&
mask
,
1.0
F
,
-
1e9
F
);
int
*
dimsPadding
=
new
int
[
padding
.
order
+
2
];
for
(
int
i
=
0
;
i
<
padding
.
order
-
1
;
i
++
)
dimsPadding
[
i
]
=
padding
.
GetDim
(
i
);
dimsPadding
[
padding
.
order
-
1
]
=
padding
.
GetDim
(
-
1
);
dimsPadding
[
padding
.
order
]
=
padding
.
GetDim
(
-
1
);
XTensor
*
padding2
=
NewTensorBuf
(
padding
.
order
+
1
,
dimsPadding
,
padding
.
dataType
,
padding
.
devID
);
for
(
int
i
=
0
;
i
<
padding2
->
order
;
i
++
)
dimsPadding
[
i
+
1
]
=
padding2
->
GetDim
(
i
);
dimsPadding
[
0
]
=
nhead
;
ScaleAndShiftMe
(
mask
,
1.0
F
,
-
1e9
F
);
//XTensor * padding3 = NewTensorBuf(padding.order + 2, dimsPadding, padding.dataType,
// padding.devID);
//
///* mask of the padding */
//_Unsqueeze(&padding, padding2, padding.order - 1, padding.GetDim(-1));
//_Unsqueeze(padding2, padding3, 0, nhead);
//
//_ScaleAndShiftMe(padding3, 1e9F, -1e9F);
//
////_Sum(&mask, padding3, &mask);
/* forward */
XTensor
encoding
;
encoding
=
MakeEncoder
(
input
,
mask
,
isTraining
);
outputLayer
->
Make
(
encoding
,
output
);
delete
[]
dims
;
delete
[]
dimsPadding
;
//DelTensorBuf(padding3);
DelTensorBuf
(
padding2
);
}
/*
...
...
@@ -183,7 +147,9 @@ make the network for machine translation (with the output softmax layer)
>> paddingDec - padding of the sequences (on the decoder side)
>> isTraining - indicates whether the model is for training
*/
void
T2TModel
::
MakeMT
(
XTensor
&
inputEnc
,
XTensor
&
inputDec
,
XTensor
&
output
,
XTensor
&
paddingEnc
,
XTensor
&
paddingDec
,
bool
isTraining
)
void
T2TModel
::
MakeMT
(
XTensor
&
inputEnc
,
XTensor
&
inputDec
,
XTensor
&
output
,
XTensor
&
paddingEnc
,
XTensor
&
paddingDec
,
bool
isTraining
)
{
XTensor
encoding
;
XTensor
decoding
;
...
...
@@ -192,10 +158,10 @@ void T2TModel::MakeMT(XTensor &inputEnc, XTensor &inputDec, XTensor &output, XTe
XTensor
maskEncDec
;
/* encoder mask */
MakeMTMaskEnc
(
inputEnc
,
paddingEnc
,
maskEnc
);
MakeMTMaskEnc
(
paddingEnc
,
maskEnc
);
/* decoder mask */
MakeMTMaskDec
(
inputEnc
,
inputDec
,
paddingEnc
,
paddingDec
,
maskDec
,
maskEncDec
);
MakeMTMaskDec
(
paddingEnc
,
paddingDec
,
maskDec
,
maskEncDec
);
encoding
=
MakeEncoder
(
inputEnc
,
maskEnc
,
isTraining
);
...
...
@@ -289,40 +255,21 @@ make the mask of the encoder
>> paddingEnc - padding of the encoder input
>> maskEnc - mask of the encoder self-attention
*/
void
T2TModel
::
MakeMTMaskEnc
(
XTensor
&
inputEnc
,
XTensor
&
paddingEnc
,
XTensor
&
maskEnc
)
void
T2TModel
::
MakeMTMaskEnc
(
XTensor
&
paddingEnc
,
XTensor
&
maskEnc
)
{
/* padding on the source side */
int
*
dimsPadding
=
new
int
[
paddingEnc
.
order
+
2
];
for
(
int
i
=
0
;
i
<
paddingEnc
.
order
-
1
;
i
++
)
dimsPadding
[
i
]
=
paddingEnc
.
GetDim
(
i
);
dimsPadding
[
paddingEnc
.
order
-
1
]
=
paddingEnc
.
GetDim
(
-
1
);
dimsPadding
[
paddingEnc
.
order
]
=
paddingEnc
.
GetDim
(
-
1
);
XTensor
*
padding2
=
NewTensorBuf
(
paddingEnc
.
order
+
1
,
dimsPadding
,
paddingEnc
.
dataType
,
paddingEnc
.
devID
);
for
(
int
i
=
0
;
i
<
padding2
->
order
;
i
++
)
dimsPadding
[
i
+
1
]
=
padding2
->
GetDim
(
i
);
dimsPadding
[
0
]
=
nhead
;
XTensor
*
padding3
=
NewTensorBuf
(
paddingEnc
.
order
+
2
,
dimsPadding
,
paddingEnc
.
dataType
,
paddingEnc
.
devID
);
XTensor
padding2
;
XTensor
padding3
;
/* mask of the padding */
_Unsqueeze
(
&
paddingEnc
,
padding2
,
paddingEnc
.
order
-
1
,
paddingEnc
.
GetDim
(
-
1
));
_Unsqueeze
(
padding2
,
padding3
,
0
,
nhead
);
_ScaleAndShiftMe
(
padding3
,
1e9
F
,
-
1e9
F
);
Unsqueeze
(
paddingEnc
,
padding2
,
paddingEnc
.
order
-
1
,
paddingEnc
.
GetDim
(
-
1
));
Unsqueeze
(
padding2
,
padding3
,
0
,
nhead
);
ScaleAndShiftMe
(
padding3
,
1e9
F
,
-
1e9
F
);
InitTensor
(
&
maskEnc
,
padding3
);
InitTensor
(
&
maskEnc
,
&
padding3
);
maskEnc
.
SetZeroAll
();
/* generate the mask on the source language side (for padding) */
_Sum
(
&
maskEnc
,
padding3
,
&
maskEnc
);
DelTensorBuf
(
padding3
);
DelTensorBuf
(
padding2
);
delete
[]
dimsPadding
;
SumMe
(
maskEnc
,
padding3
);
}
/*
...
...
@@ -334,54 +281,33 @@ make the mask of the decoder
>> maksDec - mask of the decoder self-attention
>> maksEncDec - mask of the decoder enc-dec attention
*/
void
T2TModel
::
MakeMTMaskDec
(
XTensor
&
inputEnc
,
XTensor
&
inputDec
,
XTensor
&
paddingEnc
,
XTensor
&
paddingDec
,
void
T2TModel
::
MakeMTMaskDec
(
XTensor
&
paddingEnc
,
XTensor
&
paddingDec
,
XTensor
&
maskDec
,
XTensor
&
maskEncDec
)
{
int
len
=
inputDec
.
GetDim
(
input
Dec
.
order
-
1
);
int
*
dims
=
new
int
[
input
Dec
.
order
+
2
];
for
(
int
i
=
0
;
i
<
input
Dec
.
order
;
i
++
)
dims
[
i
+
1
]
=
input
Dec
.
GetDim
(
i
);
int
len
=
paddingDec
.
GetDim
(
padding
Dec
.
order
-
1
);
int
*
dims
=
new
int
[
padding
Dec
.
order
+
2
];
for
(
int
i
=
0
;
i
<
padding
Dec
.
order
;
i
++
)
dims
[
i
+
1
]
=
padding
Dec
.
GetDim
(
i
);
dims
[
0
]
=
nhead
;
dims
[
input
Dec
.
order
+
1
]
=
len
;
InitTensor
(
&
maskDec
,
input
Dec
.
order
+
2
,
dims
,
X_FLOAT
,
paddingDec
.
devID
);
dims
[
padding
Dec
.
order
+
1
]
=
len
;
InitTensor
(
&
maskDec
,
padding
Dec
.
order
+
2
,
dims
,
X_FLOAT
,
paddingDec
.
devID
);
/* An upper triangular matrix where the cells of the upper triangular are set to -1e-9.
This matrix can be used to block the attention to current or following words in
a given sequence. */
_SetDataLowTri
(
&
maskDec
,
1e9
F
,
0
);
//maskDec.Dump(stderr, "mask: ");
_ScaleAndShiftMe
(
&
maskDec
,
1.0
F
,
-
1e9
F
);
ScaleAndShiftMe
(
maskDec
,
1.0
F
,
-
1e9
F
);
//maskDec.Dump(stderr, "mask: ");
/* encoder-decoder mask that prevents the attention to padding dummy words */
dims
[
inputDec
.
order
+
1
]
=
inputEnc
.
GetDim
(
inputEnc
.
order
-
1
);
InitTensor
(
&
maskEncDec
,
inputDec
.
order
+
2
,
dims
,
X_FLOAT
,
paddingEnc
.
devID
);
XTensor
maskEncDecTMP
;
XTensor
*
maskEncDecTMPEnc
=
NewTensorBuf
(
paddingEnc
.
order
+
1
,
dims
+
1
,
paddingEnc
.
dataType
,
paddingEnc
.
devID
);
XTensor
*
maskEncDecTMPDec
=
NewTensorBuf
(
maskEncDecTMPEnc
,
paddingEnc
.
devID
);
Unsqueeze
(
paddingEnc
,
maskEncDecTMP
,
paddingEnc
.
order
-
1
,
paddingDec
.
GetDim
(
-
1
));
ScaleAndShiftMe
(
maskEncDecTMP
,
1e9
F
,
-
1e9
F
);
Unsqueeze
(
maskEncDecTMP
,
maskEncDec
,
0
,
dims
[
0
]
);
_Unsqueeze
(
&
paddingEnc
,
maskEncDecTMPEnc
,
paddingEnc
.
order
-
1
,
paddingDec
.
GetDim
(
-
1
));
//paddingEnc.Dump(stderr, "paddingenc:");
//maskEncDecTMPEnc->Dump(stderr, "maskencdectmpenc:");
_ScaleAndShiftMe
(
maskEncDecTMPEnc
,
1e9
F
,
-
1e9
F
);
//maskEncDecTMPEnc->Dump(stderr, "maskencdectmpenc:");
_Unsqueeze
(
maskEncDecTMPEnc
,
&
maskEncDec
,
0
,
dims
[
0
]);
//maskEncDecTMPEnc->Dump(stderr, "maskencdectmpenc:");
DelTensorBuf
(
maskEncDecTMPDec
);
DelTensorBuf
(
maskEncDecTMPEnc
);
delete
[]
dims
;
}
/*
get parameter matrics
>> list - the list that keeps the parameter matrics
...
...
source/sample/transformer/T2TModel.h
查看文件 @
fb841a0a
...
...
@@ -87,11 +87,10 @@ public:
XTensor
&
maskEnc
,
XTensor
&
maskDec
,
XTensor
&
maskEncDec
);
/* make the mask of the encoder */
void
MakeMTMaskEnc
(
XTensor
&
inputEnc
,
XTensor
&
paddingEnc
,
XTensor
&
maskEnc
);
void
MakeMTMaskEnc
(
XTensor
&
paddingEnc
,
XTensor
&
maskEnc
);
/* make the mask of the decoder */
void
MakeMTMaskDec
(
XTensor
&
inputEnc
,
XTensor
&
inputDec
,
XTensor
&
paddingEnc
,
XTensor
&
paddingDec
,
void
MakeMTMaskDec
(
XTensor
&
paddingEnc
,
XTensor
&
paddingDec
,
XTensor
&
maskDec
,
XTensor
&
maskEncDec
);
/* get parameter matrics */
...
...
source/sample/transformer/T2TPredictor.cpp
查看文件 @
fb841a0a
...
...
@@ -201,7 +201,7 @@ void T2TPredictor::Predict(T2TStateBundle * next, XTensor * encoding,
XTensor
maskEncDec
;
/* decoder mask */
m
->
MakeMTMaskDec
(
*
inputEnc
,
inputDec
,
*
paddingEnc
,
paddingDec
,
maskDec
,
maskEncDec
);
m
->
MakeMTMaskDec
(
*
paddingEnc
,
paddingDec
,
maskDec
,
maskEncDec
);
/* make the decoding network */
decoding
=
decoder
.
Make
(
inputDec
,
*
encoding
,
maskDec
,
maskEncDec
,
false
);
...
...
source/sample/transformer/T2TSearch.cpp
查看文件 @
fb841a0a
...
...
@@ -89,7 +89,7 @@ void T2TSearch::Search(T2TModel * model, XTensor * input, XTensor * padding, XTe
Prepare
(
input
->
unitNum
/
input
->
GetDim
(
-
1
),
beamSize
);
/* encoder mask */
model
->
MakeMTMaskEnc
(
*
input
,
*
padding
,
maskEnc
);
model
->
MakeMTMaskEnc
(
*
padding
,
maskEnc
);
//input->Dump(stderr, "input:");
//maskEnc.Dump(stderr, "maskenc:");
...
...
source/sample/transformer/T2TTrainer.cpp
查看文件 @
fb841a0a
差异被折叠。
点击展开。
source/sample/transformer/T2TTrainer.h
查看文件 @
fb841a0a
...
...
@@ -125,28 +125,16 @@ public:
void
Train
(
const
char
*
fn
,
const
char
*
validFN
,
const
char
*
modelFN
,
T2TModel
*
model
);
/* test the model */
void
Test
(
const
char
*
fn
,
const
char
*
ofn
,
T2TModel
*
model
);
void
Validate
(
const
char
*
fn
,
const
char
*
ofn
,
T2TModel
*
model
);
/* make a checkpoint */
void
MakeCheckpoint
(
T2TModel
*
model
,
const
char
*
validFN
,
const
char
*
modelFN
,
const
char
*
label
,
int
id
);
/* get word probabilities for a batch of sequences */
float
GetProb
(
XTensor
*
output
,
XTensor
*
gold
,
XTensor
*
wordProbs
);
/* update the model by delta rule */
void
Update
(
T2TModel
*
model
,
const
float
lr
);
/* prepare model for training */
void
PrepareModel
(
T2TModel
*
model
);
/* do padding on the output */
void
PadOutput
(
XTensor
*
output
,
XTensor
*
gold
,
XTensor
*
padding
);
/* recale the output and gold tensors for normalized loss */
void
RescaleOutput
(
XTensor
*
output
,
XTensor
*
gold
,
XTensor
*
padding
);
/* perform label smoothing */
void
LabelSmooth
(
XTensor
*
gold
,
XTensor
*
smoothed
,
DTYPE
p
);
};
...
...
source/sample/transformer/Transformer.cpp
查看文件 @
fb841a0a
...
...
@@ -94,7 +94,7 @@ int TransformerMain(int argc, const char ** argv)
else
{
T2TTrainer
tester
;
tester
.
Init
(
argc
,
args
);
tester
.
Test
(
testFN
,
outputFN
,
&
model
);
tester
.
Validate
(
testFN
,
outputFN
,
&
model
);
}
}
...
...
编写
预览
Markdown
格式
0%
重试
或
添加新文件
添加附件
取消
您添加了
0
人
到此讨论。请谨慎行事。
请先完成此评论的编辑!
取消
请
注册
或者
登录
后发表评论