Skip to content
项目
群组
代码片段
帮助
当前项目
正在载入...
登录 / 注册
切换导航面板
N
NiuTrans.Tensor
概览
Overview
Details
Activity
Cycle Analytics
版本库
Repository
Files
Commits
Branches
Tags
Contributors
Graph
Compare
Charts
问题
0
Issues
0
列表
Board
标记
里程碑
合并请求
0
Merge Requests
0
CI / CD
CI / CD
流水线
作业
日程表
图表
维基
Wiki
代码片段
Snippets
成员
Collapse sidebar
Close sidebar
活动
图像
聊天
创建新问题
作业
提交
Issue Boards
Open sidebar
Emmay
NiuTrans.Tensor
Commits
35e084b0
Commit
35e084b0
authored
Oct 15, 2018
by
xiaotong
Browse files
Options
Browse Files
Download
Email Patches
Plain Diff
add padding on the encoder side for t2t MT
parent
6f90577d
隐藏空白字符变更
内嵌
并排
正在显示
4 个修改的文件
包含
115 行增加
和
58 行删除
+115
-58
source/sample/transformer/T2TModel.cpp
+35
-2
source/sample/transformer/T2TModel.h
+1
-1
source/sample/transformer/T2TTrainer.cpp
+73
-52
source/sample/transformer/T2TTrainer.h
+6
-3
没有找到文件。
source/sample/transformer/T2TModel.cpp
查看文件 @
35e084b0
...
...
@@ -174,10 +174,10 @@ make the network for machine translation (with the output softmax layer)
>> inputEnc - input tensor of the encoder
>> inputDec - input tensor of the decoder
>> output - output tensor (distribution)
>> padding
- padding of the sequences
>> padding
Enc - padding of the sequences (on the encoder side)
>> isTraining - indicates whether the model is for training
*/
void
T2TModel
::
MakeMT
(
XTensor
&
inputEnc
,
XTensor
&
inputDec
,
XTensor
&
output
,
XTensor
&
padding
,
bool
isTraining
)
void
T2TModel
::
MakeMT
(
XTensor
&
inputEnc
,
XTensor
&
inputDec
,
XTensor
&
output
,
XTensor
&
padding
Enc
,
bool
isTraining
)
{
XTensor
encoding
;
XTensor
decoding
;
...
...
@@ -199,11 +199,44 @@ void T2TModel::MakeMT(XTensor &inputEnc, XTensor &inputDec, XTensor &output, XTe
_SetDataLowTri
(
&
maskDec
,
1e9
F
,
0
);
_ScaleAndShiftMe
(
&
maskDec
,
1.0
F
,
-
1e9
F
);
/* padding on the source side */
int
*
dimsPadding
=
new
int
[
paddingEnc
.
order
+
2
];
for
(
int
i
=
0
;
i
<
paddingEnc
.
order
-
1
;
i
++
)
dimsPadding
[
i
]
=
paddingEnc
.
GetDim
(
i
);
dimsPadding
[
paddingEnc
.
order
-
1
]
=
paddingEnc
.
GetDim
(
-
1
);
dimsPadding
[
paddingEnc
.
order
]
=
paddingEnc
.
GetDim
(
-
1
);
XTensor
*
padding2
=
NewTensorBuf
(
paddingEnc
.
order
+
1
,
dimsPadding
,
paddingEnc
.
dataType
,
paddingEnc
.
denseRatio
,
paddingEnc
.
devID
,
paddingEnc
.
mem
);
for
(
int
i
=
0
;
i
<
padding2
->
order
;
i
++
)
dimsPadding
[
i
+
1
]
=
padding2
->
GetDim
(
i
);
dimsPadding
[
0
]
=
nhead
;
XTensor
*
padding3
=
NewTensorBuf
(
paddingEnc
.
order
+
2
,
dimsPadding
,
paddingEnc
.
dataType
,
paddingEnc
.
denseRatio
,
paddingEnc
.
devID
,
paddingEnc
.
mem
);
/* mask of the padding */
_Unsqueeze
(
&
paddingEnc
,
padding2
,
paddingEnc
.
order
-
1
,
paddingEnc
.
GetDim
(
-
1
));
_Unsqueeze
(
padding2
,
padding3
,
0
,
nhead
);
_ScaleAndShiftMe
(
padding3
,
1e9
F
,
-
1e9
F
);
InitTensor
(
&
maskEnc
,
padding3
);
maskEnc
.
SetZeroAll
();
/* generate the mask on the source language side (for padding) */
_Sum
(
&
maskEnc
,
padding3
,
&
maskEnc
);
encoding
=
MakeEncoder
(
inputEnc
,
maskEnc
,
isTraining
);
decoding
=
MakeDecoder
(
inputDec
,
encoding
,
maskDec
,
isTraining
);
outputLayer
.
Make
(
decoding
,
output
);
delete
[]
dims
;
delete
[]
dimsPadding
;
DelTensorBuf
(
padding2
);
DelTensorBuf
(
padding3
);
}
/*
...
...
source/sample/transformer/T2TModel.h
查看文件 @
35e084b0
...
...
@@ -78,7 +78,7 @@ public:
void
MakeLM
(
XTensor
&
input
,
XTensor
&
output
,
XTensor
&
padding
,
bool
isTraining
);
/* make the network for machine translation (with the output softmax layer) */
void
MakeMT
(
XTensor
&
inputEnc
,
XTensor
&
inputDec
,
XTensor
&
output
,
XTensor
&
padding
,
bool
isTraining
);
void
MakeMT
(
XTensor
&
inputEnc
,
XTensor
&
inputDec
,
XTensor
&
output
,
XTensor
&
padding
Enc
,
bool
isTraining
);
/* get parameter matrics */
void
GetParams
(
XList
&
list
);
...
...
source/sample/transformer/T2TTrainer.cpp
查看文件 @
35e084b0
...
...
@@ -183,7 +183,8 @@ void T2TTrainer::Train(const char * fn, const char * validFN, const char * model
XTensor
batch
;
/* padding */
XTensor
padding
;
XTensor
paddingEnc
;
XTensor
paddingDec
;
/* gold standard */
XTensor
gold
;
...
...
@@ -191,7 +192,8 @@ void T2TTrainer::Train(const char * fn, const char * validFN, const char * model
/* label smoothed gold standard (if needed) */
XTensor
goldSmoothed
;
while
(
LoadBatch
(
file
,
model
->
isLM
,
&
batch
,
&
padding
,
&
gold
,
NULL
,
vSize
,
vSizeTgt
,
while
(
LoadBatch
(
file
,
model
->
isLM
,
&
batch
,
&
paddingEnc
,
&
gold
,
&
paddingDec
,
NULL
,
vSize
,
vSizeTgt
,
sBatchSize
,
wBatchSize
,
isLenSorted
,
wc
,
devID
,
mem
))
{
...
...
@@ -202,9 +204,9 @@ void T2TTrainer::Train(const char * fn, const char * validFN, const char * model
/* make the network */
if
(
model
->
isLM
)
model
->
MakeLM
(
batch
,
output
,
padding
,
true
);
model
->
MakeLM
(
batch
,
output
,
padding
Enc
,
true
);
else
if
(
model
->
isMT
)
model
->
MakeMT
(
batch
,
gold
,
output
,
padding
,
true
);
model
->
MakeMT
(
batch
,
gold
,
output
,
padding
Enc
,
true
);
else
{
ShowNTErrors
(
"Illegal model type!"
);
}
...
...
@@ -215,7 +217,7 @@ void T2TTrainer::Train(const char * fn, const char * validFN, const char * model
/* make paddings for the output */
if
(
output
.
GetDim
(
0
)
>
1
)
PadOutput
(
&
output
,
&
gold
,
&
padding
);
PadOutput
(
&
output
,
&
gold
,
&
padding
Dec
);
//output.Dump(tmpFILE, "output: ");
//fflush(tmpFILE);
...
...
@@ -230,7 +232,7 @@ void T2TTrainer::Train(const char * fn, const char * validFN, const char * model
if
(
doUpdate
)
{
/* recale the output for normalized loss */
RescaleOutput
(
&
output
,
&
g
,
&
padding
);
RescaleOutput
(
&
output
,
&
g
,
&
padding
Dec
);
/* back-propagation */
net
.
Backward
(
output
,
g
,
CROSSENTROPY
);
...
...
@@ -331,7 +333,8 @@ void T2TTrainer::Test(const char * fn, const char * ofn, T2TModel * model)
XTensor
batch
;
/* padding */
XTensor
padding
;
XTensor
paddingEnc
;
XTensor
paddingDec
;
/* gold standard */
XTensor
gold
;
...
...
@@ -341,7 +344,8 @@ void T2TTrainer::Test(const char * fn, const char * ofn, T2TModel * model)
ClearBuf
();
while
(
LoadBatch
(
file
,
model
->
isLM
,
&
batch
,
&
padding
,
&
gold
,
seqs
,
vSize
,
vSizeTgt
,
while
(
LoadBatch
(
file
,
model
->
isLM
,
&
batch
,
&
paddingEnc
,
&
gold
,
&
paddingDec
,
seqs
,
vSize
,
vSizeTgt
,
1
,
1
,
false
,
wc
,
devID
,
mem
))
{
...
...
@@ -352,9 +356,9 @@ void T2TTrainer::Test(const char * fn, const char * ofn, T2TModel * model)
/* make the network */
if
(
model
->
isLM
)
model
->
MakeLM
(
batch
,
output
,
padding
,
false
);
model
->
MakeLM
(
batch
,
output
,
padding
Enc
,
false
);
else
if
(
model
->
isMT
)
model
->
MakeMT
(
batch
,
gold
,
output
,
padding
,
false
);
model
->
MakeMT
(
batch
,
gold
,
output
,
padding
Enc
,
false
);
else
{
ShowNTErrors
(
"Illegal model type!"
);
}
...
...
@@ -560,6 +564,7 @@ int T2TTrainer::LoadBuf(FILE * file, bool isSorted, int step)
buf
=
buf2
;
buf2
=
tmp
;
tmp
=
seqLen
;
seqLen
=
seqLen2
;
seqLen2
=
tmp
;
...
...
@@ -580,9 +585,10 @@ void T2TTrainer::ClearBuf()
load a batch of sequences
>> file - the handle to the data file
>> isLM - indicates whether the data is used for training lms
>> batch - the batch of the input sequences
>> padding - padding of the input sequences
>> output - the batch of the output sequences
>> batchEnc - the batch of the input sequences
>> paddingEnc - padding of the input sequences
>> batchDec - the batch of the output sequences
>> paddingDec - padding of the output sequences
>> seqs - keep the sequences in an array
>> vsEnc - size of the encoder vocabulary
>> vsDec - size of the decoder vocabulary
...
...
@@ -594,19 +600,20 @@ load a batch of sequences
>> mem - memory pool
*/
int
T2TTrainer
::
LoadBatch
(
FILE
*
file
,
bool
isLM
,
XTensor
*
batch
,
XTensor
*
padding
,
XTensor
*
output
,
XTensor
*
batchEnc
,
XTensor
*
paddingEnc
,
XTensor
*
batchDec
,
XTensor
*
paddingDec
,
int
*
seqs
,
int
vsEnc
,
int
vsDec
,
int
sBatch
,
int
wBatch
,
bool
isSorted
,
int
&
wCount
,
int
devID
,
XMem
*
mem
)
{
if
(
isLM
){
return
LoadBatchLM
(
file
,
batch
,
padding
,
output
,
seqs
,
return
LoadBatchLM
(
file
,
batch
Enc
,
paddingEnc
,
batchDec
,
paddingDec
,
seqs
,
vsEnc
,
sBatch
,
wBatch
,
isSorted
,
wCount
,
devID
,
mem
);
}
else
{
return
LoadBatchMT
(
file
,
batch
,
padding
,
output
,
seqs
,
return
LoadBatchMT
(
file
,
batch
Enc
,
paddingEnc
,
batchDec
,
paddingDec
,
seqs
,
vsEnc
,
vsDec
,
sBatch
,
wBatch
,
isSorted
,
wCount
,
devID
,
mem
);
}
...
...
@@ -616,9 +623,10 @@ int T2TTrainer::LoadBatch(FILE * file, bool isLM,
load a batch of sequences (for LM)
>> file - the handle to the data file
>> isLM - indicates whether the data is used for training lms
>> batch - the batch of the input sequences
>> padding - padding of the input sequences
>> output - the batch of the output sequences
>> batchEnc - the batch of the input sequences
>> paddingEnc - padding of the input sequences
>> batchDec - the batch of the output sequences
>> paddingDec - padding of the output sequences
>> seqs - keep the sequences in an array
>> vs - vocabulary size
>> sBatch - batch size of sequences
...
...
@@ -629,7 +637,8 @@ load a batch of sequences (for LM)
>> mem - memory pool
*/
int
T2TTrainer
::
LoadBatchLM
(
FILE
*
file
,
XTensor
*
batch
,
XTensor
*
padding
,
XTensor
*
output
,
XTensor
*
batchEnc
,
XTensor
*
paddingEnc
,
XTensor
*
batchDec
,
XTensor
*
paddingDec
,
int
*
seqs
,
int
vs
,
int
sBatch
,
int
wBatch
,
bool
isSorted
,
int
&
wCount
,
...
...
@@ -669,20 +678,24 @@ int T2TTrainer::LoadBatchLM(FILE * file,
dims
[
1
]
=
max
;
dims
[
2
]
=
vs
;
InitTensor
(
batch
,
3
,
dims
,
X_FLOAT
,
1.0
F
,
devID
,
mem
);
InitTensor2D
(
padding
,
sc
,
max
,
X_FLOAT
,
devID
,
mem
);
InitTensor
(
output
,
3
,
dims
,
X_FLOAT
,
1.0
F
,
devID
,
mem
);
InitTensor
(
batchEnc
,
3
,
dims
,
X_FLOAT
,
1.0
F
,
devID
,
mem
);
InitTensor2D
(
paddingEnc
,
sc
,
max
,
X_FLOAT
,
devID
,
mem
);
InitTensor
(
batchDec
,
3
,
dims
,
X_FLOAT
,
1.0
F
,
devID
,
mem
);
InitTensor2D
(
paddingDec
,
sc
,
max
,
X_FLOAT
,
devID
,
mem
);
XNoder
::
MakeGrad
(
batch
);
XNoder
::
MakeGrad
(
padding
);
XNoder
::
MakeGrad
(
output
);
XNoder
::
MakeGrad
(
batchEnc
);
XNoder
::
MakeGrad
(
paddingEnc
);
XNoder
::
MakeGrad
(
batchDec
);
XNoder
::
MakeGrad
(
paddingDec
);
batch
->
SetZeroAll
();
padding
->
SetZeroAll
();
output
->
SetZeroAll
();
batch
->
grad
->
SetZeroAll
();
padding
->
grad
->
SetZeroAll
();
output
->
grad
->
SetZeroAll
();
batchEnc
->
SetZeroAll
();
paddingEnc
->
SetZeroAll
();
batchDec
->
SetZeroAll
();
paddingDec
->
SetZeroAll
();
batchEnc
->
grad
->
SetZeroAll
();
paddingEnc
->
grad
->
SetZeroAll
();
batchDec
->
grad
->
SetZeroAll
();
paddingDec
->
grad
->
SetZeroAll
();
int
seqSize
=
0
;
...
...
@@ -693,15 +706,18 @@ int T2TTrainer::LoadBatchLM(FILE * file,
int
len
=
isDoubledEnd
?
seqLen
[
s
]
:
seqLen
[
s
]
-
1
;
CheckNTErrors
(
len
<=
max
,
"Something is wrong!"
);
for
(
int
w
=
0
;
w
<
len
;
w
++
){
batch
->
Set3D
(
1.0
F
,
s
-
seq
,
w
,
buf
[
seqOffset
[
s
]
+
w
]);
padding
->
Set2D
(
1.0
F
,
s
-
seq
,
w
);
if
(
w
>
0
)
output
->
Set3D
(
1.0
F
,
s
-
seq
,
w
-
1
,
buf
[
seqOffset
[
s
]
+
w
]);
batchEnc
->
Set3D
(
1.0
F
,
s
-
seq
,
w
,
buf
[
seqOffset
[
s
]
+
w
]);
paddingEnc
->
Set2D
(
1.0
F
,
s
-
seq
,
w
);
if
(
w
>
0
)
{
batchDec
->
Set3D
(
1.0
F
,
s
-
seq
,
w
-
1
,
buf
[
seqOffset
[
s
]
+
w
]);
paddingDec
->
Set2D
(
1.0
F
,
s
-
seq
,
w
-
1
);
}
if
(
w
==
len
-
1
){
if
(
isDoubledEnd
)
output
->
Set3D
(
1.0
F
,
s
-
seq
,
w
,
buf
[
seqOffset
[
s
]
+
w
]);
batchDec
->
Set3D
(
1.0
F
,
s
-
seq
,
w
,
buf
[
seqOffset
[
s
]
+
w
]);
else
output
->
Set3D
(
1.0
F
,
s
-
seq
,
w
,
buf
[
seqOffset
[
s
]
+
w
+
1
]);
batchDec
->
Set3D
(
1.0
F
,
s
-
seq
,
w
,
buf
[
seqOffset
[
s
]
+
w
+
1
]);
paddingDec
->
Set2D
(
1.0
F
,
s
-
seq
,
w
);
}
wCount
++
;
/*fprintf(tf, "%d", buf[seqOffset[s] + w]);
...
...
@@ -727,9 +743,10 @@ int T2TTrainer::LoadBatchLM(FILE * file,
/*
load a batch of sequences (for MT)
>> file - the handle to the data file
>> batch - the batch of the input sequences
>> padding - padding of the input sequences
>> output - the batch of the output sequences
>> batchEnc - the batch of the input sequences
>> paddingEnc - padding of the input sequences
>> batchDec - the batch of the output sequences
>> paddingDec - padding of the output sequences
>> seqs - keep the sequences in an array
>> vsEnc - size of the encoder vocabulary
>> vsDec - size of the decoder vocabulary
...
...
@@ -741,7 +758,8 @@ load a batch of sequences (for MT)
>> mem - memory pool
*/
int
T2TTrainer
::
LoadBatchMT
(
FILE
*
file
,
XTensor
*
batch
,
XTensor
*
padding
,
XTensor
*
output
,
XTensor
*
batchEnc
,
XTensor
*
paddingEnc
,
XTensor
*
batchDec
,
XTensor
*
paddingDec
,
int
*
seqs
,
int
vsEnc
,
int
vsDec
,
int
sBatch
,
int
wBatch
,
bool
isSorted
,
int
&
wCount
,
...
...
@@ -794,13 +812,15 @@ int T2TTrainer::LoadBatchMT(FILE * file,
int
dimsEnc
[
3
]
=
{
sCount
,
maxEnc
,
vsEnc
};
int
dimsDec
[
3
]
=
{
sCount
,
maxDec
,
vsDec
};
InitTensor
(
batch
,
3
,
dimsEnc
,
X_FLOAT
,
1.0
F
,
devID
,
mem
);
InitTensor2D
(
padding
,
sCount
,
maxDec
,
X_FLOAT
,
devID
,
mem
);
InitTensor
(
output
,
3
,
dimsDec
,
X_FLOAT
,
1.0
F
,
devID
,
mem
);
InitTensor
(
batchEnc
,
3
,
dimsEnc
,
X_FLOAT
,
1.0
F
,
devID
,
mem
);
InitTensor2D
(
paddingEnc
,
sCount
,
maxEnc
,
X_FLOAT
,
devID
,
mem
);
InitTensor
(
batchDec
,
3
,
dimsDec
,
X_FLOAT
,
1.0
F
,
devID
,
mem
);
InitTensor2D
(
paddingDec
,
sCount
,
maxDec
,
X_FLOAT
,
devID
,
mem
);
batch
->
SetZeroAll
();
padding
->
SetZeroAll
();
output
->
SetZeroAll
();
batchEnc
->
SetZeroAll
();
paddingEnc
->
SetZeroAll
();
batchDec
->
SetZeroAll
();
paddingDec
->
SetZeroAll
();
wCount
=
0
;
...
...
@@ -809,7 +829,8 @@ int T2TTrainer::LoadBatchMT(FILE * file,
int
len
=
seqLen
[
s
];
int
sent
=
(
s
-
seq
)
/
2
;
for
(
int
w
=
0
;
w
<
len
;
w
++
){
batch
->
Set3D
(
1.0
F
,
sent
,
w
,
buf
[
seqOffset
[
s
]
+
w
]);
batchEnc
->
Set3D
(
1.0
F
,
sent
,
w
,
buf
[
seqOffset
[
s
]
+
w
]);
paddingEnc
->
Set2D
(
1.0
F
,
sent
,
w
);
wCount
++
;
}
}
...
...
@@ -819,11 +840,11 @@ int T2TTrainer::LoadBatchMT(FILE * file,
int
len
=
seqLen
[
s
];
int
sent
=
(
s
-
seq
-
1
)
/
2
;
for
(
int
w
=
0
;
w
<
len
;
w
++
){
padding
->
Set2D
(
1.0
F
,
sent
,
w
);
padding
Dec
->
Set2D
(
1.0
F
,
sent
,
w
);
if
(
w
>
0
)
output
->
Set3D
(
1.0
F
,
sent
,
w
-
1
,
buf
[
seqOffset
[
s
]
+
w
]);
batchDec
->
Set3D
(
1.0
F
,
sent
,
w
-
1
,
buf
[
seqOffset
[
s
]
+
w
]);
if
(
w
==
len
-
1
)
output
->
Set3D
(
1.0
F
,
sent
,
w
,
buf
[
seqOffset
[
s
]
+
w
]);
batchDec
->
Set3D
(
1.0
F
,
sent
,
w
,
buf
[
seqOffset
[
s
]
+
w
]);
wCount
++
;
if
(
seqs
!=
NULL
)
...
...
source/sample/transformer/T2TTrainer.h
查看文件 @
35e084b0
...
...
@@ -166,7 +166,8 @@ public:
/* load a batch of sequences */
int
LoadBatch
(
FILE
*
file
,
bool
isLM
,
XTensor
*
batch
,
XTensor
*
padding
,
XTensor
*
output
,
XTensor
*
batchEnc
,
XTensor
*
paddingEnc
,
XTensor
*
batchDec
,
XTensor
*
paddingDec
,
int
*
seqs
,
int
vsEnc
,
int
vsDec
,
int
sBatch
,
int
wBatch
,
bool
isSorted
,
int
&
wCount
,
...
...
@@ -174,14 +175,16 @@ public:
/* load a batch of sequences (for language modeling) */
int
LoadBatchLM
(
FILE
*
file
,
XTensor
*
batch
,
XTensor
*
padding
,
XTensor
*
output
,
XTensor
*
batchEnc
,
XTensor
*
paddingEnc
,
XTensor
*
batchDec
,
XTensor
*
paddingDec
,
int
*
seqs
,
int
vs
,
int
sBatch
,
int
wBatch
,
bool
isSorted
,
int
&
wCount
,
int
devID
,
XMem
*
mem
);
/* load a batch of sequences (for machine translation) */
int
LoadBatchMT
(
FILE
*
file
,
XTensor
*
batch
,
XTensor
*
padding
,
XTensor
*
output
,
XTensor
*
batchEnc
,
XTensor
*
paddingEnc
,
XTensor
*
batchDec
,
XTensor
*
paddingDec
,
int
*
seqs
,
int
vsEnc
,
int
vsDec
,
int
sBatch
,
int
wBatch
,
bool
isSorted
,
int
&
wCount
,
int
devID
,
XMem
*
mem
);
...
...
编写
预览
Markdown
格式
0%
重试
或
添加新文件
添加附件
取消
您添加了
0
人
到此讨论。请谨慎行事。
请先完成此评论的编辑!
取消
请
注册
或者
登录
后发表评论