HardTanH.cpp 3.65 KB
Newer Older
xiaotong committed
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22
/* NiuTrans.Tensor - an open-source tensor library
 * Copyright (C) 2017, Natural Language Processing Lab, Northestern University. 
 * All rights reserved.
 *
 * Licensed under the Apache License, Version 2.0 (the "License");
 * you may not use this file except in compliance with the License.
 * You may obtain a copy of the License at
 *
 *   http://www.apache.org/licenses/LICENSE-2.0
 *
 * Unless required by applicable law or agreed to in writing, software
 * distributed under the License is distributed on an "AS IS" BASIS,
 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
 * See the License for the specific language governing permissions and
 * limitations under the License.
 */

/*
* $Created by: XIAO Tong (email: xiaotong@mail.neu.edu.cn) 2018-04-25
*/

#include <stdlib.h>
23
#include "../XName.h"
xiaotong committed
24 25
#include "HardTanH.h"
#include "HardTanH.cuh"
26
#include "CrossEntropy.h"
xiaotong committed
27 28 29 30 31 32 33 34 35 36 37

namespace nts{ // namespace nts(NiuTrans.Tensor)

/*
hard tanh function 
y =  1    if x > 1
     x    if -1 <= x <= 1
    -1    if x < -1
>> x - input tensor
>> y - result
*/
38
void _HardTanH(const XTensor * x, XTensor * y)
xiaotong committed
39 40 41
{
#ifdef USE_CUDA
    if(x->devID >= 0 || y->devID >= 0){
42
        _CudaHardTanH(x, y);
xiaotong committed
43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62
        return;
    }
#endif
    if(x->dataType == DEFAULT_DTYPE && y->dataType == DEFAULT_DTYPE){
        int n = x->GetSize();
        DTYPE * ip = (DTYPE*)x->data;
        DTYPE * op = (DTYPE*)y->data;
        for(int i = 0; i < n; i++){
            DTYPE p = ip[i];
            if(p > 1.0)
                p = 1.0;
            else if(p < -1.0)
                p = -1.0;
            op[i] = p;
        }
    }
    else
        ShowNTErrors("TODO!");
}

63 64 65 66 67 68 69 70 71 72 73 74 75
/* 
hard tanh function (return a XTensor structure) 
make a new tensor to keep the result and return it

y =  1    if x > 1
     x    if -1 <= x <= 1
    -1    if x < -1
>> x - input tensor
<< return - y
*/
XTensor HardTanH(const XTensor &x)
{
    XTensor y(&x);
76
    y.SetTMPFlag();
77 78 79 80 81 82 83 84 85 86

    /* call _HardTanH function */
    _HardTanH(&x, &y);

    /* tensor connection */
    XLink::MakeLink(&x, NULL, &y, FUNC_HARDTANH);

    return y;
}

xiaotong committed
87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105
/*
backward computation

dE/dx = dE/dy * dy/dx

hard tanh: y =  1    if x > 1
                x    if -1 <= x <= 1
               -1    if x< -1

   and dy/dx =  1    if -1 <= x <= 1
                0    otherwise

>> gold - gold standard to measure error (or loss)
>> y - output of the function
>> x - input of the function
>> dedy - dE/dy
>> dedx - dE/dx
>> lossName - type of loss function, e.g., cross entropy
*/
106 107 108
void _HardTanHBackward(XTensor * gold, XTensor * y, XTensor * x, 
                       XTensor * dedy, XTensor * dedx,
                       LOSS_FUNCTION_NAME lossName)
xiaotong committed
109
{
110
    CheckNTErrors((gold == NULL || XTensor::IsSameShaped(gold, y)), 
111
                   "The tensors must be of the same size!");
xiaotong committed
112 113 114

#ifdef USE_CUDA
    if(x->devID >= 0 || y->devID >= 0){
115
        _CudaHardTanHBackward(gold, y, x, dedy, dedx, lossName);
xiaotong committed
116 117 118 119
        return;
    }
#endif

120
    if(x->dataType == DEFAULT_DTYPE && y->dataType == DEFAULT_DTYPE){
xiaotong committed
121
        /* calculate dE/dy */
122 123 124
        if(lossName == CROSSENTROPY)
            _CrossEntropyBackward(dedy, y, gold);
        else if(lossName != NOLOSS)
125
            _LossBackward(dedy, gold, y, lossName);
xiaotong committed
126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145

        DTYPE * dedyp = (DTYPE*)dedy->data;
        DTYPE * dedxp = (DTYPE*)dedx->data;
        DTYPE * ip = (DTYPE*)x->data;
        int size = y->unitNum;

        /* dE/dx = dE/dy * dy/dx */
        for(int i = 0; i < size; i++){
            DTYPE s =ip[i];
            if(s > 1.0 || s < -1.0)
                dedxp[i] = 0;
            else
                dedxp[i] = dedyp[i];
        }
    }
    else
        ShowNTErrors("TODO!");
}

} // namespace nts(NiuTrans.Tensor)