Home | History | Annotate | Download | only in kernels

Lines Matching refs:features

31   // features: any shape.
32 // activations: same shape as "features".
33 void operator()(const Device& d, typename TTypes<T>::ConstTensor features,
35 activations.device(d) = features.cwiseMax(static_cast<T>(0));
45 // features: either the inputs that were passed to the Relu or, or its
49 typename TTypes<T>::ConstTensor features,
55 gradients * (features > static_cast<T>(0)).template cast<T>();
64 // features: any shape.
65 // activations: same shape as "features".
66 void operator()(const Device& d, typename TTypes<T>::ConstTensor features,
69 features.cwiseMax(static_cast<T>(0)).cwiseMin(static_cast<T>(6));
79 // features: inputs that where passed to the Relu6 op, or its outputs.
82 typename TTypes<T>::ConstTensor features,
86 // value. This allows "features" to be either the input or the output of
88 backprops.device(d) = gradients * ((features > static_cast<T>(0)) *
89 (features < static_cast<T>(6)))
99 // features: any shape.
100 // activations: same shape as "features".
101 void operator()(const Device& d, typename TTypes<T>::ConstTensor features,
103 // features.constant(?)
105 (features < static_cast<T>(0))
106 .select(features.exp() - features.constant(static_cast<T>(1)),
107 features);
133 // features: any shape.
134 // activations: same shape as "features".
135 void operator()(const Device& d, typename TTypes<T>::ConstTensor features,
137 // features.constant(?)
143 (features < zero)
144 .select(scale_alpha * (features.exp() - features.constant(one)),
145 scale * features);