36092e2ae66342ee1d7ad5678cd99c46090373eb,src/gluonnlp/layers.py,PositionwiseFFN,__init__,#PositionwiseFFN#,496
Before Change
])
self.dropout_layer = nn.Dropout(dropout)
self.activation_dropout_layer = nn.Dropout(activation_dropout)
self.ffn_1 = nn.Dense(units=hidden_size,
in_units=units,
flatten=False,
weight_initializer=weight_initializer,
bias_initializer=bias_initializer,
dtype=dtype)
self.activation = get_activation(activation)
self.ffn_2 = nn.Dense(units=units,
in_units=hidden_size,
flatten=False,
After Change
bias_initializer=bias_initializer,
dtype=dtype)
self.activation = get_activation(activation)
self.ffn_2 = nn.Dense(units=units,
in_units=hidden_size,
flatten=False,
use_bias=use_bias,
weight_initializer=weight_initializer,
bias_initializer=bias_initializer,
dtype=dtype)
// TODO(sxjscience) We may need to set the dtype flag in LayerNorm, need to double check
self.layer_norm = get_norm_layer(in_channels=units,
normalization=normalization,
epsilon=layer_norm_eps,
In pattern: SUPERPATTERN
Frequency: 3
Non-data size: 2
Instances
Project Name: dmlc/gluon-nlp
Commit Name: 36092e2ae66342ee1d7ad5678cd99c46090373eb
Time: 2021-01-18
Author: wyy@cmu.edu
File Name: src/gluonnlp/layers.py
Class Name: PositionwiseFFN
Method Name: __init__
Project Name: dmlc/gluon-nlp
Commit Name: 36092e2ae66342ee1d7ad5678cd99c46090373eb
Time: 2021-01-18
Author: wyy@cmu.edu
File Name: src/gluonnlp/layers.py
Class Name: PositionwiseFFN
Method Name: __init__
Project Name: d2l-ai/d2l-zh
Commit Name: 406584a66f5b3887b0ea2b1c1cc6ac51b892a15a
Time: 2018-07-17
Author: asv325@gmail.com
File Name: gluonbook/utils.py
Class Name:
Method Name: resnet18