JavaScript
x
6
1
model = Sequential()
2
3
model.add(Conv2D(256, (3, 3), input_shape=X.shape[1:]))
4
model.add(Activation('relu'))
5
model.add(MaxPooling2D(pool_size=(2, 2)))
6
How do I use Maxout instead of’relu’ for activation?
Advertisement
Answer
You can use tensorflow_addons.layers.Maxout
to add Maxout
Activation function
JavaScript
1
7
1
import tensorflow_addons as tfa
2
3
model = Sequential()
4
model.add(Conv2D(256, (3, 3), input_shape=X.shape[1:]))
5
model.add(tfa.layers.Maxout(256))
6
model.add(MaxPooling2D(pool_size=(2, 2)))
7
You can install tensorflow_addons by:
JavaScript
1
2
1
pip install tensorflow-addons
2