Python keras.activations.linear() Examples
The following are 24
code examples of keras.activations.linear().
You can vote up the ones you like or vote down the ones you don't like,
and go to the original project or source file by following the links above each example.
You may also want to check out all available functions/classes of the module
keras.activations
, or try the search function
.
Example #1
Source File: keras_mnist_vis.py From keras-mnist-workshop with Apache License 2.0 | 6 votes |
def keras_digits_vis(model, X_test, y_test): layer_idx = utils.find_layer_idx(model, 'preds') model.layers[layer_idx].activation = activations.linear model = utils.apply_modifications(model) for class_idx in np.arange(10): indices = np.where(y_test[:, class_idx] == 1.)[0] idx = indices[0] f, ax = plt.subplots(1, 4) ax[0].imshow(X_test[idx][..., 0]) for i, modifier in enumerate([None, 'guided', 'relu']): heatmap = visualize_saliency(model, layer_idx, filter_indices=class_idx, seed_input=X_test[idx], backprop_modifier=modifier) if modifier is None: modifier = 'vanilla' ax[i+1].set_title(modifier) ax[i+1].imshow(heatmap) plt.imshow(heatmap) plt.show()
Example #2
Source File: activations_test.py From DeepLearning_Wavelet-LSTM with MIT License | 6 votes |
def test_get_fn(): """Activations has a convenience "get" function. All paths of this function are tested here, although the behaviour in some instances seems potentially surprising (e.g. situation 3) """ # 1. Default returns linear a = activations.get(None) assert a == activations.linear # 2. Passing in a layer raises a warning layer = Dense(32) with pytest.warns(UserWarning): a = activations.get(layer) # 3. Callables return themselves for some reason a = activations.get(lambda x: 5) assert a(None) == 5 # 4. Anything else is not a valid argument with pytest.raises(ValueError): a = activations.get(6)
Example #3
Source File: activations_test.py From DeepLearning_Wavelet-LSTM with MIT License | 6 votes |
def test_get_fn(): """Activations has a convenience "get" function. All paths of this function are tested here, although the behaviour in some instances seems potentially surprising (e.g. situation 3) """ # 1. Default returns linear a = activations.get(None) assert a == activations.linear # 2. Passing in a layer raises a warning layer = Dense(32) with pytest.warns(UserWarning): a = activations.get(layer) # 3. Callables return themselves for some reason a = activations.get(lambda x: 5) assert a(None) == 5 # 4. Anything else is not a valid argument with pytest.raises(ValueError): a = activations.get(6)
Example #4
Source File: activations_test.py From DeepLearning_Wavelet-LSTM with MIT License | 6 votes |
def test_get_fn(): """Activations has a convenience "get" function. All paths of this function are tested here, although the behaviour in some instances seems potentially surprising (e.g. situation 3) """ # 1. Default returns linear a = activations.get(None) assert a == activations.linear # 2. Passing in a layer raises a warning layer = Dense(32) with pytest.warns(UserWarning): a = activations.get(layer) # 3. Callables return themselves for some reason a = activations.get(lambda x: 5) assert a(None) == 5 # 4. Anything else is not a valid argument with pytest.raises(ValueError): a = activations.get(6)
Example #5
Source File: activations_test.py From DeepLearning_Wavelet-LSTM with MIT License | 6 votes |
def test_get_fn(): """Activations has a convenience "get" function. All paths of this function are tested here, although the behaviour in some instances seems potentially surprising (e.g. situation 3) """ # 1. Default returns linear a = activations.get(None) assert a == activations.linear # 2. Passing in a layer raises a warning layer = Dense(32) with pytest.warns(UserWarning): a = activations.get(layer) # 3. Callables return themselves for some reason a = activations.get(lambda x: 5) assert a(None) == 5 # 4. Anything else is not a valid argument with pytest.raises(ValueError): a = activations.get(6)
Example #6
Source File: activations_test.py From DeepLearning_Wavelet-LSTM with MIT License | 6 votes |
def test_get_fn(): """Activations has a convenience "get" function. All paths of this function are tested here, although the behaviour in some instances seems potentially surprising (e.g. situation 3) """ # 1. Default returns linear a = activations.get(None) assert a == activations.linear # 2. Passing in a layer raises a warning layer = Dense(32) with pytest.warns(UserWarning): a = activations.get(layer) # 3. Callables return themselves for some reason a = activations.get(lambda x: 5) assert a(None) == 5 # 4. Anything else is not a valid argument with pytest.raises(ValueError): a = activations.get(6)
Example #7
Source File: activations_test.py From DeepLearning_Wavelet-LSTM with MIT License | 6 votes |
def test_get_fn(): """Activations has a convenience "get" function. All paths of this function are tested here, although the behaviour in some instances seems potentially surprising (e.g. situation 3) """ # 1. Default returns linear a = activations.get(None) assert a == activations.linear # 2. Passing in a layer raises a warning layer = Dense(32) with pytest.warns(UserWarning): a = activations.get(layer) # 3. Callables return themselves for some reason a = activations.get(lambda x: 5) assert a(None) == 5 # 4. Anything else is not a valid argument with pytest.raises(ValueError): a = activations.get(6)
Example #8
Source File: activations_test.py From DeepLearning_Wavelet-LSTM with MIT License | 6 votes |
def test_get_fn(): """Activations has a convenience "get" function. All paths of this function are tested here, although the behaviour in some instances seems potentially surprising (e.g. situation 3) """ # 1. Default returns linear a = activations.get(None) assert a == activations.linear # 2. Passing in a layer raises a warning layer = Dense(32) with pytest.warns(UserWarning): a = activations.get(layer) # 3. Callables return themselves for some reason a = activations.get(lambda x: 5) assert a(None) == 5 # 4. Anything else is not a valid argument with pytest.raises(ValueError): a = activations.get(6)
Example #9
Source File: activations_test.py From DeepLearning_Wavelet-LSTM with MIT License | 6 votes |
def test_get_fn(): """Activations has a convenience "get" function. All paths of this function are tested here, although the behaviour in some instances seems potentially surprising (e.g. situation 3) """ # 1. Default returns linear a = activations.get(None) assert a == activations.linear # 2. Passing in a layer raises a warning layer = Dense(32) with pytest.warns(UserWarning): a = activations.get(layer) # 3. Callables return themselves for some reason a = activations.get(lambda x: 5) assert a(None) == 5 # 4. Anything else is not a valid argument with pytest.raises(ValueError): a = activations.get(6)
Example #10
Source File: activations_test.py From DeepLearning_Wavelet-LSTM with MIT License | 5 votes |
def test_serialization(): all_activations = ['softmax', 'relu', 'elu', 'tanh', 'sigmoid', 'hard_sigmoid', 'linear', 'softplus', 'softsign', 'selu'] for name in all_activations: fn = activations.get(name) ref_fn = getattr(activations, name) assert fn == ref_fn config = activations.serialize(fn) fn = activations.deserialize(config) assert fn == ref_fn
Example #11
Source File: similarity_layer.py From bidaf-keras with GNU General Public License v3.0 | 5 votes |
def compute_similarity(self, repeated_context_vectors, repeated_query_vectors): element_wise_multiply = repeated_context_vectors * repeated_query_vectors concatenated_tensor = K.concatenate( [repeated_context_vectors, repeated_query_vectors, element_wise_multiply], axis=-1) dot_product = K.squeeze(K.dot(concatenated_tensor, self.kernel), axis=-1) return linear(dot_product + self.bias)
Example #12
Source File: activations_test.py From DeepLearning_Wavelet-LSTM with MIT License | 5 votes |
def test_serialization(): all_activations = ['softmax', 'relu', 'elu', 'tanh', 'sigmoid', 'hard_sigmoid', 'linear', 'softplus', 'softsign', 'selu'] for name in all_activations: fn = activations.get(name) ref_fn = getattr(activations, name) assert fn == ref_fn config = activations.serialize(fn) fn = activations.deserialize(config) assert fn == ref_fn
Example #13
Source File: activations_test.py From DeepLearning_Wavelet-LSTM with MIT License | 5 votes |
def test_linear(): xs = [1, 5, True, None] for x in xs: assert(x == activations.linear(x))
Example #14
Source File: activations_test.py From DeepLearning_Wavelet-LSTM with MIT License | 5 votes |
def test_serialization(): all_activations = ['softmax', 'relu', 'elu', 'tanh', 'sigmoid', 'hard_sigmoid', 'linear', 'softplus', 'softsign', 'selu'] for name in all_activations: fn = activations.get(name) ref_fn = getattr(activations, name) assert fn == ref_fn config = activations.serialize(fn) fn = activations.deserialize(config) assert fn == ref_fn
Example #15
Source File: activations_test.py From DeepLearning_Wavelet-LSTM with MIT License | 5 votes |
def test_serialization(): all_activations = ['softmax', 'relu', 'elu', 'tanh', 'sigmoid', 'hard_sigmoid', 'linear', 'softplus', 'softsign', 'selu'] for name in all_activations: fn = activations.get(name) ref_fn = getattr(activations, name) assert fn == ref_fn config = activations.serialize(fn) fn = activations.deserialize(config) assert fn == ref_fn
Example #16
Source File: activations_test.py From DeepLearning_Wavelet-LSTM with MIT License | 5 votes |
def test_linear(): xs = [1, 5, True, None] for x in xs: assert(x == activations.linear(x))
Example #17
Source File: test_activations.py From CAPTCHA-breaking with MIT License | 5 votes |
def test_linear(): ''' This function does no input validation, it just returns the thing that was passed in. ''' from keras.activations import linear as l xs = [1, 5, True, None, 'foo'] for x in xs: assert x == l(x)
Example #18
Source File: activations_test.py From DeepLearning_Wavelet-LSTM with MIT License | 5 votes |
def test_serialization(): all_activations = ['softmax', 'relu', 'elu', 'tanh', 'sigmoid', 'hard_sigmoid', 'linear', 'softplus', 'softsign', 'selu'] for name in all_activations: fn = activations.get(name) ref_fn = getattr(activations, name) assert fn == ref_fn config = activations.serialize(fn) fn = activations.deserialize(config) assert fn == ref_fn
Example #19
Source File: activations_test.py From DeepLearning_Wavelet-LSTM with MIT License | 5 votes |
def test_linear(): xs = [1, 5, True, None] for x in xs: assert(x == activations.linear(x))
Example #20
Source File: activations_test.py From DeepLearning_Wavelet-LSTM with MIT License | 5 votes |
def test_serialization(): all_activations = ['softmax', 'relu', 'elu', 'tanh', 'sigmoid', 'hard_sigmoid', 'linear', 'softplus', 'softsign', 'selu'] for name in all_activations: fn = activations.get(name) ref_fn = getattr(activations, name) assert fn == ref_fn config = activations.serialize(fn) fn = activations.deserialize(config) assert fn == ref_fn
Example #21
Source File: activations_test.py From DeepLearning_Wavelet-LSTM with MIT License | 5 votes |
def test_serialization(): all_activations = ['softmax', 'relu', 'elu', 'tanh', 'sigmoid', 'hard_sigmoid', 'linear', 'softplus', 'softsign', 'selu'] for name in all_activations: fn = activations.get(name) ref_fn = getattr(activations, name) assert fn == ref_fn config = activations.serialize(fn) fn = activations.deserialize(config) assert fn == ref_fn
Example #22
Source File: activations_test.py From DeepLearning_Wavelet-LSTM with MIT License | 5 votes |
def test_linear(): xs = [1, 5, True, None] for x in xs: assert(x == activations.linear(x))
Example #23
Source File: activations_test.py From DeepLearning_Wavelet-LSTM with MIT License | 5 votes |
def test_serialization(): all_activations = ['softmax', 'relu', 'elu', 'tanh', 'sigmoid', 'hard_sigmoid', 'linear', 'softplus', 'softsign', 'selu'] for name in all_activations: fn = activations.get(name) ref_fn = getattr(activations, name) assert fn == ref_fn config = activations.serialize(fn) fn = activations.deserialize(config) assert fn == ref_fn
Example #24
Source File: model.py From Keras-progressive_growing_of_gans with MIT License | 4 votes |
def Generator( num_channels =1, resolution =32, label_size =0, fmap_base =4096, fmap_decay =1.0, fmap_max =256, latent_size =None, normalize_latents =True, use_wscale =True, use_pixelnorm =True, use_leakyrelu =True, use_batchnorm =False, tanh_at_end =None, **kwargs): R = int(np.log2(resolution)) assert resolution == 2 ** R and resolution >= 4 cur_lod = K.variable(np.float32(0.0), dtype='float32', name='cur_lod') def numf(stage): return min(int(fmap_base / (2.0 ** (stage * fmap_decay))), fmap_max) if latent_size is None: latent_size = numf(0) (act, act_init) = (lrelu, lrelu_init) if use_leakyrelu else (relu, relu_init) inputs = [Input(shape=[latent_size], name='Glatents')] net = inputs[-1] #print("DEEEEEEEE") if normalize_latents: net = PixelNormLayer(name='Gnorm')(net) if label_size: inputs += [Input(shape=[label_size], name='Glabels')] net = Concatenate(name='G1na')([net, inputs[-1]]) net = Reshape((1, 1,K.int_shape(net)[1]), name='G1nb')(net) net = G_convblock(net, numf(1), 4, act, act_init, pad='full', use_wscale=use_wscale, use_batchnorm=use_batchnorm, use_pixelnorm=use_pixelnorm, name='G1a') net = G_convblock(net, numf(1), 3, act, act_init, pad=1, use_wscale=use_wscale, use_batchnorm=use_batchnorm, use_pixelnorm=use_pixelnorm, name='G1b') lods = [net] for I in range(2, R): net = UpSampling2D(2, name='G%dup' % I)(net) net = G_convblock(net, numf(I), 3, act, act_init, pad=1, use_wscale=use_wscale, use_batchnorm=use_batchnorm, use_pixelnorm=use_pixelnorm, name='G%da' % I) net = G_convblock(net, numf(I), 3, act, act_init, pad=1, use_wscale=use_wscale, use_batchnorm=use_batchnorm, use_pixelnorm=use_pixelnorm, name='G%db' % I) lods += [net] lods = [NINblock(l, num_channels, linear, linear_init, use_wscale=use_wscale, name='Glod%d' % i) for i, l in enumerate(reversed(lods))] output = LODSelectLayer(cur_lod, name='Glod')(lods) if tanh_at_end is not None: output = Activation('tanh', name='Gtanh')(output) if tanh_at_end != 1.0: output = Lambda(lambda x: x * tanh_at_end, name='Gtanhs') model = Model(inputs=inputs, outputs=[output]) model.cur_lod = cur_lod return model