博主
https://www.jianshu.com/p/673fd07954e9
tf.contrib.layers.fully_connected(F, num_outputs,activation_fn)
F ---[batch_size,images_pixels],tensor
num_outputs --- numbers of outputs,[batch_size,num_outputs]
activation_fn ---采用指定的非线性激励函数,默认不是None,如果不需要的话,要赋值None
API解释
https://docs.w3cub.com/tensorflow~python/tf/contrib/layers/fully_connected/
tf.contrib.layers.fully_connected
tf.contrib.layers.fully_connected(
inputs,
num_outputs,
activation_fn=tf.nn.relu,
normalizer_fn=None,
normalizer_params=None,
weights_initializer=initializers.xavier_initializer(),
weights_regularizer=None,
biases_initializer=tf.zeros_initializer(),
biases_regularizer=None,
reuse=None,
variables_collections=None,
outputs_collections=None,
trainable=True,
scope=None
)
Defined in tensorflow/contrib/layers/python/layers/layers.py.
See the guide: Layers (contrib) > Higher level ops for building neural network layers
Adds a fully connected layer.
fully_connected creates a variable called weights, representing a fully connected weight matrix, which is multiplied by the inputs to produce a Tensor of hidden units. If a normalizer_fn is provided (such as batch_norm), it is then applied. Otherwise, if normalizer_fn is None and a biases_initializer is provided then a biases variable would be created and added the hidden units. Finally, if activation_fn is not None, it is applied to the hidden units as well.
Note: that ifinputshave a rank greater than 2, theninputsis flattened prior to the initial matrix multiply byweights.
Args:
inputs: A tensor of at least rank 2 and static value for the last dimension; i.e.[batch_size, depth],[None, None, None, channels].num_outputs: Integer or long, the number of output units in the layer.activation_fn: Activation function. The default value is a ReLU function. Explicitly set it to None to skip it and maintain a linear activation.normalizer_fn: Normalization function to use instead ofbiases. Ifnormalizer_fnis provided thenbiases_initializerandbiases_regularizerare ignored andbiasesare not created nor added. default set to None for no normalizer functionnormalizer_params: Normalization function parameters.weights_initializer: An initializer for the weights.weights_regularizer: Optional regularizer for the weights.biases_initializer: An initializer for the biases. If None skip biases.biases_regularizer: Optional regularizer for the biases.reuse: Whether or not the layer and its variables should be reused. To be able to reuse the layer scope must be given.variables_collections: Optional list of collections for all the variables or a dictionary containing a different list of collections per variable.outputs_collections: Collection to add the outputs.trainable: IfTruealso add variables to the graph collectionGraphKeys.TRAINABLE_VARIABLES(see tf.Variable).scope: Optional scope for variable_scope.
Returns:
The tensor variable representing the result of the series of operations.
Raises:
ValueError: If x has rank less than 2 or if its last dimension is not set.
2290

被折叠的 条评论
为什么被折叠?



