Please enable JavaScript.
Coggle requires JavaScript to display documents.
Deep Learning, architecture - Coggle Diagram
Deep Learning
Activation Function
-
Sigmoid Function
-
-
-
monotonic, continuous, and easy to derive.
-
-
-
-
-
-
Softplus
-
-
-
has a continuous derivative, and a relatively smooth defined surface
-
single layer perceptron
-
-
Ketika W^T X <= 0, maka O = -1 sample diklasifikasikan ke dalam other type
Ketika W^T X > 0, maka O = 1 sample diklasifikasikan ke dalam type
-
deep neural network
-
neural network
information processing system designed to imitate the human brain structure and functions based on its source, feature, and explanations.
-
-
-
Type
CNN
-
pooling layer
combines nearby units, reduces a size of a feature map, and reduces dimensions
Common pooling layers
max pooling layer
divides a feature map into several regions and uses the maximum value of each region to represent the entire region
average pooling layer
divides a feature map into several regions and uses the average value of each region to represent the entire region
-
-
does not involve arrangement of elements in each small region, and concerns only statistical features of these elements
-
-
-
-
-
Training Rule
Gradient Decent
-
Enables the loss function to search along the negative gradient direction and update the parameters iteratively
-
-
-
-
-
Optimizer
-
-
Common optimizers
momentum optimizer
-
-
-
Advantages
-
where the gradient direction is stable, accelerates convergence
-
Disadvantages
The learning rate π and momentum πΌ need to be manually set, requires more experiments to
determine the appropriate value.
AdaGrad Optimizer
-
π indicates the global LR, which needs to be set manually
-
π continues increasing while the overall learning rate keeps decreasing as the algorithm iterates.
Pros
The learning rate is automatically updated. As the number of updates increases, the learning rate decreases.
Cons
The denominator keeps accumulating so that the learning rate will eventually become very small, and the algorithm will become ineffective.
initial value of π is 0, which
increases continuously
RMSProp Optimizer
-
-
the initial value of π is 0, which may not increase and needs to be adjusted using a parameter
π indicates the global LR, which needs to be set manually
-
-
-
-