Fine-tuning Stacked AEs
From Ufldl
for
Fine-tuning Stacked AEs
Jump to:
navigation
,
search
=== Introduction === Fine tuning is a strategy that is commonly used to reduce the run time of a stacked autoencoder. It involves viewing all layers of a stacked autoencoder as a single model. === Strategy === Conceptually, fine tuning is quite simple. In order to view all layers of a stacked autoencoder as a single model, the gradients at each step are computed using the [[Backpropagation Algorithm]], as discussed in the sparse autoencoder section.
Template:CNN
(
view source
)
Template:Languages
(
view source
)
Template:Quote
(
view source
)
Return to
Fine-tuning Stacked AEs
.
Views
Page
Discussion
View source
History
Personal tools
Log in
ufldl resources
UFLDL Tutorial
Recommended Readings
wiki
Main page
Recent changes
Random page
Help
Search
Toolbox
What links here
Related changes
Special pages