Editing
Transfer Learning
(section)
Jump to navigation
Jump to search
Warning:
You are not logged in. Your IP address will be publicly visible if you make any edits. If you
log in
or
create an account
, your edits will be attributed to your username, along with other benefits.
Anti-spam check. Do
not
fill this in!
== <span style="color: #FFFFFF;">Remembering</span> == * '''Pre-trained model''' β A model trained on a large, general dataset (e.g., ImageNet for vision, Wikipedia for NLP) whose weights serve as a starting point for a new task. * '''Fine-tuning''' β The process of continuing training of a pre-trained model on task-specific data, adjusting weights to specialize behavior. * '''Feature extraction''' β Using a pre-trained model as a fixed feature extractor: freeze all its weights and add a new trainable head for the new task. * '''Domain adaptation''' β Adapting a model trained on one data distribution (source domain) to perform well on a different but related distribution (target domain). * '''Source domain''' β The original domain on which the model was pre-trained. * '''Target domain''' β The new domain or task to which you are transferring knowledge. * '''Domain shift''' β The difference in statistical distribution between source and target domains. * '''Frozen layers''' β Layers of a pre-trained model whose weights are not updated during fine-tuning. * '''Trainable layers''' β Layers whose weights are updated during fine-tuning (typically the head and possibly last few blocks). * '''Head''' β The task-specific output layer(s) added on top of a pre-trained backbone for a new task. * '''Backbone''' β The main body of a pre-trained model (the feature extractor), as opposed to the task-specific head. * '''ImageNet''' β A 1.2-million-image classification dataset; models pre-trained on ImageNet are the standard starting point for most computer vision tasks. * '''BERT''' β A pre-trained transformer encoder; the standard starting point for many NLP fine-tuning tasks. * '''Domain-adaptive pre-training''' β Additional pre-training on in-domain unlabeled data before task-specific fine-tuning. * '''Zero-shot transfer''' β Applying a model trained on one task directly to a new task without any task-specific training (e.g., CLIP for zero-shot image classification). </div> <div style="background-color: #006400; color: #FFFFFF; padding: 20px; border-radius: 8px; margin-bottom: 15px;">
Summary:
Please note that all contributions to BloomWiki may be edited, altered, or removed by other contributors. If you do not want your writing to be edited mercilessly, then do not submit it here.
You are also promising us that you wrote this yourself, or copied it from a public domain or similar free resource (see
BloomWiki:Copyrights
for details).
Do not submit copyrighted work without permission!
Cancel
Editing help
(opens in new window)
Navigation menu
Personal tools
Not logged in
Talk
Contributions
Create account
Log in
Namespaces
Page
Discussion
English
Views
Read
Edit
View history
More
Search
Navigation
Main page
Recent changes
Random page
Help about MediaWiki
Tools
What links here
Related changes
Special pages
Page information