# [Paper筆記] : A Survey on Transfer Learning (2010)
### [Paper筆記] : A Survey on Transfer Learning (2010)
### **<Introduction>**
--------------------------------------
Common machine learning algorithms work under an assumption:
> **Training and Testing data are drawn from the same feature space and have the same distribution**
Therefore, when the distribution changes, models need to be rebuild from scratch using new data. **There's no knowledge transfer between tasks.**.
In reality, there are many scenarios where such assumption can be problematic when modeling.
> * **Limited data for our target task**
> * There may exist similar tasks with abundant dataset and more sophisticated models.
> * However, by traditional methodology, we can't harvest the knowledge from these similar tasks.
> * **The data may be easily outdated due to dynamic environments**
> * Models need to adapt quickly, while based on existing knowledge.
> * (ex) Wifi data to infer location.
---------------------------------------
Focus of this paper:
* Definition of terms and topics in transfer learning
* Survey transfer learning techniques for classification, regression, and clustering.
--------------------------------------------
### **<Overview>**
-------------------------------------------------
#### Brief History of Transfer Learning
------------------------------------------------
Many techniques have been proposed to address the problem of not enough labeled data.
For example, **semi-supervised learning** and **active learning** are two of them.
> - **Semi-Superised Learning:**
> - Large amount of unlabeled data + Small amount of labeled data
> - **Active Learning:**
> - Select the next candidate data to be labeled, for cost-effective concern.
However, most of these techniques assume the distributions between labeled and unlabeled data are the same.
Transfer Learning, in constrast, allows the domains/tasks/distributions in training and testing data to be different.
------------------------------------------------
A closely-related technique to transfer learning is **multi-task learning**.
The difference lies in the asymmetric role of source and target tasks in transfer learning:
- Transfer Learning:
- Extract knowledge from one or more source tasks, and apply the knowledge to a target task.
- Multi-Task Learning:
- Learn multiple tasks simultenously
Transfer Learning cares most about the target task.
-------------------------------------------------
#### Notations and Definitions
-------------------------------------------------