Useful Papers
===
[TOC]
# in urgent
- Few Shot Network Compression via Cross Distillation
- [code-official (PyTorch)](https://github.com/haolibai/Cross-Distillation)
- 改造 layer-wise distillation loss
- 變成 correction loss, imitation loss
- Adversarial Complementary Learning for Weakly Supervised Object Localization. CVPR 2018
- 互補的 classifier
- Borrowing treasures from the wealthy: Deep transfer learning through selective joint fine- tuning. In CVPR, 2017.
- Learning Deep Feature Representations with Domain Guided Dropout for Person Re-identification. CVPR 2016
- [中文](https://blog.csdn.net/He_is_all/article/details/55522302)
- Addressing Model Vulnerability to Distributional Shifts over Image Transformation Sets. ICCV 2019
- define new **data augmentation** rules according to the image transformations that the **current model is most vulnerable to**, over iterations. (in terms of **N−tuples of image transformations**.)
- show that **random search** and, in particular, **evolution-based search** are effective approaches to face this problem.
- Understanding how feature structure transfers in transfer learning. 2017, IJCAI
- 本篇 paper 有助於理解 feature 如何 transfer
- 這篇似乎比較理論性質,有空再看
- justified that feature structure can be transferred, **independently of the change of $P_{y|x}$** over domains
- discussed how feature structure can be transferred in **domain adaptation** and **learning to learn** settings from a regularization perspective.
- implies that a **tuning parameter is necessary** to help transfer feature structure information from the source domain to the target domain.
- 應該是指 fine-tune 是必須的?
- Recent Advances in Transfer Learning for Cross-Dataset Visual Recognition: A Problem-Oriented Perspective. 2017
- https://arxiv.org/pdf/1705.04396.pdf
- Knowledge Transfer in Vision Recognition: A Survey
- https://hal.archives-ouvertes.fr/hal-02101005/document
- Characterizing and Avoiding Negative Transfer. CVPR 2019
- This paper proposes a formal definition of negative transfer and analyzes three important aspects thereof. Stemming from this analysis, a novel technique is proposed to circumvent negative transfer by **filtering out unrelated source data**. Based on **adversarial networks**, the technique is highly generic and can **be applied to a wide range of transfer learning algorithms**.
- Disjoint Label Space Transfer Learning with Common Factorised Space. AAAI'19
- Generalizing to unseen domains via adversarial data augmentation. NeurIPS 2018
- Instance Normalization
- 好像就是把單個 sample、單一 channel 內的所有 pixel 做 normalize
- Adaptive Instance Normalization
- Adaptive batch normalization for practical domain adaptation. ICLR 2017 Workshop, *Pattern Recognition* 2018
- **AdaBN**
- 跟這篇是同一篇 Revisiting batch normalization for practical domain adaptation. arXiv 16
- [中文](https://zhuanlan.zhihu.com/p/56162416)
- [my note](https://hackmd.io/@johnnyasd12/ry7xlIA64)
- the **label related** knowledge is stored in the **weight matrix** of each layer, whereas **domain related** knowledge is represented by the **statistics of the Batch Normalization** (BN) layer
# Domain & Label Shift
- Decaf: A deep convolutional activation feature for generic visual recognition. ICML 2014
- Label Efficient Learning of Transferable Representations across Domains and Tasks. NIPS 2017 (Li Fei-Fei)
- Our method shows compelling results on **novel classes within a new domain** even when **only a few labeled examples per class** are available, outperforming the prevalent fine-tuning approach.
- 請移駕到 [Awesome Few-shot/Meta Learning](https://hackmd.io/2e6l5BmYS2ebhE__dwOcgQ?both#Label-Efficient-Learning-of-Transferable-Representations-across-Domains-and-Tasks-NIPS-2017-Li-Fei-Fei)
- Learning to cluster in order to transfer across domains and tasks. ICLR 2018
- **cross domain + label**- This paper introduces a novel method to perform transfer learning across domains and tasks, formulating it as a problem of **learning to cluster**.
- The key insight is that, in addition to features, we can transfer **similarity information** and this is sufficient to learn a **similarity function** and **clustering network** to perform both domain adaptation and cross-task transfer learning.
- We begin by reducing categorical information to **pairwise** constraints, which only considers whether two instances **belong to the same class or not (similarity network)**
- We then present two novel approaches for performing transfer learning **using this similarity function**
1. for unsupervised domain adaptation, we design a new loss function to **regularize classification** with a **constrained clustering loss**, hence learning a **clustering network** with the transferred similarity metric **generating the training inputs(??)**.
2. for cross-task learning (i.e., unsupervised clustering with unseen categories), we propose a framework to reconstruct and estimate the number of semantic clusters, **again** using the **clustering network**.
- Since the similarity network is noisy, the key is to use a **robust clustering** algorithm,
- Our results show that we can reconstruct semantic clusters with high accuracy.
- Our approach doesn’t explicitly deal with domain discrepancy. **If we combine with a domain adaptation loss, it shows further improvement**.
- Deep Cocktail Network: Multi-source Unsupervised Domain Adaptation with Category Shift. CVPR 2018
- **cross domain + label**
- **multi-source domain**
- Split Batch Normalization: Improving Semi-Supervised Learning under Domain Shift. ICLR 2019 Workshop LLD
- Recent work has shown that using **unlabeled data** in semi-supervised learning is **not always beneficial** and can even hurt generalization, especially when there is a class mismatch between the unlabeled and labeled examples.
- Our main contribution is showing how to benefit from additional **unlabeled data that comes from a shifted distribution** in **batch-normalized** neural networks.
- We achieve it by simply using **separate batch normalization statistics for unlabeled examples**.
- Transfer Learning via Learning to Transfer. ICML 2018
- 這篇有點難
- Partial Adversarial Domain Adaptation. ECCV 2018
- 想參考在 source 中和 target 無關的 class 是怎處理的
# Remove Bias in Dataset
- Unbiased look at dataset bias. CVPR 2011
- REPAIR: Removing Representation Bias by Dataset Resampling. CVPR'19
- [code - official(PyTorch)](https://github.com/JerryYLi/Dataset-REPAIR/)
# Useful
- Making Convolutional Networks Shift-Invariant Again. ICML 2019
- SpotTune: Transfer Learning through Adaptive Fine-tuning. CVPR'19
- Multi-class Classification without Multi-class Labels. ICLR'19
- pairwise similarity
- M-ADDA: Unsupervised Domain Adaptation with Deep Metric Learning. arXiv 1807
- Improving generalization via scalable neighborhood component analysis. ECCV 2018
- supervised domain adaptation
- Cross-entropy adversarial view adaptation for person re-identification. IEEE Transactions on Circuits and Systems for Video Technology
- Learning what you can do before doing anything. ICLR'19
- Learning to remember more with less memorization. ICLR'19
- Unsupervised Domain Adaptation for Distance Metric Learning. ICLR'19
- Divide and Conquer the Embedding Space for Metric Learning. CVPR'19
- AutoAugment: Learning Augmentation Strategies from Data. CVPR'19
- AutoDIAL: Automatic Domain Alignment Layers. ICCV 2017
# Conditional VAE
- Learning structured output representation using deep conditional generative models. NIPS 2015
# VAE-GAN
- Adversarial feature learning. 2016
- Autoencoding beyond pixels using a learned similarity metric. ICML 2016
- Adversarial variational bayes: Unifying variational autoencoders and generative adversarial networks. ICML 2017
- Cvae-gan: Fine-grained image generation through asymmetric training. ICCV 2017
- conditional VAE-GAN
# Disentangled Feature Learning
- CVAE-GAN: Fine-Grained Image Generation through Asymmetric Training. ICCV 2017
- InfoGAN: Interpretable representation learning by information maximizing generative adversarial nets. NIPS 2016
- Conditional image synthesis with auxiliary classifier GANs. ICML 2017
- AC-GAN
- Independently Controllable Factors. arXiv 1708
- Multi-Task Adversarial Network for Disentangled Feature Learning. CVPR 2018
- 训练出一个网络能提取出只与任务相关的Feature,从而提升在此网络上针对改任务的泛化性和可迁移性
- Multi-Level Variational Autoencoder: Learning Disentangled Representations from Grouped Observations. AAAI 2018
- Domain Agnostic Learning with Disentangled Representations. ICML'19
- single source & multi-target domain adaptation

- beta-VAE: Learning Basic Visual Concepts with a Constrained Variational Framework. ICLR 2017
- Isolating Sources of Disentanglement in Variational Autoencoders. NeurIPS 2018
- A two-step disentanglement method. CVPR 2018
- Towards open- set identity preserving face synthesis. CVPR 2018
- requires at least two inputs for training
- Neural face editing with intrinsic image disentangling. CVPR 2017
- Disentangling by Factorising. ICML 2018
- factor-VAE
- [code - official (PyTorch)](https://github.com/1Konny/FactorVAE)
- Disentangling Latent Space for VAE by Label Relevant/Irrelevant Dimensions. CVPR 2019
- [code (under-developed?) - official (TF)](https://github.com/ZhilZheng/Lr-LiVAE)
- different from cVAE, we present a method for disentangling the latent space into the **label relevant** and **irrelevant** dimensions $z_s$ and $z_u$
- $z_u$ represent the **common characteristics** of all inputs, hence they are constrained by the **standard Gaussian**
- $z_s$ is assumed to follow the **Gaussian mixture distribution** in which each component corresponds to a particular class.
# Open-set Domain Adaptation
- Universal Domain Adaptation. CVPR 2019
- Open Set Domain Adaptation. ICCV 2017
- [中文](https://zhuanlan.zhihu.com/p/31230331)
- [code - official (MatLab)](https://github.com/Heliot7/open-set-da)
- Learning Factorized Representations for Open-set Domain Adaptation. ICLR'19
# Partial Domain Adaptation / Partial Transfer Learning
- $Y_T\subset Y_S$
- Partial Adversarial Domain Adaptation. ECCV 2018
- based on DANN
- (1) Mitigate negative transfer by filtering out unrelated source labeled
data belonging to the outlier label space $C_s\ C_t$.
減輕 source labeled data 在 outlier label 的影響,等於減輕 negative transfer 的結果;
- (2) Promote positive transfer by maximally matching the data distributions $p_{Ct}$ and $q$ in the shared label space $C_t$.
另一方面,減少 target 與 source 共同 label 分布之差異(positive transfer)。
- predict target domain data by source classifier and take mean value of the output probability
- $\gamma = \dfrac{1}{n_t}\sum_\limits{i=1}^{n_t}\hat y_i$
- then
- original DANN: 
- proposed PADA: 
- $y_i$ 是 source data $x_i$ 的 ground truth;$\gamma_{y_i}$ 是相應的 class weight
- Learning to Transfer Examples for Partial Domain Adaptation. CVPR'19
- Partial transfer learning with selective adversarial networks. 2017
# Explainable / visualizing NN
# Domain Generalization
- Generalizing Across Domains via Cross-Gradient Training. ICLR 2018
- Episodic Training for Domain Generalization. ICCV'19 (oral)
- Metareg: Towards domain generalization using meta-regularization. NeurIPS 2018
- seems to be the same label space between source and target, focus on **domain generalization** problem.
# Domain Adaptation
- Adversarial Dropout Regularization. ICLR 2018
- 待讀
- Bridging Theory and Algorithm for Domain Adaptation. ICML'19
- Boosting for transfer learning. ICML 2007. (著名的 TrAdaboost)
- [Awesome domain adaptation](https://github.com/zhaoxin94/awsome-domain-adaptation)
- Domain Adaptation with Asymmetrically-Relaxed Distribution Alignment. ICLR 2019 Workshop LLD
- **only cross domain & label "distribution"**
- seems not for label space shift
- **Recently-proposed domain-adversarial approaches** consist of aligning source and target encodings, they can **break down under shifting label distributions**.
- contributions
- We propose **asymmetrically-relaxed distribution alignment**, a **relaxed distance** for aligning data across domains that can be minimized **without requiring latent-space distributions to match exactly**.
- We propose **several distances** that satisfy the desired properties and are **optimizable by adversarial training**.
- Regularized Learning for Domain Adaptation under Label Shifts. ICLR'19
- seems still **only cross domain + label "distribution"**
- Transferable meta learning across domains. UAI 2018
- 似乎用到 target domain 的 unlabeled data
- MAML + DANN?
- 這篇也跟樓上一樣不是真的在做 few-shot,**source 跟 target 是相同 label space**
- [my paper note](https://hackmd.io/yh6uPnEwQzOfuvYOynB06Q)
- Meta-learning algorithms require **sufficient tasks** for meta model training and resulted model can **only solve new similar tasks**.
- to address these two problems, we propose a new **transferable meta learning (TML)** algorithm
- Bidirectional One-Shot Unsupervised Domain Mapping. ICCV 2019
- one encoder and one decoder for **each domain**
- domain $A$: single training sample (per class???)
- domain $B$: richer training set
- For example, we can transfer all MNIST images to the visual domain captured by a single SVHN image and transform the SVHN image to the domain of the MNIST images.
- seems a one-shot domain adaptation problem...
- A DIRT-T Approach to Unsupervised Domain Adaptation. ICLR 2018
- Virtual Adversarial Domain Adaptation (VADA)
- using Virtual Adversarial Training (VAT)
- Taking A Closer Look at Domain Shift: Category-level Adversaries for Semantics Consistent Domain Adaptation. CVPR 2019 Oral
- Consensus Adversarial Domain Adaptation. AAAI 19
- few-shot domain adaptation scheme (F-CADA)
- just **domain adaptation with few labeled data**
- gives freedom to both **target encoder** and **source encoder** to get domain-invariant features.
- **After obtaining** a **source encoder** and a **source classifier** as a good reference in the source domain, CADA **trains a target encoder** and also gives freedom to the **source encoder by fine-tuning** it through **adversarial learning**.
- 
- i think
- it's not so many novelty...
- the experiments are not compared to SOTA
- Few-shot adversarial domain adaptation. NIPS 2017
- [中文](https://blog.csdn.net/Adupanfei/article/details/85164925)
- [中文2](https://www.twblogs.net/a/5c1f39d2bd9eee16b3da81c0)
- [code (PyTorch)](https://github.com/Coolnesss/fada-pytorch)
- [My paper note](https://hackmd.io/8H_J9XauQgWrGfkLz88dKQ?view)
- supervised domain adaptation
- 並不真的 focus 在 few-shot learning
<!--
- 先在 source target 上 pre-train 一個 variational auto-encoder(VAE),複製給 target task。兩個 task share 一些 layer,target task 只能 update task-specific layer;source task 可以 update shared 跟 他自己的 task-specific layer
-->
- Transferrable Prototypical Networks for Unsupervised Domain Adaptation. CVPR'19
- 單純做 domain adaptation 而不是 few-shot,只是方法借用 few-shot 的 ProtoNet
- Maximum classifier discrepancy for unsupervised domain adaptation. CVPR 2018
- 找到 classifier 預測結果的 domain 差異上下界,使之越來越小
- [中文](https://zhuanlan.zhihu.com/p/57083034)
- [code - official (PyTorch)](https://github.com/mil-tokyo/MCD_DA)
- Transferable Attention for Domain Adaptation. AAAI 2019
- [中文笔记](https://zhuanlan.zhihu.com/p/52591143)
- Deep coral: Correlation alignment for deep domain adaptation. ECCV 2016
- Simultaneous deep transfer across domains and tasks. ICCV 2015
- 說是 transfer tasks,其實 source 跟 target 是同 label space。**下標題的人很有當記者的潛力**
- 所謂 transfer across tasks 其實是指:利用 source 某類別 predicted probability 的平均得到類別間的 information,再對 target domain 的 labeled data 用 Knowledge Distillation 的 soft label loss 做 regularization
- Improving the Generalization of Adversarial Training with Domain Adaptation. ICLR'19
- Learning to generalize: Meta-learning for domain generalization. AAAI 18
- d-SNE: Domain Adaptation using Stochastic Neighborhood Embedding. CVPR'19
- supervised domain adaptation with few labeled data
- Learning What and Where to Transfer. ICML'19
- [code - official (PyTorch)](https://github.com/alinlab/L2T-ww)
- [中文](https://zhuanlan.zhihu.com/p/66130006)
- When Samples Are Strategically Selected. ICML'19
- 沒看懂 abstract 0.0
- Depthwise Convolution is All You Need for Learning Multiple Visual Domains. AAAI'19
- Domain Agnostic Real-Valued Specificity Prediction. AAAI'19
# Zero-shot Learning
- [Awesome Zero-shot Learning](https://github.com/chichilicious/awesome-zero-shot-learning)
- Zero-Shot Deep Domain Adaptation. ECCV 2018
- To the best of our knowledge, ZDDA is the first domain adaptation and sensor fusion method which **requires no task-relevant target-domain data**.
- Preserving Semantic Relations for Zero-Shot Learning. CVPR 2018
- Learning Class Prototypes via Structure Alignment for Zero-Shot Recognition. ECCV 2018
- Zero-Shot Task Transfer. CVPR 2019
# Datasets
- A Large-scale Attribute Dataset for Zero-shot Learning. 2018
- Meta-dataset: A dataset of datasets for learning to learn from few examples. arXiv 2019
- other datasets
1. Market-1501: 750类 每类平均17.2张图片
1. CUHK03: 1367类 每类平均9.6张图片
1. DukeMTMC-reID: 702类 每类平均23.5张图片
1. CUBird: 200类 每类平均29.97张图片
# Interesting
- Rethinking feature distribution for loss functions in image classification. CVPR 2018
- proposed Gaussian Mixture Cross Entropy
- Do better imagenet models transfer better? CVPR 2019
- Understanding Deep Learning Requires Rethinking Generalization. arXiv 1612
- Google
- [如何评价 ICLR 2017 中关于 Rethinking Generalization 的那篇文章?](https://www.zhihu.com/question/56151007)
- self-supervised learning
- Adaptive Softmax
- Accurate, large minibatch sgd: Training imagenet in 1 hour
- Squeeze-and-Excitation Networks. CVPR 2018
- Attention is all you need (Transformer)
- [中文](https://zhuanlan.zhihu.com/p/47282410)
- Image Transformer. ICML 2018
- Hidden Technical Debt in Machine Learning Systems
- zero shot imitation learning. ICLR 2018
- spectral normalization
- spectral clustering
- Wide-ResidualNet
- How Important is a Neuron. ICLR'19
- Like What You Like: Knowledge Distill via Neuron Selectivity Transfer. ICLR'19
- All You Need is a Few Shifts: Designing Efficient Convolutional Neural Networks for Image Classification. CVPR'19
- SiamRPN++: Evolution of Siamese Visual Tracking with Very Deep Networks. CVPR'19
- FML: Face Model Learning from Videos. CVPR'19
- (單目標追蹤)
- arithmetic in CNN
- bag of tricks for CNN
###### tags: `papers` `domain adaptation` `transfer learning`