# Notes on "FixBi: Bridging Domain Spaces for Unsupervised Domain Adaptation" [[Paper]](https://arxiv.org/pdf/2011.09230.pdf)
###### tags: `notes` `unsupervised` `domain-adaptation`
Notes Author: [Rohit Lal](https://rohitlal.net/)
---
## Brief Outline
- propose a UDA method that effectively handles large domain discrepancies.
- introduce a fixed ratio-based mixup to augment multiple intermediate domains between the source and target domain.
## Introduction
- an augmented domain close to the source domain has more reliable label information, but it has a lower correlation with the target domain.
- apply self penalization, which penalizes its own model to improve performance through self-training.
- to prevent divergence of the augmented models generated in different domains, we propose a consistency regularization using an augmented domain with the same ratio of source and target samples.
## Methodology
FixBi algorithm
![](https://i.imgur.com/FcrYFgS.png)
### Fixed Ratio based Mixup
- Given a pair of input samples and their corresponding one-hot labels in the source and target domain: $(x_i^s, y_i^s)$ and $(x_i^t, \hat{y}_i^t)$, mixup settings are defined as follows:
![](https://i.imgur.com/IMN17CZ.png)
Note: $\hat{y}_i^t$ obtained from pseudo labels got from DANN
$\lambda \in \{\lambda_{sd}, \lambda_{td}\}$ such that $\lambda_{sd} + \lambda_{td}\ = 1$
- Taking advantage of the fixed ratio-based mixup, we construct two network models that act as bridges between the source and target domain
- The source-dominant model has strong supervision for the source domain but relatively weak supervision for the target domain. By contrast, the target-dominant model has strong target supervision but weak source supervision.
### Confidence based Learning
- one model teaches the other model using the positive pseudo-labels or teach itself using the negative pseudo-labels.
#### Bidirectional Matching with positive pseudo-labels
- when one network assigns the class probability of input above a certain threshold $\tau$ , we assume that this predicted label as a pseudo-label.
- Then we train the peer network to make its predictions match these positive pseudo-labels via a standard cross-entropy loss.
![](https://i.imgur.com/KCjN9dk.png)
#### Self-penalization with negative pseudo-labels.
- the negative pseudo-label indicates the most confident label (top-1 label) predicted by the network with a confidence lower than the threshold $\tau$
- Since the negative pseudo-label is unlikely to be a correct label, we need to increase the probability values of all other classes except for this negative pseudo-label. Therefore, we optimize the output probability corresponding to the negative pseudo-label to be close to zero.
![](https://i.imgur.com/mA88ZoH.png)
- $\tau$ changes adaptively by the sample mean and standard deviation of a mini-batch, not a fixed one.
#### Consistency Regularization
- assume that the well-trained models should be regularized to have consistent results in the same space.
- ![](https://i.imgur.com/n95y3e5.png)