Benchmarks results ================== Digits datasets --------------- We use the same network architecture, number of epochs, batch\_size for all experiments. These are the results obtained with the default configuration in ``digits_network.json``. The feature extractor is a convolutional neural network with 3 convolutional layers. The classifier is a 3-layer fully connected neural network. MNIST -> MNIST-M (5 runs) ~~~~~~~~~~~~~~~~~~~~~~~~~ +----------+-----------------+-----------------+ | Method | source acc | target acc | +==========+=================+=================+ | Source | 89.0% +- 2.52 | 34.0% +- 1.71 | +----------+-----------------+-----------------+ | DANN | 94.2% +- 1.57 | 37.5% +- 2.85 | +----------+-----------------+-----------------+ | CDAN | 98.7% +- 0.19 | 68.4% +- 1.80 | +----------+-----------------+-----------------+ | CDAN-E | 98.7% +- 0.12 | 69.6% +- 1.51 | +----------+-----------------+-----------------+ | DAN | 98.0% +- 0.68 | 47.0% +- 1.85 | +----------+-----------------+-----------------+ | JAN | 96.4% +- 4.57 | 52.9% +- 2.16 | +----------+-----------------+-----------------+ | WDGRL | 93.9% +- 2.70 | 52.0% +- 4.82 | +----------+-----------------+-----------------+ MNIST -> USPS (5 runs) ~~~~~~~~~~~~~~~~~~~~~~ +----------+-----------------+-----------------+ | Method | source acc | target acc | +==========+=================+=================+ | Source | 99.2% +- 0.08 | 94.2% +- 1.07 | +----------+-----------------+-----------------+ | DANN | 99.1% +- 0.15 | 93.8% +- 1.06 | +----------+-----------------+-----------------+ | CDAN | 98.8% +- 0.17 | 90.7% +- 1.17 | +----------+-----------------+-----------------+ | CDAN-E | 98.9% +- 0.11 | 90.3% +- 0.98 | +----------+-----------------+-----------------+ | DAN | 99.0% +- 0.14 | 95.0% +- 0.83 | +----------+-----------------+-----------------+ | JAN | 98.6% +- 0.30 | 89.5% +- 2.00 | +----------+-----------------+-----------------+ | WDGRL | 98.7% +- 0.13 | 85.7% +- 6.57 | +----------+-----------------+-----------------+ SVHN -> MNIST (5 runs) ~~~~~~~~~~~~~~~~~~~~~~ +----------+-----------------+-----------------+ | Method | source acc | target acc | +==========+=================+=================+ | Source | 80.4% +- 1.65 | 60.2% +- 1.98 | +----------+-----------------+-----------------+ | DANN | 80.7% +- 2.09 | 61.7% +- 2.75 | +----------+-----------------+-----------------+ | CDAN | 80.5% +- 1.91 | 79.0% +- 3.13 | +----------+-----------------+-----------------+ | CDAN-E | 82.0% +- 0.74 | 77.9% +- 5.59 | +----------+-----------------+-----------------+ | DAN | 80.7% +- 1.26 | 54.8% +- 2.76 | +----------+-----------------+-----------------+ | JAN | 79.6% +- 0.67 | 57.9% +- 2.35 | +----------+-----------------+-----------------+ | WDGRL | 80.5% +- 1.09 | 59.5% +- 3.61 | +----------+-----------------+-----------------+ MNIST -> SVHN (5 runs) ~~~~~~~~~~~~~~~~~~~~~~ This problem is much harder than the others and results are usually not reported. Indeed, most methods fail to improve performance -- on the contrary, aligning features results in decreased global performance. +----------+------------------+-----------------+ | Method | source acc | target acc | +==========+==================+=================+ | Source | 91.6% +- 2.28 | 16.4% +- 3.31 | +----------+------------------+-----------------+ | DANN | 96.2% +- 0.24 | 19.5% +- 2.60 | +----------+------------------+-----------------+ | CDAN | 67.0% +- 12.74 | 11.5% +- 1.62 | +----------+------------------+-----------------+ | CDAN-E | 59.8% +- 18.99 | 11.3% +- 1.08 | +----------+------------------+-----------------+ | DAN | 93.3% +- 3.94 | 16.7% +- 1.19 | +----------+------------------+-----------------+ | JAN | 68.4% +- 12.72 | 11.5% +- 1.53 | +----------+------------------+-----------------+ | WDGRL | 77.4% +- 3.11 | 13.8% +- 1.75 | +----------+------------------+-----------------+ Office 31 dataset ----------------- We use the same network architecture, number of epochs, batch\_size for all experiments. These are the results obtained with the default configuration in ``office31_network.json``. The feature extractor is a ResNet50 with the last layer removed. The task classifier is linear. The results on the source look a bit lower than results reported in the literature (which I found from 68 to 80%). Note, however, that the parameters haven't been full finetuned, and that the confidence intervals are wider than those reported in the literature. Amazon to Webcam (5 runs) ~~~~~~~~~~~~~~~~~~~~~~~~~ +----------+-----------------+-----------------+ | Method | source acc | target acc | +==========+=================+=================+ | Source | 83.0% +- 1.04 | 59.7% +- 3.67 | +----------+-----------------+-----------------+ | DANN | 82.7% +- 1.56 | 73.4% +- 5.63 | +----------+-----------------+-----------------+ | CDAN | 82.0% +- 1.30 | 82.3% +- 3.52 | +----------+-----------------+-----------------+ | CDAN-E | 82.7% +- 0.74 | 81.6% +- 4.67 | +----------+-----------------+-----------------+ | DAN | 83.0% +- 1.65 | 68.1% +- 1.96 | +----------+-----------------+-----------------+ | JAN | 82.0% +- 2.02 | 64.6% +- 3.70 | +----------+-----------------+-----------------+ | WDGRL | 83.4% +- 1.72 | 75.5% +- 3.13 | +----------+-----------------+-----------------+ Amazon to DSLR (5 runs) ~~~~~~~~~~~~~~~~~~~~~~~ +----------+-----------------+-----------------+ | Method | source acc | target acc | +==========+=================+=================+ | Source | 81.9% +- 1.02 | 62.7% +- 3.25 | +----------+-----------------+-----------------+ | DANN | 80.2% +- 0.99 | 68.8% +- 2.78 | +----------+-----------------+-----------------+ | CDAN | 81.2% +- 2.06 | 72.0% +- 2.09 | +----------+-----------------+-----------------+ | CDAN-E | 80.1% +- 1.35 | 70.9% +- 3.24 | +----------+-----------------+-----------------+ | DAN | 82.3% +- 2.35 | 68.7% +- 1.36 | +----------+-----------------+-----------------+ | JAN | 78.1% +- 4.77 | 62.3% +- 2.39 | +----------+-----------------+-----------------+ | WDGRL | 79.6% +- 2.53 | 68.8% +- 1.63 | +----------+-----------------+-----------------+ DSLR to Amazon (5 runs) ~~~~~~~~~~~~~~~~~~~~~~~ +----------+-----------------+-----------------+ | Method | source acc | target acc | +==========+=================+=================+ | Source | 95.6% +- 2.08 | 55.6% +- 6.44 | +----------+-----------------+-----------------+ | DANN | 91.6% +- 2.42 | 53.8% +- 5.61 | +----------+-----------------+-----------------+ | CDAN | 87.2% +- 3.09 | 56.6% +- 4.87 | +----------+-----------------+-----------------+ | CDAN-E | 89.1% +- 2.09 | 54.1% +- 6.71 | +----------+-----------------+-----------------+ | DAN | 95.9% +- 2.05 | 58.1% +- 5.49 | +----------+-----------------+-----------------+ | JAN | 92.8% +- 2.75 | 54.1% +- 5.92 | +----------+-----------------+-----------------+ | WDGRL | 88.8% +- 2.08 | 56.6% +- 4.03 | +----------+-----------------+-----------------+