Mixup-Inference

Tianyu Pang, Kun Xu, Jun Zhu / Mixup Inference Better Exploiting Mixup to Defend Adversarial Attacks / 2020 The International Conference on Learning Representations (ICLR)

๋ณธ ๋…ผ๋ฌธ์€ ๋ฏน์Šค์—…์„ ํ†ตํ•ด adversarial Robustness๋ฅผ ํ–ฅ์ƒ์‹œํ‚ค๋Š” ๋ฐฉ๋ฒ•์„ ์ œ์‹œํ•œ๋‹ค. ๊ธฐ์กด ๋ฏน์Šค์—…๊ณผ ๋‹ฌ๋ฆฌ Training ๋‹จ๊ณ„์—์„œ input์— noise๋ฅผ ์ถ”๊ฐ€ํ•˜๋Š” ๋ฐฉ๋ฒ•์ด ์•„๋‹Œ inference๋œ ๊ฒฐ๊ณผ๋ฅผ mixupํ•˜๋Š” ๋ฐฉ๋ฒ•์ด๋‹ค.

๋…ผ๋ฌธ๋งํฌ https://arxiv.org/abs/1909.11515

1. Problem Definition

1) Adversarial Attacks

data label pair (x, y)์— ์ถ”๊ฐ€๋กœ adversaial binary ๋ณ€์ˆ˜ z(1์ผ ๊ฒฝ์šฐ adversarial)๋ฅผ ์ถ”๊ฐ€ํ•ด์„œ ์‚ฌ์šฉํ•œ๋‹ค. ๋ณธ ๋…ผ๋ฌธ์—์„œ๋Š” $l_p$-norm ์–ดํƒ์„ ์‚ฌ์šฉํ•˜๋ฉฐ $(||\delta||_p \leq\epsilon)$, clean sample $x_0$์— ๋Œ€ํ•ด ๋…ธ์ด์ฆˆ๊ฐ€ ์ถ”๊ฐ€๋œ $x$๊ฐ’์€ ๋‹ค์Œ๊ณผ ๊ฐ™๋‹ค.

adversarial input

2) Mixup in Training

mixup ๋ฐฉ๋ฒ•์€ Beyond Empirical Risk Minimizationarrow-up-right์—์„œ ์ฒ˜์Œ ์ œ์‹œ๋œ ๋ฐฉ๋ฒ•์œผ๋กœ ๋‘ ์ƒ˜ํ”Œ $(x_i, y_i), (x_j, y_j)$์˜ ์„ ํ˜•๊ฒฐํ•ฉ์„ ํ†ตํ•œ data augmentation๊ธฐ๋ฒ•์ด๋ผ๊ณ  ๋ณผ ์ˆ˜ ์žˆ๋‹ค. ์ƒˆ๋กญ๊ฒŒ ๋งŒ๋“ค์–ด์ง„ ๊ฐ€์ƒ์˜ ๋ฐ์ดํ„ฐ $(\tilde{x}, \tilde{y})$๋Š” ๋‹ค์Œ๊ณผ ๊ฐ™๋‹ค. $\tilde x = \lambda x_i + (1-\lambda) x_j$, $\tilde y = \lambda y_i + (1-\lambda) y_j$, where $\lambda \sim Beta(\alpha, \alpha)$ ๋‘ ์ƒ˜ํ”Œ๊ฐ„์˜ ๋นˆ๊ณต๊ฐ„์— ์ƒˆ๋กœ์šด ๊ฐ€์ƒ ๋ฐ์ดํ„ฐ๋ฅผ ์ฑ„์›Œ๋„ฃ์Œ์œผ๋กœ์„œ ๋‹ค์–‘ํ•œ ์ƒ˜ํ”Œ์„ ํ•™์Šตํ•˜๊ธฐ ๋•Œ๋ฌธ์— model์˜ ์ „๋ฐ˜์ ์ธ ์„ฑ๋Šฅ์„ ํ–ฅ์ƒ์‹œํ‚ค๋Š” ๋™์‹œ์— adversarial robustness๋„ ํ–ฅ์ƒ์‹œํ‚จ๋‹ค๊ณ  ์•Œ๋ ค์ ธ ์žˆ๋‹ค. ๋˜ํ•œ, ์ผ๋ฐ˜์ ์ธ Adversarial Training๊ณผ ๋น„๊ตํ–ˆ์„๋•Œ ์—ฐ์‚ฐ๋Ÿ‰์ด ํ˜„์ „ํžˆ ๋‚ฎ๊ณ , clean data์— ๋Œ€ํ•œ ์„ฑ๋Šฅ๋„ ๋ณด์žฅํ•˜๊ธฐ ๋•Œ๋ฌธ์— ์ด ๋ถ€๋ถ„์—์„œ ์žฅ์ ์„ ๊ฐ€์ง€๊ณ  ์žˆ๋‹ค.

2. Motivation

๋ณธ ๋…ผ๋ฌธ์—์„œ๋Š” ๊ธฐ์กด์˜ mixup training์ฒ˜๋Ÿผ input sample์„ ์„ž์–ด์„œ training data๋กœ๋งŒ ํ™œ์šฉํ•˜๋Š” ๊ฒฝ์šฐ, input์˜ locality์—์„œ ํฌ๊ฒŒ ๋ฒ—์–ด๋‚˜์ง€ ๋ชปํ•˜๊ธฐ ๋•Œ๋ฌธ์— mixup ๋ณธ์—ฐ์˜ 'globally linear behavior'๋ฅผ ๊ทน๋Œ€ํ™”ํ•˜์ง€ ๋ชป ํ•œ๋‹ค๊ณ  ์ฃผ์žฅํ•œ๋‹ค. (์—ฌ๊ธฐ์„œ 'globally linear behavior'๋ž€ ์œ„์—์„œ ์–˜๊ธฐํ•œ ๋‘ ์ƒ˜ํ”Œ์˜ ๋นˆ๊ณต๊ฐ„์— ์กด์žฌํ•˜๋Š” ์ƒˆ๋กœ์šด ๊ฐ€์ƒ ๋ฐ์ดํ„ฐ๋ฅผ ์‚ฌ์šฉํ•˜๋Š” ๊ฒƒ์„ ์˜๋ฏธํ•œ๋‹ค.) ์ฆ‰, ์ €์ž๋Š” mixup์˜ ํšจ๊ณผ๋ฅผ ๋” ๊ทน๋Œ€ํ™”ํ•˜๊ธฐ ์œ„ํ•ด inference phase์—์„œ mixup ํ•  ๊ฒƒ์„ ์ œ์•ˆํ•œ๋‹ค.

3. Method

1) Notations

$y$ : ground truth $\hat y$ : predicted label $y_s \sim p_x(y)$ : sampled label $x_s \sim p_s(x|y_s)$: sampled data $\tilde x = \lambda x + (1-\lambda) x_s$ $z$ : adversarial flag

$F$ : mixup-trained model $H$ : linear function $G$ : extra non-linear part of F

2) Mixup Inference

์ €์ž๋Š” ์ž˜ training ๋œ mixup ๋ชจ๋ธ์€ ์•„๋ž˜์™€ ๊ฐ™์ด ๊ฐ clean input๋“ค์˜ ์„ ํ˜•ํ•จ์ˆ˜์˜ ๊ฒฐํ•ฉ์œผ๋กœ ๋‚˜ํƒ€๋‚ผ ์ˆ˜ ์žˆ๋‹ค๊ณ  ์„ค๋ช…ํ•œ๋‹ค. method_H ๋ณธ ๋…ผ๋ฌธ์—์„œ๋Š” ์ด๋Ÿฌํ•œ ์„ ํ˜•๊ฒฐํ•ฉ์œผ๋กœ ์ „๊ฐœ๋˜๋Š” ๋‚ด์šฉ์— ๋Œ€ํ•œ ์ž์„ธํ•œ ๋‚ด์šฉ์€ ์–ธ๊ธ‰๋˜์–ด ์žˆ์ง€ ์•Š๊ณ , MixUp as Locally Linear Out-Of-Manifold Regularizationarrow-up-right ๋…ผ๋ฌธ์„ ์ฐธ์กฐํ•˜์—ฌ ๋ฐ”๋กœ ์‚ฌ์šฉํ•˜๋Š” ๊ฒƒ์œผ๋กœ ์„ค๋ช…๋˜์–ด ์žˆ๋‹ค.

๋‹ค๋งŒ, Adversarial Training ์˜ ๊ฒฝ์šฐ noise์— ๋Œ€ํ•œ non-linear part G๊ฐ€ ์ถ”๊ฐ€๋˜๊ณ ,

method_H_andG

์ตœ์ข…์ ์œผ๋กœ mixup ๊ฐ’ $\tilde x$์— ๋Œ€ํ•œ ๊ฒฐ๊ณผ๋Š” ๋‹ค์Œ๊ณผ ๊ฐ™์ด ์ „๊ฐœ๋œ๋‹ค.

method_H_xtilde

Mixup Inference๋Š” ์ด $F(\tilde x)$๊ฐ’์˜ N๋ฒˆ ํ‰๊ท ์„ ์‚ฌ์šฉํ•ด์„œ model์„ ์—…๋ฐ์ดํŠธ ํ•˜๋Š” ๋ฐฉ๋ฒ•์ด๋‹ค.

method_MI

(5)์‹์„ ๋” ์ •๋ฆฌํ•˜๋ฉด, clean data $x_0$์™€ sampled data $x_s$์— ๋Œ€ํ•ด $H_y(x_0) = 1$, $H_{y_s}(x_S) = 1$ ์ด๊ณ , ์•„๋ž˜ ์ˆ˜์‹์€ $F$ ๊ฒฐ๊ณผ๋ฅผ ๊ฐ๊ฐ $y$, $\hat y$์— ๋Œ€ํ•ด ์•„๋ž˜์™€ ๊ฐ™์ด ๋‚˜ํƒ€๋‚ผ ์ˆ˜ ์žˆ๋‹ค.

method_MI_y_yhat

$y$, $\hat y$ ๋‘ ๊ฒฝ์šฐ ๋ชจ๋‘ $y_s$(sampled label)์˜ ์˜ํ–ฅ์„ ๋ฐ›๊ธฐ ๋•Œ๋ฌธ์— ๋…ผ๋ฌธ์—์„œ๋Š” MI-PL($y=\hat y$), MI-OL($y\neq\hat y$) ๋‘ ๊ฐ€์ง€ ๋ฒ„์ „์„ ๋‚˜๋ˆ ์„œ ํ•จ๊ป˜ ์‚ดํŽด๋ณผ ํ•„์š”๊ฐ€ ์žˆ๋‹ค๊ณ  ์„ค๋ช…ํ•œ๋‹ค

method_PL_OL

๊ฐ๊ฐ์˜ ๊ฒฝ์šฐ $F$๊ฐ’์„ ์š”์•ฝํ•˜๋ฉด ๋‹ค์Œ๊ณผ ๊ฐ™๋‹ค. ($z=1$์ธ ๊ฒฝ์šฐ adversarial sample, $z=0$์ธ ๊ฒฝ์šฐ clean sample์„ ์˜๋ฏธํ•œ๋‹ค.)

method_PL_OL

์ถ”๊ฐ€๋กœ Mixup Inference ์ „/ํ›„ robustness ํ–ฅ์ƒ์ •๋„, ์‹ค์ œ attack๋œ ์ƒ˜ํ”Œ ํƒ์ง€์ •๋„์— ๋Œ€ํ•œ ํ‰๊ฐ€์ง€ํ‘œ๋กœ ๊ฐ๊ฐ Robustness Improving Condition(RIC)์™€ Detection Gap(DG)๋ฅผ ์ •์˜ํ–ˆ๋‹ค. method_RIC_DG RIC(10๋ฒˆ ์‹)๋Š” adversarial sample์— ๋Œ€ํ•ด ํ•™์Šต ์ดํ›„์˜ ์˜ˆ์ธก๋œ F๊ฐ’ ์ฆ‰, confidence๊ฐ€ ๋‚ฎ์•„์งˆ์ˆ˜๋ก, DG(11๋ฒˆ ์‹)๋Š” adversarial atack์ด ๋œ sample๊ณผ ์•„๋‹Œ sample๊ฐ„์˜ confidence ์ฐจ์ด๊ฐ€ ํด ์ˆ˜๋ก ํ•™์Šต์ด ์ž˜ ๋œ ๊ฒฐ๊ณผ์ž„์„ ์˜๋ฏธํ•œ๋‹ค.

3) Theoretical Analysis

์œ„์—์„œ ์ œ์‹œํ•œ RIC ์‹์€ MI-PL, MI-OL ๊ฐ๊ฐ ์•„๋ž˜์™€ ๊ฐ™์ด ์ •๋ฆฌํ•  ์ˆ˜ ์žˆ๋‹ค.

  • MI-PL (Predicted Label) MI_analysis_PL

  • MI-OL (Other Label) MI_analysis_OL

  • Analysis results MI_analysis

๊ฐ€์žฅ ์™ผ์ชฝ plot๊ณผ ๊ฐ€์šด๋ฐ plot์—์„œ adversarial inputs(์ฃผํ™ฉ์ƒ‰ ์‹ค์„ )๋ฅผ ๋ณด๋ฉด ์‹ค์ œ๋กœ MI๋ฅผ ์ ์šฉํ•˜์ง€ ์•Š์•˜์„ ๋•Œ($\lambda = 1$)๋ณด๋‹ค MI๋ฅผ ์ ์šฉํ–ˆ์„ ๋•Œ($\lambda \neq 1$), $F_y$๋Š” ์ฆ๊ฐ€ํ•˜๊ณ  $F_{\hat y}$์€ ๊ฐ์†Œํ•˜๊ธฐ ๋•Œ๋ฌธ์— ์ด๋Š” RIC(10๋ฒˆ์‹)์„ ๋งŒ์กฑํ•˜๋Š” ๊ฒฐ๊ณผ์ž„์„ ์•Œ ์ˆ˜ ์žˆ๋‹ค. ๊ฐ€์žฅ ์˜ค๋ฅธ์ชฝ์˜ plot์€ [$G_k(\delta;x_0)-G_k(\lambda\delta;\tilde x_0)$]์„ ๊ทธ๋ฆฐ ๊ทธ๋ž˜ํ”„์ด๋‹ค. ๊ทธ๋ž˜ํ”„ ๊ฐ’์„ ๋ณด๋ฉด ์œ„์˜ 12, 15(ํŽธ์˜์ƒ ์›๋ž˜ ์ˆ˜์‹์˜ minus๊ฐ’์„ ๊ทธ๋ž˜ํ”„์— ํ‘œ์‹œํ•จ) ์‹์—์„œ ์ œ์‹œํ•œ ์กฐ๊ฑด์„ ๋ชจ๋‘ ๋งŒ์กฑํ•˜๊ณ  ์žˆ๊ธฐ ๋•Œ๋ฌธ์—(์ฆ‰, RIC ์„ฑ์งˆ์„ ๋งŒ์กฑํ•œ๋‹ค๋Š” ์˜๋ฏธ) MI ๋ฐฉ๋ฒ•์ด adversarial training์— ํšจ๊ณผ์ ์ธ ๋ฐฉ๋ฒ•์ž„์„ ๋ณด์—ฌ์ค€๋‹ค.

4. Experiment

Experiment setup

  • Dataset : CIFAR-10, CIFAR-100

  • Model : ResNet-50

  • Adversarial Attack : Gaussian noise, Random rotation, Random cropping and resizing, Random cropping and padding

  • Baseline : Mixup(๊ธฐ๋ณธ์ ์ธ mixup training ํ›„ attack์— ๋Œ€ํ•œ accuracy ์ธก์ •), Interpolated AT(Interpolated Adversarial Trainingarrow-up-right์—์„œ ์†Œ๊ฐœ๋œ Mixup ๋ฐฉ๋ฒ•์„ ์ด์šฉํ•œ AT method)

Result

๋…ผ๋ฌธ์—์„œ๋Š” MI-PL๊ณผ MI-OL์„ ๊ฒฐํ•ฉํ•œ Mi-Combined ๋ฒ„์ „๋„ ์‹คํ—˜๊ฒฐ๊ณผ์— ํฌํ•จ์‹œ์ผฐ๋‹ค. MI_PL์„ ์ ์šฉํ•˜๋‹ค๊ฐ€ adversarial input detection ๊ฐ’์ด ํŠน์ • ์ž„๊ณ„๊ฐ’์„ ๋„˜์–ด๊ฐ€๋ฉด MI_OL์„ ์ ์šฉํ•˜๋Š” ๋ฐฉ๋ฒ•์ด๋‹ค.

์•„๋ž˜ ์‹คํ—˜๊ฒฐ๊ณผ๋ฅผ ํ†ตํ•ด, Mixup, Interpolated AT ๋ชจ๋‘ Mixup Inference method๋ฅผ ํ•จ๊ป˜ ์‚ฌ์šฉํ–ˆ์„๋•Œ ๋” ์ข‹์€ ์„ฑ๋Šฅ์„ ๋ณด์—ฌ์ฃผ๋Š” ๊ฒƒ์„ ์•Œ ์ˆ˜ ์žˆ๋‹ค.

CIFAR-10

result1

CIFAR-100

result2

5. Conclusion

๋ชจ๋ธ์˜ ์˜ˆ์ธก๊ฐ’์„ Mixup ํ•˜๋Š” ๋ฐœ์ƒ์ด ์ƒˆ๋กœ์›Œ์„œ ๊ด€์‹ฌ์žˆ๊ฒŒ ๋ณธ ๋…ผ๋ฌธ์ด์—ˆ๋‹ค. Input mixup, Manifold Mixup์— ์ด์–ด์„œ ์ƒˆ๋กœ์šด ๋ฐฉ๋ฒ•์˜ mixup ๋ฐฉ๋ฒ•์œผ๋กœ ์ƒ๊ฐ ํ•  ์ˆ˜ ์žˆ์„ ๊ฒƒ ๊ฐ™๋‹ค. ํ•˜์ง€๋งŒ motivation์—์„œ ์ œ์‹œํ–ˆ๋“ฏ์ด Inference๋‹จ๊ณ„์—์„œ Mixupํ•˜๋Š” ๊ฒƒ์ด Mixup ๋ณธ์—ฐ์˜ 'globally linear behavior' ์„ฑ์งˆ์„ ํ™•๋Œ€์‹œํ‚ฌ ์ˆ˜ ์žˆ์„ ๊ฑฐ๋ผ๋Š” ์ฃผ์žฅ์— ๋Œ€ํ•œ ๊ทผ๊ฑฐ๋Š” ๋ช…ํ™•ํ•˜๊ฒŒ ์ฆ๋ช…๋˜์ง€ ์•Š๊ณ  ์‹คํ—˜์ ์œผ๋กœ ์„ฑ๋Šฅ๋น„๊ต๋งŒ ์ œ์‹œ๋œ ๊ฒƒ์ด ์•„์‰ฌ์šด ์ ์ด์—ˆ๋‹ค.


Author Information

  • ๊น€์ •ํ—Œ(JUNGHURN KIM): Master student, KSE, KAIST

6. Reference & Additional materials

  • github https://github.com/P2333/Mixup-Inference

Last updated