๋…ผ๋ฌธ๋ช…: Adversarial Retain-free Unlearning for Bearing Prognostics and Health Management

์ €์ž: Chaewon Yoon, Jiyoung Lee, Hoki Kim

๊ฒŒ์žฌ์ง€: IEEE Transactions on Industrial Informatics

URL: https://ieeexplore.ieee.org/document/11414433

์„œ๋ก 


์ œ4์ฐจ ์‚ฐ์—…ํ˜๋ช…์˜ ๋„๋ž˜๋กœ ์‚ฐ์—… ํ˜„์žฅ์—์„œ ์„ผ์„œ ๋ฐ์ดํ„ฐ์™€ ์ธ๊ณต์ง€๋Šฅ์„ ๊ธฐ๋ฐ˜์œผ๋กœ ํ•œ PHM(Prognostics and Health Management) ๊ธฐ์ˆ ์ด ์ ๊ทน์ ์œผ๋กœ ์‚ฌ์šฉ๋˜๊ณ  ์žˆ์Šต๋‹ˆ๋‹ค. ํŠนํžˆ ๋ฒ ์–ด๋ง ๊ฒฐํ•จ ์ง„๋‹จ๊ณผ ๊ฐ™์€ ํ•ต์‹ฌ ์„ค๋น„ ๊ด€๋ฆฌ ๋ถ„์•ผ์—์„œ ๋”ฅ๋Ÿฌ๋‹ ๋ชจ๋ธ์ด ๋†’์€ ์ •ํ™•๋„๋ฅผ ๋ณด์ด๋ฉฐ ์‚ฐ์—… ๊ฒฝ์Ÿ๋ ฅ์„ ์ขŒ์šฐํ•˜๋Š” ์ฃผ์š” ๊ธฐ์ˆ ๋กœ ์ž๋ฆฌ ์žก์•˜์Šต๋‹ˆ๋‹ค.

๋จธ์‹  ์–ธ๋Ÿฌ๋‹(Machine Unlearning) ๊ธฐ์ˆ ์€ ์ฃผ์š” ๋ฐ์ดํ„ฐ ๋ฒ•๋ น์ธ GDPR์—์„œ ๋ช…์‹œํ•˜๋Š” โ€œ์žŠํ˜€์งˆ ๊ถŒ๋ฆฌ(Right To Be Forgotten)โ€์ฒ˜๋Ÿผ, ๋ชจ๋ธ์ด ํ•™์Šตํ•œ ๋ฐ์ดํ„ฐ ์ค‘ ์ผ๋ถ€๋ฅผ ๋‚˜์ค‘์— ์‚ญ์ œํ•ด์•ผ ํ•˜๋Š” ์ƒํ™ฉ์—์„œ ์œ ์šฉํ•˜๊ฒŒ ์“ฐ์ž…๋‹ˆ๋‹ค.

๋ณธ ์—ฐ๊ตฌ์‹ค์˜ ๋…ผ๋ฌธ โ€œAdversarial Retain-free Unlearning for Bearing Prognostics and Health Managementโ€ [Paper]์€ ๊ณต์ • ํ™˜๊ฒฝ์˜ ํ˜„์‹ค์ ์ธ ์ œ์•ฝ ์กฐ๊ฑด์„ ๊ณ ๋ คํ•ด, ์‚ญ์ œ ๋Œ€์ƒ ๋ฐ์ดํ„ฐ์™€ ์ด๋ฏธ ํ•™์Šต๋œ ๋ชจ๋ธ๋งŒ์œผ๋กœ ๋จธ์‹  ์–ธ๋Ÿฌ๋‹์„ ์ˆ˜ํ–‰ํ•  ์ˆ˜ ์žˆ๋Š” ์ƒˆ๋กœ์šด ํ”„๋ ˆ์ž„์›Œํฌ๋ฅผ ์ œ์•ˆํ•ฉ๋‹ˆ๋‹ค.

์ด ๊ธ€์—์„œ๋Š” ์ œ์กฐ ๊ณต์ • ํ™˜๊ฒฝ์—์„œ ๊ธฐ์กด ๋จธ์‹  ์–ธ๋Ÿฌ๋‹ ๋ฐฉ์‹์ด ๊ฐ–๋Š” ํ•œ๊ณ„์™€ ์ œ์•ˆ๋œ ARU ๋ฐฉ๋ฒ•์„ ํ•จ๊ป˜ ์‚ดํŽด๋ณด๊ฒ ์Šต๋‹ˆ๋‹ค.

ARU ๊ฐœ์š”

์‚ฌ์ „ ์ง€์‹


์ ๋Œ€์  ์˜ˆ์ œ

์ ๋Œ€์  ์˜ˆ์ œ(Adversarial Attack)๋Š” ํŠน์ • ์ƒ˜ํ”Œ์— ์ ๋Œ€์  ๊ณต๊ฒฉ์œผ๋กœ ์•„์ฃผ ์ž‘์€ ๊ต๋ž€(\(\epsilon\))์„ ์ถ”๊ฐ€ํ•ด ๋ชจ๋ธ์ด ์ƒ˜ํ”Œ์„ ์™„์ „ํžˆ ๋‹ค๋ฅธ ํด๋ž˜์Šค๋กœ ๋ถ„๋ฅ˜ํ•˜๋„๋ก ์œ ๋„ํ•ฉ๋‹ˆ๋‹ค.

๋Œ€ํ‘œ์ ์ธ ์ ๋Œ€์  ๊ณต๊ฒฉ ๋ฐฉ๋ฒ•์œผ๋กœ๋Š” FGSM๊ณผ PGD๊ฐ€ ์žˆ์Šต๋‹ˆ๋‹ค :

1. FGSM(Fast Gradient Sign Method) : ํ•œ ๋ฒˆ์˜ gradient ๊ณ„์‚ฐ์œผ๋กœ ๊ต๋ž€์„ ์ถ”๊ฐ€ํ•ฉ๋‹ˆ๋‹ค.

\[\begin{equation} \delta = \epsilon \cdot \operatorname{sign} (\nabla_{x}\mathcal{L} (f_{\theta} (x)), y), \end{equation}\]

2. PGD(Projected Gradient Descent) : FGSM์„ ์—ฌ๋Ÿฌ ๋‹จ๊ณ„๋กœ ํ™•์žฅํ•ด ๋” ๊ฐ•๋ ฅํ•œ ๊ณต๊ฒฉ์„ ์ˆ˜ํ–‰ํ•ฉ๋‹ˆ๋‹ค.

\[\begin{equation} x^{t+1} = \Pi_{\lVert \delta \rVert_{\infty}\le \epsilon} \left (x^t + \alpha \cdot \operatorname{sign} (\nabla_{x^t}\mathcal{L} (f_{\theta} (x^t), y)) \right), \end{equation}\]

๋จธ์‹  ์–ธ๋Ÿฌ๋‹

๋ชจ๋ธ์ด ํ•™์Šตํ•œ ์ „์ฒด ๋ฐ์ดํ„ฐ๋ฅผ \(\mathcal{D}\) ๋ผ๊ณ  ํ•  ๋•Œ, ์‚ญ์ œ ์—ฌ๋ถ€์— ๋”ฐ๋ผ ๋ฐ์ดํ„ฐ๋ฅผ ์•„๋ž˜์™€ ๊ฐ™์ด ๋‘ ๊ฐ€์ง€๋กœ ๋‚˜๋ˆŒ ์ˆ˜ ์žˆ์Šต๋‹ˆ๋‹ค :

1. ์‚ญ์ œ ๋Œ€์ƒ ๋ฐ์ดํ„ฐ (Forget data) = \(\mathcal{D_F}\)

2. ์œ ์ง€ ๋Œ€์ƒ ๋ฐ์ดํ„ฐ (Retain data) = \(\mathcal{D_R}\)

๋จธ์‹  ์–ธ๋Ÿฌ๋‹์€ ๋ชจ๋ธ์ด \(\mathcal{D_R}\)๋กœ๋งŒ ํ•™์Šตํ•œ ๋ชจ๋ธ์„ ์ด์ƒ์ ์ธ ์ƒํƒœ๋กœ ๊ฐ€์ •ํ•ฉ๋‹ˆ๋‹ค. ์ด ๊ธ€์—์„œ๋Š” ์ด๋Ÿฌํ•œ ์ด์ƒ์ ์ธ ํ•™์Šต ๋ฐฉ๋ฒ•์„ Retrain์ด๋ผ๊ณ  ์ง€์นญํ•ฉ๋‹ˆ๋‹ค. ํ•˜์ง€๋งŒ ๋ชจ๋ธ์„ ์ฒ˜์Œ๋ถ€ํ„ฐ ์žฌํ•™์Šตํ•˜๋Š” Retrain์€ ์‹œ๊ฐ„, ๋น„์šฉ์ ์ธ ๋ถ€๋‹ด์ด ํฝ๋‹ˆ๋‹ค. ๋”ฐ๋ผ์„œ ์ด๋ฏธ ํ•™์Šต๋œ ๋ชจ๋ธ์„ ํŠน์ • ์•Œ๊ณ ๋ฆฌ์ฆ˜์œผ๋กœ ํ•™์Šตํ•ด ์‚ญ์ œ ๋Œ€์ƒ ๋ฐ์ดํ„ฐ๋ฅผ ์ œ๊ฑฐํ•˜๋Š” ๋จธ์‹  ์–ธ๋Ÿฌ๋‹ ๊ธฐ์ˆ ์ด ํ•„์š”ํ•ฉ๋‹ˆ๋‹ค.

๋จธ์‹  ์–ธ๋Ÿฌ๋‹์€ ๋ชจ๋ธ์ด \(\mathcal{D_F}\)์˜ ์ •๋ณด๋ฅผ ์žŠ๋Š” ๋™์‹œ์—, \(\mathcal{D_R}\)์˜ ์ •๋ณด๋ฅผ ์œ ์ง€ํ•˜๋„๋ก ์œ ๋„ํ•ด, Retrain ์ƒํƒœ์™€ ๋น„์Šทํ•ด์ง€๋Š” ๊ฒƒ์„ ๋ชฉํ‘œ๋กœ ํ•ฉ๋‹ˆ๋‹ค.

retain-free ์–ธ๋Ÿฌ๋‹

๊ธฐ์กด ์—ฐ๊ตฌ์—์„œ ์ œ์•ˆ๋œ retain-free ์–ธ๋Ÿฌ๋‹ ๋ฐฉ๋ฒ•๋ก ์„ ์†Œ๊ฐœํ•ฉ๋‹ˆ๋‹ค. ๋ชจ๋ธ์˜ ๊ฐ€์ค‘์น˜๋ฅผ \(\theta\)๋ผ๊ณ  ํ•  ๋•Œ, ๊ฐ ๋ฐฉ๋ฒ•์€ ๋‹ค์Œ๊ณผ ๊ฐ™์ด ์—…๋ฐ์ดํŠธํ•ฉ๋‹ˆ๋‹ค :

1. Gradient Ascent (GA) : \(\mathcal{D_F}\)์— ๋Œ€ํ•œ ์†์‹ค์„ ์ตœ๋Œ€ํ™”ํ•ด ์–ธ๋Ÿฌ๋‹ํ•ฉ๋‹ˆ๋‹ค. \(\begin{equation} \theta \leftarrow \theta + \eta\cdot \nabla_{\theta} \mathcal{L} (f_{\theta} (x_f), y_f) \text{ where } (x_f, y_f) \in \mathcal{D_F} \end{equation}\)

2. Random Labeling (RL) : \(\mathcal{D_F}\) ์ƒ˜ํ”Œ์— ๋žœ๋ค ๋ผ๋ฒจ์„ ๋ถ™์—ฌ ์–ธ๋Ÿฌ๋‹ํ•ฉ๋‹ˆ๋‹ค. \(\begin{equation} \theta \leftarrow \theta - \eta \cdot \nabla_{\theta}\,\mathcal{L}\big(f_{\theta}(x_f), \tilde{y}\big) \end{equation}\)

3. Adversarial Machine UNlearning (AMUN) : \(\mathcal{D_F}\)์— ์ตœ์†Œํ•œ์˜ ๊ต๋ž€(\(\epsilon^{\star}\))์„ ๊ฐ€ํ•ด ์ ๋Œ€์  ์˜ˆ์ œ \(\mathcal{D_A}\)๋ฅผ ์–ป๊ณ , \(\mathcal{D_F} \cup \mathcal{D_A}\)์œผ๋กœ ๋ชจ๋ธ(\(\theta\))์„ ์žฌํ•™์Šตํ•ฉ๋‹ˆ๋‹ค.

์ ๋Œ€์  ์˜ˆ์ œ \(\mathcal{D_A} = (x_f^{adv}, y_f^{adv})\)๋Š” ๋‹ค์Œ๊ณผ ๊ฐ™์ด ๊ตฌ์„ฑํ•ฉ๋‹ˆ๋‹ค :

๋จผ์ € ์ ๋Œ€์  ๊ณต๊ฒฉ ์•Œ๊ณ ๋ฆฌ์ฆ˜ \(A\)์— ๋Œ€ํ•ด \(\mathcal{D_F}\)๋ฅผ ์˜ค๋ถ„๋ฅ˜ํ•˜๋Š” ์ตœ์†Œ ๊ต๋ž€ ํฌ๊ธฐ์ธ \(\epsilon^{\star}\)๋ฅผ ๊ตฌํ•ฉ๋‹ˆ๋‹ค. \(\begin{equation} \epsilon^\star = \min \{\epsilon : \arg\max_i f_{\theta} (A (x_f, \epsilon))_i \neq y_f \} \end{equation}\)

๋‹ค์Œ์œผ๋กœ, ์ตœ์†Œ ๊ต๋ž€ ํฌ๊ธฐ \(\epsilon^{\star}\)๋ฅผ ์ ์šฉํ•œ ์ ๋Œ€์  ์˜ˆ์ œ์ธ \(\mathcal{D_A} = $(x_f^{adv}, y_f^{adv})\)๋ฅผ ๊ตฌํ•ฉ๋‹ˆ๋‹ค. \(\begin{equation} x_f^{adv} = A(x_f, \epsilon^{star}), y_f^{adv} = arg \max_i f_{\theta}(x_f^{adv})_i \end{equation}\)

๋งˆ์ง€๋ง‰์œผ๋กœ ๋ชจ๋ธ์„ \(\mathcal{D_F}\cup\mathcal{D_A}\)๋กœ ์žฌํ•™์Šตํ•ฉ๋‹ˆ๋‹ค. \(\begin{equation} \theta \leftarrow \theta - \eta \sum_{ (x,y)\in \mathcal{D_F} \cup \mathcal{D_A}} \nabla_{\theta}\mathcal{L} (f_{\theta} (x),y) \end{equation}\)

๋ณธ๋ก 


๊ธฐ์กด retain-free ์–ธ๋Ÿฌ๋‹ ์•Œ๊ณ ๋ฆฌ์ฆ˜์€ ๋ชจ๋ธ์˜ ๊ตฌ์กฐ๋ฅผ ์‹ฌํ•˜๊ฒŒ ๋ฌด๋„ˆ๋œจ๋ฆฌ๋Š” ๊ฒฝ์šฐ๊ฐ€ ๋งŽ์•˜์Šต๋‹ˆ๋‹ค. ๋˜ํ•œ ๊ณต์ • ๋ฐ์ดํ„ฐ๋กœ ์‹คํ—˜ ์‹œ ๋ชจ๋ธ์˜ ๋ถ„๋ฅ˜ ๋Šฅ๋ ฅ ๋˜ํ•œ ์ €ํ•˜๋˜๋Š” ๋ฌธ์ œ๋ฅผ ๋ณด์˜€์Šต๋‹ˆ๋‹ค.

๋ณธ ๋…ผ๋ฌธ์€ ์ „์ฒด ๋ฐ์ดํ„ฐ์— ๋Œ€ํ•œ ์ ‘๊ทผ์ด ์ œํ•œ๋œ ํ™˜๊ฒฝ์—์„œ ๊ธฐ์กด retain-free ์–ธ๋Ÿฌ๋‹ ์•Œ๊ณ ๋ฆฌ์ฆ˜์˜ ๋ชจ๋ธ ๋ถ•๊ดด ๋ฌธ์ œ๋ฅผ ์ •๋Ÿ‰ํ™”ํ•ด ํ™•์ธํ•˜๊ณ , ์ด๋ฅผ ํ•ด๊ฒฐํ•˜๋Š” ARU ํ”„๋ ˆ์ž„์›Œํฌ๋ฅผ ์ œ์•ˆํ•ฉ๋‹ˆ๋‹ค.

Adversarial Retain-free Unlearning (ARU)๋Š” 2๊ฐ€์ง€ ๋‹จ๊ณ„๋กœ ์ด๋ฃจ์–ด์ง‘๋‹ˆ๋‹ค :

1.์ ๋Œ€์  ์˜ˆ์ œ ์ƒ์„ฑ : \(\mathcal{D_F}\)๋ฅผ ์ด์šฉํ•ด ๋ชจ๋ธ์ด ๊ธฐ์–ตํ•˜๋Š” ๋ฒ”์œ„ ๋‚ด์—์„œ \(\mathcal{D_R}\)๋กœ ์ธ์ง€ํ•˜๋Š” ์ ๋Œ€์  ์˜ˆ์ œ (\(\mathcal{\tilde{D}_R}\))๋ฅผ ์ƒ์„ฑํ•ฉ๋‹ˆ๋‹ค

2.๋ชจ๋ธ ์žฌํ•™์Šต : ์†์‹คํ•จ์ˆ˜ \(\mathcal{l}_{\texttt{SDA}}\)๋ฅผ ์ด์šฉํ•ด, ํ•™์Šต ๋ฐ์ดํ„ฐ๊ฐ€ \(\mathcal{D_F}\)๊ณผ๋Š” ๋ฉ€์–ด์ง€๋„๋ก, \(\mathcal{\tilde{D}_R}\)์™€๋Š” ๊ฐ€๊นŒ์›Œ์ง€๋„๋ก ์œ ๋„ํ•ฉ๋‹ˆ๋‹ค

๋ชจ๋ธ์˜ ๊ตฌ์กฐ ์ •๋Ÿ‰ํ™”

๋ณธ ๋…ผ๋ฌธ์€ Structural Similarity Correlation (SSC)๋ฅผ ํ†ตํ•ด ๊ธฐ์กด ์—ฐ๊ตฌ๋œ retain-free ์–ธ๋Ÿฌ๋‹ ๋ฐฉ๋ฒ•๋“ค์ด ๋ชจ๋ธ์˜ ๊ตฌ์กฐ๋ฅผ ์–ผ๋งˆ๋‚˜ ๋ณด์กดํ•˜๋Š”์ง€ ์ •๋Ÿ‰ํ™”ํ•ฉ๋‹ˆ๋‹ค.

SSC๋Š” ๋‹ค์Œ๊ณผ ๊ฐ™์€ ์ ˆ์ฐจ๋กœ ๊ณ„์‚ฐ๋ฉ๋‹ˆ๋‹ค :

SSC ์ง€ํ‘œ ๊ณ„์‚ฐ ๊ณผ์ •. Spearman's rank correlation์„ ์ด์šฉํ•ด Retrain ๋ชจ๋ธ์˜ RDM๊ณผ ์–ธ๋Ÿฌ๋‹ ํ›„ ๋ชจ๋ธ์˜ RDM ์‚ฌ์ด์˜ ์œ ์‚ฌ๋„๋ฅผ ์ •๋Ÿ‰ํ™”ํ•ฉ๋‹ˆ๋‹ค.

1. ๊ฐ retain ํด๋ž˜์Šค์˜ prototype ๋ฒกํ„ฐ ๊ณ„์‚ฐ

๋ฐ์ดํ„ฐ์…‹์˜ ํด๋ž˜์Šค \(C\)์— ๋Œ€ํ•ด, \(k\in C\)์ธ \(k\)์— ๋Œ€ํ•ด prototype ๋ฒกํ„ฐ \(p_k\)๋Š” ๋‹ค์Œ๊ณผ ๊ฐ™์ด ๊ณ„์‚ฐํ•ฉ๋‹ˆ๋‹ค :

\[\begin{equation} p_k = \frac{1}{|S_k|} \sum_{x \in S_k} f (x), \end{equation}\]

2. ํด๋ž˜์Šค ๊ฐ„ ์œ ์‚ฌ๋„ matrix ๊ตฌ์„ฑ

\(i,j \in C\)์ธ \(i,j\)์— ๋Œ€ํ•ด์„œ ํด๋ž˜์Šค prototype ๋ฒกํ„ฐ ๊ฐ„ ์ฝ”์‚ฌ์ธ ๊ฑฐ๋ฆฌ๋กœ ํด๋ž˜์Šค ๊ฐ„ ์œ ์‚ฌ๋„๋ฅผ ์ธก์ •ํ•ฉ๋‹ˆ๋‹ค.

ํด๋ž˜์Šค \(i,j\)๊ฐ„ ์œ ์‚ฌ๋„๋ฅผ ๊ธฐ๋ฐ˜์œผ๋กœ Representational Dissimilarity Matrix(RDM) \(M_{i,j}\)์„ ๊ตฌ์„ฑํ•ฉ๋‹ˆ๋‹ค.

\[\begin{equation} M_{ij} = 1 - \frac{p_i \cdot p_j}{\|p_i\| \|p_j\|} \end{equation}\]

3. Retrain์˜ RDM๊ณผ ์–ธ๋Ÿฌ๋‹ ํ›„ ๋ชจ๋ธ์˜ RDM ์‚ฌ์ด์˜ ์œ ์‚ฌ๋„ ์ธก์ •

ํ•ด๋‹น ๋‹จ๊ณ„๋Š” ๋‘ ํด๋ž˜์Šค์˜ ๊ตฌ์กฐ ์‚ฌ์ด์˜ ์ •๋Ÿ‰์ ์ธ ์œ ์‚ฌ๋„๊ฐ€ ์•„๋‹Œ, ๊ตฌ์กฐ์  ์œ ์‚ฌ๋„๋ฅผ ์ธก์ •ํ•˜๊ธฐ ์œ„ํ•œ ๋‹จ๊ณ„์ž…๋‹ˆ๋‹ค. ๋”ฐ๋ผ์„œ ๋‘ RDM ์‚ฌ์ด์˜ ์œ ์‚ฌ๋„๋ฅผ ์ธก์ •ํ•  ๋•Œ, Spearmanโ€™s rank correlation์„ ์ด์šฉํ•ด ๋‘ ํ–‰๋ ฌ์˜ ์œ ์‚ฌ๋„์ธ SSC๋ฅผ ๊ณ„์‚ฐํ•ฉ๋‹ˆ๋‹ค.

\[\begin{equation} SSC = 1 - {6\cdot \sum d_i^2 \over n(n^2-1)} \end{equation}\]

SSC๊ฐ’์ด ๋†’์„์ˆ˜๋ก, ์–ธ๋Ÿฌ๋‹ ํ›„ ๋ชจ๋ธ๊ณผ Retrain ๋ชจ๋ธ์˜ ๊ตฌ์กฐ๊ฐ€ ์œ ์‚ฌํ•จ์„ ์˜๋ฏธํ•ฉ๋‹ˆ๋‹ค.

๊ธฐ์กด retain-free ์–ธ๋Ÿฌ๋‹ ์•Œ๊ณ ๋ฆฌ์ฆ˜์˜ SSC ๊ฒฐ๊ณผ. CWRU ๋ฒ ์–ด๋ง ๊ณต์ • ๋ฐ์ดํ„ฐ๋กœ ์‚ฌ์ „ ํ•™์Šต๋œ ๋‹ค์„ฏ ๊ฐ€์ง€ ๋”ฅ๋Ÿฌ๋‹ ๋ชจ๋ธ์„ ๊ธฐ์กด retain-free ์–ธ๋Ÿฌ๋‹ ์•Œ๊ณ ๋ฆฌ์ฆ˜๊ณผ ARU๋กœ ์–ธ๋Ÿฌ๋‹ํ•œ ๊ฒฐ๊ณผ๋ฅผ ๋‚˜ํƒ€๋ƒ…๋‹ˆ๋‹ค. ๊ธฐ์กด retain-free ์–ธ๋Ÿฌ๋‹ ์•Œ๊ณ ๋ฆฌ์ฆ˜์˜ ๊ฒฝ์šฐ SSC ์ง€ํ‘œ๊ฐ€ ๋งค์šฐ ๋‚ฎ์•„, Retrain ๋ชจ๋ธ, ์ฆ‰ ์‚ฌ์ „ํ•™์Šต๋œ ๋ชจ๋ธ๊ณผ์˜ retain ๋ฐ์ดํ„ฐ์— ๋Œ€ํ•œ ๋ชจ๋ธ ๊ตฌ์กฐ์˜ ์œ ์‚ฌ๋„๊ฐ€ ๋‚ฎ์Œ์„ ์ •๋Ÿ‰์ ์œผ๋กœ ํ™•์ธํ•  ์ˆ˜ ์žˆ์Šต๋‹ˆ๋‹ค.

์ ๋Œ€์  ์˜ˆ์ œ ์ƒ์„ฑ

retain-free ํ™˜๊ฒฝ์—์„œ๋Š” \(\mathcal{D_R}\)์— ์ ‘๊ทผํ•  ์ˆ˜ ์—†์Šต๋‹ˆ๋‹ค. ๋”ฐ๋ผ์„œ ์‚ฌ์ „ํ•™์Šต๋œ ๋ชจ๋ธ \(f\)์™€ forget ๋ฐ์ดํ„ฐ \(\mathcal{D_F}\)๋งŒ์œผ๋กœ retain ๋ฐ์ดํ„ฐ์™€ ๋น„์Šทํ•œ \(\mathcal{\tilde{D}_R}\)๋ฅผ ์ƒ์„ฑํ•ด์•ผ ํ•ฉ๋‹ˆ๋‹ค.

๊ธฐ์กด retain-free ์–ธ๋Ÿฌ๋‹ ์•Œ๊ณ ๋ฆฌ์ฆ˜ ์ค‘ AMUN์€ \(\mathcal{D_F}\)์— ๋Œ€ํ•ด untargeted adversarial attack์„ ์ˆ˜ํ–‰ํ•ด ์˜ค๋ถ„๋ฅ˜๊ฐ€ ๋ฐœ์ƒํ•  ๋•Œ๊นŒ์ง€ ์ž„์˜์˜ ๋ฐฉํ–ฅ์œผ๋กœ ๋…ธ์ด์ฆˆ์˜ ๋ฐ˜๊ฒฝ์„ ์ฆ๊ฐ€์‹œํ‚ต๋‹ˆ๋‹ค.

๊ทธ๋Ÿฌ๋‚˜ PHM ๋„๋ฉ”์ธ์—์„œ ์ด ๋ฐฉ์‹์œผ๋กœ ์–ธ๋Ÿฌ๋‹์„ ์ˆ˜ํ–‰ํ•  ๊ฒฝ์šฐ, ๋ชจ๋ธ์ด ํŠน์ • ํด๋ž˜์Šค ๋ฐฉํ–ฅ์œผ๋กœ ๋ถ•๊ดด๋˜๋Š” embedding collapse ํ˜„์ƒ์ด ๋ฐœ์ƒํ•ฉ๋‹ˆ๋‹ค.

๋ณธ ๋…ผ๋ฌธ์€ ์ด๋ฅผ ํ•ด๊ฒฐํ•˜๊ธฐ ์œ„ํ•ด targeted adversarial attack์„ ์ˆ˜ํ–‰ํ•ฉ๋‹ˆ๋‹ค.

๋จผ์ €, forget ์ƒ˜ํ”Œ \((x_f, y_f)\in\mathcal{D_F}\)์— ๋Œ€ํ•ด ์›๋ž˜ ํด๋ž˜์Šค \(y_f\)๋ฅผ ์ œ์™ธํ•œ ํด๋ž˜์Šค ์ค‘ softmax ํ™•๋ฅ ์ด ๊ฐ€์žฅ ๋†’์€ ํด๋ž˜์Šค \(\tilde{y}\)๋ฅผ ๋‹ค์Œ๊ณผ ๊ฐ™์ด ์ •์˜ํ•ฉ๋‹ˆ๋‹ค :

\[\begin{equation} \tilde{y} = \arg\max_{f (x_f)_k \neq y_f} f (x_f)_k, \end{equation}\]

์ด๋Š” ๋ชจ๋ธ์ด ํ•ด๋‹น ์ƒ˜ํ”Œ์„ ๊ฐ€์žฅ ํ˜ผ๋™ํ•˜๋Š” ํด๋ž˜์Šค ๋ฐฉํ–ฅ์„ ์˜๋ฏธํ•ฉ๋‹ˆ๋‹ค. ์ฆ‰, ์ž„์˜์˜ ๋ฐฉํ–ฅ์ด ์•„๋‹Œ ๋ชจ๋ธ์ด ๊ฐ€์žฅ ํ˜ผ๋™ํ•˜๋Š” ํด๋ž˜์Šค ๋ฐฉํ–ฅ์œผ๋กœ์˜ targeted update๋ฅผ ์ˆ˜ํ–‰ํ•ฉ๋‹ˆ๋‹ค.

์ดํ›„, \(\mathcal{\tilde{D}_R}\)๋Š” PGD๋ฅผ ๋ณ€ํ˜•ํ•œ ๋‹ค์Œ๊ณผ ๊ฐ™์€ ๋ฐ˜๋ณต๋œ ์—…๋ฐ์ดํŠธ๋ฅผ ํ†ตํ•ด ์ƒ์„ฑ๋ฉ๋‹ˆ๋‹ค :

\[\begin{equation} x^{t+1} = \Pi_{\mathcal{B}_\epsilon} \Big( (x^t + \zeta^t) + \alpha \cdot \operatorname{sign}\Big( \nabla_{x^t + \zeta^t} \big[ \mathcal{L}(f(x^t + \zeta^t; \theta), \tilde{y}) \big] \Big) \Big), \end{equation}\]

์—ฌ๊ธฐ์„œ \(t\)๋Š” ๊ณต๊ฒฉ์˜ ๋ฐ˜๋ณต ๋‹จ๊ณ„, \(\Pi_{\mathcal{B}_\epsilon}\)๋Š” ๋…ธ์ด์ฆˆ๊ฐ€ ๋ฐ˜๊ฒฝ \(\epsilon\)์„ ๋„˜์ง€ ์•Š๋„๋ก \(\mathcal{l}_{\inf}\)-ball์œผ๋กœ ํˆฌ์˜ํ•˜๋Š” ์—ฐ์‚ฐ, ๊ทธ๋ฆฌ๊ณ  \(\zeta^t\)๋Š” ๊ฐ ๊ณต๊ฒฉ ๋‹จ๊ณ„๋งˆ๋‹ค ์ฃผ์ž…๋˜๋Š” ์ž‘์€ ํ™•๋ฅ ์  ๋…ธ์ด์ฆˆ์ž…๋‹ˆ๋‹ค.

ํ•ต์‹ฌ์€ per-step stochasticity์ž…๋‹ˆ๋‹ค. ๊ฐ ๋‹จ๊ณ„๋งˆ๋‹ค ์ž‘์€ ๋…ธ์ด์ฆˆ๋ฅผ ์ถ”๊ฐ€ํ•ด ๊ตญ์†Œ ์†์‹ค ์ง€ํ˜•์„ ๋ถ€๋“œ๋Ÿฝ๊ฒŒ ๋งŒ๋“ค๊ณ , ๋‹จ์ผ ๋ฐฉํ–ฅ์œผ๋กœ์˜ ๊ณผ๋„ํ•œ ์ˆ˜๋ ด์„ ๋ฐฉ์ง€ํ•ฉ๋‹ˆ๋‹ค.

๊ฒฐ๊ณผ์ ์œผ๋กœ ์ƒ์„ฑ๋œ \((\tilde{x}, \tilde{y})\)๋Š” ์–ธ๋Ÿฌ๋‹ ์ „ ๋ชจ๋ธ์˜ manifold ๊ทผ์ฒ˜์— ์œ„์น˜ํ•˜๋ฉด์„œ๋„ retain ํด๋ž˜์Šค์˜ ๊ธฐํ•˜ํ•™์  ๊ตฌ์กฐ์™€ ์ •๋ ฌ๋œ retain-like ์ƒ˜ํ”Œ๋กœ์„œ ์ž‘๋™ํ•ฉ๋‹ˆ๋‹ค.

๋ชจ๋ธ ๊ตฌ์กฐ ๋ณด์กด์„ ์œ„ํ•œ ์–ธ๋Ÿฌ๋‹ ์•Œ๊ณ ๋ฆฌ์ฆ˜

ARU์˜ ๋ชฉํ‘œ๋Š” retain ๋ฐ์ดํ„ฐ์— ์ง์ ‘ ์ ‘๊ทผํ•˜์ง€ ์•Š๊ณ ๋„ ๋ชจ๋ธ์˜ ๊ตฌ์กฐ๋ฅผ ๋ณด์กดํ•˜๋ฉฐ forget ๋ฐ์ดํ„ฐ์˜ ์ •๋ณด๋ฅผ ์ œ๊ฑฐํ•˜๋Š” ๊ฒƒ์ž…๋‹ˆ๋‹ค. ์ด๋ฅผ ๋‹ค์Œ๊ณผ ๊ฐ™์ด ๊ณต์‹ํ™”ํ•  ์ˆ˜ ์žˆ์Šต๋‹ˆ๋‹ค :

\[\begin{equation} \mathcal{L}_{ARU}(x_f, \tilde{x}, y;\alpha) = \mathcal{L}_\texttt{SDA}(x_f, \tilde{x};\alpha) + \mathcal{L}_\texttt{CE}(\tilde{x}, y) \end{equation}\]

์œ„ ์†์‹คํ•จ์ˆ˜๋Š” ๋‘ ๊ฐ€์ง€ ๊ตฌ์„ฑ ์š”์†Œ๋ฅผ ๊ฐ€์ง‘๋‹ˆ๋‹ค :

1. ์–ธ๋Ÿฌ๋‹ ํ•ญ : \(\mathcal{L}_\texttt{SDA}\)๋Š” forget ์ƒ˜ํ”Œ \(x_f\)๋ฅผ ๊ธฐ์กด forget ํด๋ž˜์Šค ์ค‘์‹ฌ์—์„œ ๋ฉ€์–ด์ง€๊ฒŒ ํ•˜๊ณ , retain-like ๋ฐ์ดํ„ฐ์˜ ์ค‘์‹ฌ ๋ฐฉํ–ฅ์œผ๋กœ ์ด๋™ํ•˜๋„๋ก ์œ ๋„ํ•ฉ๋‹ˆ๋‹ค.

SDA ์†์‹ค ํ•จ์ˆ˜์˜ ๊ธฐํ•˜ํ•™์  ํ•ด์„. forget ๋ฐ์ดํ„ฐ๋ฅผ ์›๋ž˜ forget ๋ฐ์ดํ„ฐ์˜ ์˜๋ฏธ ์ค‘์‹ฌ์—์„œ ๋ฉ€์–ด์ง€๋„๋ก ์œ ๋„ํ•˜๊ณ , retain-like ๋ฐ์ดํ„ฐ์˜ ์˜๋ฏธ ์ค‘์‹ฌ์œผ๋กœ ๊ฐ€๊นŒ์›Œ์ง€๋„๋ก ์œ ๋„ํ•ฉ๋‹ˆ๋‹ค. ๋ชจ๋ธ์˜ ๊ธฐ์กด ๊ตฌ์กฐ๋Š” ์œ ์ง€ํ•˜๋ฉฐ forget ๋ฐ์ดํ„ฐ์˜ ์ •๋ณด๋ฅผ ์žŠ๋„๋ก ์œ ๋„ํ•ฉ๋‹ˆ๋‹ค.

2. ๋ชจ๋ธ ๊ตฌ์กฐ ๋ณด์กด ํ•ญ : \(\mathcal{L}_\texttt{CE}\)๋Š” retain-like ๋ฐ์ดํ„ฐ \(\mathcal{\tilde{D}_R}\)๋กœ ์žฌํ•™์Šตํ•ด ๋ชจ๋ธ์˜ ๊ตฌ์กฐ๋ฅผ ๋ณด์กดํ•ฉ๋‹ˆ๋‹ค.

์ฆ‰, ๋‘ ํ•ญ์„ ๊ฒฐํ•ฉํ•จ์œผ๋กœ์จ \(\mathcal{L}_\texttt{SDA}\)๋ฅผ ํ†ตํ•œ forget ๋ฐ์ดํ„ฐ ๋ถ„๋ฆฌ์™€ \(\mathcal{L}_\texttt{CE}\)๋ฅผ ํ†ตํ•œ ๋ชจ๋ธ์˜ ๊ตฌ์กฐ ๋ณด์กด์„ ๋™์‹œ์— ๋‹ฌ์„ฑํ•ฉ๋‹ˆ๋‹ค.

embedding space์—์„œ์˜ ์„ฑ๋Šฅ

ARU๋ฅผ ์„ธ ๊ฐ€์ง€ ๊ณต์ • ๋ฐ์ดํ„ฐ์…‹์—์„œ ํ‰๊ฐ€ํ–ˆ์Šต๋‹ˆ๋‹ค :

1. CWRU ๋ฐ์ดํ„ฐ์…‹ : ๋ฒ ์–ด๋ง ๊ฒฐํ•จ ํƒ์ง€์—์„œ ๋„๋ฆฌ ์‚ฌ์šฉ๋˜๋Š” ๋ฐ์ดํ„ฐ์…‹์ž…๋‹ˆ๋‹ค. CWRU ๋ฐ์ดํ„ฐ์…‹์€ ๊ฒฐํ•จ ๋ถ€์œ„์— ๋”ฐ๋ผ ์ •์ƒ, ๋ณผ ๊ฒฐํ•จ, ๋‚ด๋ฅœ ๊ฒฐํ•จ, ์™ธ๋ฅœ ๊ฒฐํ•จ์˜ ๋„ค ๊ฐ€์ง€ ์ƒํƒœ๋กœ ๋‚˜๋‰˜๋ฉฐ, ๊ฐ ์ƒํƒœ์˜ ๊ณต์ • ์ฃผํŒŒ์ˆ˜๋ฅผ ์ƒ˜ํ”Œ๋กœ ์ˆ˜์ง‘ํ•œ ๋ฐ์ดํ„ฐ์…‹์ž…๋‹ˆ๋‹ค.

2. PRivate Bearing Dataset(PRBD) ๋ฐ์ดํ„ฐ์…‹ : ๋ณธ ์—ฐ๊ตฌ์—์„œ ์ง์ ‘ ๊ตฌ์ถ•ํ•œ ๋ฒ ์–ด๋ง ๊ฒฐํ•จ ์ง„๋‹จ ๋ฐ์ดํ„ฐ์…‹์ž…๋‹ˆ๋‹ค. CWRU ๋ฐ์ดํ„ฐ์…‹๊ณผ ๋งˆ์ฐฌ๊ฐ€์ง€๋กœ ๋„ค ๊ฐ€์ง€์˜ ๊ฒฐํ•จ ์ƒํƒœ๋ฅผ ๊ฐ€์ง‘๋‹ˆ๋‹ค.

3. CWRU + PRBD ๋ฐ์ดํ„ฐ์…‹ : ๊ณต๊ฐœ๋œ ๋ฐ์ดํ„ฐ์…‹๊ณผ ๊ณต๊ฐœ๋˜์ง€ ์•Š์€ ๋ฐ์ดํ„ฐ์…‹์„ ๋™์‹œ์— ์‚ฌ์šฉํ•ด ํ˜„์‹ค์ ์ธ retain-free ํ™˜๊ฒฝ์„ ๊ฐ€์ •ํ–ˆ์Šต๋‹ˆ๋‹ค.

๋˜ํ•œ MLP, TCN, CNN1D, Transformer์˜ ๋„ค ๊ฐ€์ง€ ๋”ฅ๋Ÿฌ๋‹ ๋ชจ๋ธ์„ ์‚ฌ์šฉํ•ด ๋‹ค์–‘ํ•œ PHM ํ™˜๊ฒฝ์„ ๊ตฌํ˜„ํ–ˆ์Šต๋‹ˆ๋‹ค.

์„ธ ๊ฐ€์ง€ ๋ฐ์ดํ„ฐ์…‹์— ๋Œ€ํ•ด ์‹คํ—˜ํ•œ ๊ฒฐ๊ณผ, ARU๋Š” ๋ชจ๋“  ๋ฐ์ดํ„ฐ์…‹์—์„œ ๊ฐ€์žฅ ๋‚ฎ์€ \(\Delta\)Acc๋ฅผ ๋‹ฌ์„ฑํ•˜๋ฉฐ Retrain๊ณผ ๊ฐ€์žฅ ๋น„์Šทํ•œ ์ •ํ™•๋„๋ฅผ ๊ธฐ๋กํ–ˆ์Šต๋‹ˆ๋‹ค.

์˜ค๋ฅธ์ชฝ๋ถ€ํ„ฐ ์ˆœ์„œ๋Œ€๋กœ CWRU, PRBD, CWRU+PRBD ์— ๋Œ€ํ•œ ๊ฒฐ๊ณผ. ARU๋Š” ๋ชจ๋“  ๋ฐ์ดํ„ฐ์…‹์—์„œ ๊ฐ€์žฅ ๋‚ฎ์€ $$\Delta$$Acc๋ฅผ ๋‹ฌ์„ฑํ•˜๋ฉฐ Retrain์— ๊ฐ€์žฅ ๊ฐ€๊นŒ์šด ์„ฑ๋Šฅ์„ ๋ณด์ž…๋‹ˆ๋‹ค.

๊ฒฐ๋ก 


์‚ฐ์—… ํ˜„์žฅ์—์„œ๋Š” ์ธ๊ณต์ง€๋Šฅ์„ ํ™œ์šฉํ•œ ๋ชจ๋ธ ์ง„๋‹จ ์‹œ์Šคํ…œ์ด ๋„๋ฆฌ ํ™œ์šฉ๋˜๊ณ  ์žˆ์Šต๋‹ˆ๋‹ค. ์ด๋•Œ ๋ฒ•์ , ๋ณด์•ˆ์  ์ด์œ ๋กœ ํŠน์ • ๋ฐ์ดํ„ฐ๋ฅผ ๋ชจ๋ธ์—์„œ ์‚ญ์ œํ•ด์•ผ ํ•˜๋Š” ์ƒํ™ฉ์—์„œ ๋ชจ๋ธ์ด ํ•ด๋‹น ๋ฐ์ดํ„ฐ๋ฅผ โ€œ์žŠ๋„๋กโ€ ๋งŒ๋“ค์–ด์•ผ ํ•ฉ๋‹ˆ๋‹ค. ํŠนํžˆ ์‹ค์ œ ์‚ฐ์—… ํ˜„์žฅ์—์„œ๋Š” ์ „์ฒด ํ•™์Šต ๋ฐ์ดํ„ฐ์— ๋‹ค์‹œ ์ ‘๊ทผํ•˜๊ธฐ ์–ด๋ ค์šด ๊ฒฝ์šฐ๊ฐ€ ๋งŽ์•„, ์ด๋ฅผ ํ•ด๊ฒฐํ•  ์ˆ˜ ์žˆ๋Š” ์ƒˆ๋กœ์šด ๋จธ์‹  ์–ธ๋Ÿฌ๋‹ ์ ‘๊ทผ๋ฒ•์ด ์š”๊ตฌ๋ฉ๋‹ˆ๋‹ค.

๋ณธ ๋…ผ๋ฌธ์—์„œ๋Š” ์ด๋Ÿฌํ•œ ๋ฌธ์ œ๋ฅผ ํ•ด๊ฒฐํ•˜๊ธฐ ์œ„ํ•ด Adversarial Retain-free Unlearning(ARU) ํ”„๋ ˆ์ž„์›Œํฌ๋ฅผ ์ œ์•ˆํ•ฉ๋‹ˆ๋‹ค. ARU๋Š” ๋ชจ๋ธ์ด ํ•™์Šตํ•œ ์ „์ฒด ๋ฐ์ดํ„ฐ ๋Œ€์‹ , ์‚ญ์ œํ•  ๋ฐ์ดํ„ฐ๋งŒ์„ ์ด์šฉํ•ด ํ•„์š”ํ•œ ์ •๋ณด๋Š” ์œ ์ง€ํ•˜๋ฉด์„œ๋„ ์‚ญ์ œํ•  ๋ฐ์ดํ„ฐ๋ฅผ ํšจ๊ณผ์ ์œผ๋กœ ์ œ๊ฑฐํ•ฉ๋‹ˆ๋‹ค. ๋‹ค์–‘ํ•œ ๋ฐ์ดํ„ฐ์…‹์„ ์ด์šฉํ•ด ARU๊ฐ€ ๋ชจ๋ธ์˜ ๊ตฌ์กฐ์™€ ์ •ํ™•๋„๋ฅผ ์œ ์ง€ํ•˜๋ฉด์„œ๋„ ํŠน์ • ํด๋ž˜์Šค๋ฅผ ์•ˆ์ •์ ์œผ๋กœ ๋ถ„๋ฆฌํ•ด๋ƒ„์„ ํ™•์ธํ–ˆ์Šต๋‹ˆ๋‹ค.

๋จธ์‹  ์–ธ๋Ÿฌ๋‹์€ ์•ž์œผ๋กœ ํ”„๋ผ์ด๋ฒ„์‹œ ๋ณดํ˜ธ์™€ ์‚ฐ์—… ๊ทœ์ œ ์ค€์ˆ˜๋ฅผ ์œ„ํ•ด ๋”์šฑ ์ค‘์š”ํ•ด์งˆ ๊ธฐ์ˆ ์ž…๋‹ˆ๋‹ค. ๋ณธ ์—ฐ๊ตฌ๊ฐ€ ์‹ค์ œ PHM ํ™˜๊ฒฝ์—์„œ ์ ์šฉ ๊ฐ€๋Šฅํ•œ ์–ธ๋Ÿฌ๋‹ ๊ธฐ์ˆ ์— ๊ธฐ์—ฌํ•˜๊ธฐ๋ฅผ ๋ฐ”๋ž๋‹ˆ๋‹ค.

์—ฐ๊ตฌ์‹ค ๊ด€๋ จ ๋…ผ๋ฌธ๋“ค

  • Unlearning-Aware Minimization [NeurIPS 2025] | [Paper] | [Article]| [Code]
  • Evaluating practical adversarial robustness of fault diagnosis systems via spectrogram-aware ensemble method [EAAI] | [Paper] | [Article]
  • Black-box adversarial examples via frequency distortion against fault diagnosis systems [EAAI] | [Paper]
  • Generating transferable adversarial examples for speech classification [Pattern Recognition] | [Paper]