[Paper reading] Dataset Condensation reading

2023. 9. 11. 00:11ใ†ArtificialIntelligence/PaperReading

 

 

 

Dataset Condensation with Gradient Matching (ICLR 2021)

 

 

 

Abstract

  • As the state-of-the-art machine learning methods in many fields rely on larger datasets, storing datasets and training models on them become significantly more expensive. This paper proposes a training set synthesis technique for data-efficient learning, called Dataset Condensation, that learns to condense large dataset into a small set of informative synthetic samples for training deep neural networks from scratch.
  • We formulate this goal as a gradient matching problem between the gradients of deep neural network weights that are trained on the original and our synthetic data. We rigorously evaluate its performance in several computer vision benchmarks and demonstrate that it significantly outperforms the state-of-the-art methods.
  • Finally we explore the use of our method in continual learning and neural architecture search and report promising gains when limited memory and computations are available.

 

 

 

Conclusion

  • In this paper, we propose a dataset condensation method that learns to synthesize a small set of informative images. We show that these images are significantly more data-efficient than the same number of original images and the ones produced by the previous method, and they are not architecture dependent, can be used to train different deep networks.
  • Once trained, they can be used to lower the memory print of datasets and efficiently train numerous networks which are crucial in continual learning and neural architecture search respectively.
  • For future work, we plan to explore the use of condensed images in more diverse and thus challenging datasets like ImageNet (Deng et al., 2009) that contain higher resolution images with larger variations in appearance and pose of objects, background.

 

 

data distillation vs data condensation

 

 

 

Question

  • Few-shot learning vs zero-shot learning
    • ๋ชจ๋ธ์„ ํ•™์Šตํ•˜๋Š” ๋ฐ ์‚ฌ์šฉ๋˜๋Š” ๋ฐ์ดํ„ฐ์˜ ์–‘์„ ๋‚˜ํƒ€๋‚ด๋Š” ์šฉ์–ด

    • few shot learning
      • ํ•œ ํด๋ž˜์Šค๋‹น ์ผ๋ถ€ ์ƒ˜ํ”Œ ์ด๋ฏธ์ง€๋งŒ ์‚ฌ์šฉํ•˜์—ฌ classification
      • ์ œํ•œ๋œ, ์ ์€ ์ˆ˜์˜ ์ƒ˜ํ”Œ ์ด๋ฏธ์ง€๋ฅผ ์‚ฌ์šฉํ•˜์—ฌ ๋ถ„๋ฅ˜ ์ž‘์—…์„ ์ˆ˜ํ–‰
      • Few shot learning refers to a machine learning methodology that can learn from a limited number of samples, less than 10.
      • The idea is that just as humans can quickly pick up new concepts with just a few examples, so can models. (Machine learning aims to bridge the gap between human-like learning capabilities.)

    • zero shot learning
      • ๋ผ๋ฒจ๋ง ๋˜์ง€ ์•Š์€ ์ƒˆ๋กœ์šด ํด๋ž˜์Šค์— ๋Œ€ํ•ด, ์ด์ „์— ํ•™์Šต๋œ ๋ชจ๋ธ์„ ์‚ฌ์šฉํ•˜์—ฌ ๋ถ„๋ฅ˜
      • ์ด์ „์— ํ•™์Šต๋œ ๋ชจ๋ธ์„ ์‚ฌ์šฉํ•˜์—ฌ ํ•™์Šต ๋ฐ์ดํ„ฐ์— ์—†๋Š” ์ƒˆ๋กœ์šด ํด๋ž˜์Šค๋ฅผ ์ธ์‹ํ•˜๊ณ  ๋ถ„๋ฅ˜ํ•  ์ˆ˜ ์žˆ๋‹ค.
      • ์ด๋•Œ ์ƒˆ๋กœ์šด class์— ๋Œ€ํ•œ ์ •๋ณด๋ฅผ input์œผ๋กœ ์ฃผ์–ด์•ผ ํ•œ๋‹ค.
      •  ZSL is a machine learning methodology that allows a model to learn to recognize a new class without ever seeing an example of it during training
      • model can be taught to recognize and classify it accurately using information about its attributes or characteristics without explicit training on the new class.

 

 

  • Dataset distillation vs Knowledge distillation vs Dataset condensation
    • Knowledge distillation
      • ํฐ ์•™์ƒ๋ธ” ๋ชจ๋ธ์˜ ์ง€์‹์„ ๋” ์ž‘์€ ๋ชจ๋ธ(compact network)๋กœ ์ „๋‹ฌํ•˜๋Š” ๋ฐฉ๋ฒ•
      • ๋Œ€๋Ÿ‰์˜ ๋ฐ์ดํ„ฐ๋ฅผ ์†Œ์ˆ˜์˜ ๋ฐ์ดํ„ฐ๋กœ ์••์ถ•ํ•˜๋Š” ์•„์ด๋””์–ด์—์„œ ์ถœ๋ฐœ
      • ์ „์ฒด ๋ฐ์ดํ„ฐ์˜ loss์™€ ์ฆ๋ฅ˜ ๋ฐ์ดํ„ฐ(_x)์˜ loss๊ฐ€ ์„œ๋กœ ์ตœ๋Œ€ํ•œ ๋‚ฎ๊ฒŒ ๋‚˜์˜ค๋Š” _x๋ฅผ ์ฐพ๋Š” ๋ฌธ์ œ๋กœ ๊ท€๊ฒฐ
        • Dataset distillation

 

 

 

์ฐธ๊ณ  ์ž๋ฃŒ ๋ฐ ์ฝ”๋“œ

https://github.com/VICO-UoE/DatasetCondensation

 

GitHub - VICO-UoE/DatasetCondensation: Dataset Condensation (ICLR21 and ICML21)

Dataset Condensation (ICLR21 and ICML21). Contribute to VICO-UoE/DatasetCondensation development by creating an account on GitHub.

github.com

https://velog.io/@nomaday/n-shot-learning

https://www.thedatahunt.com/en-insight/guide-for-few-shot-learning

 

Guide for few shot learning - concept, learning method, process

Few-shot learning requires less data for training, making it valuable in domains with limited or costly data collection.

www.thedatahunt.com