• Skip to primary navigation
  • Skip to content
  • Skip to footer
Machine Learning / Deep Learning Study
      Hong Yong Man

      Hong Yong Man

      I am an amazing person.

      • Republic of Korea
      • Email
      • Facebook
      • LinkedIn
      • GitHub

      An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale

      less than 1 minute read

      An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale

      - Vision Transformer(ViT) Method

      • 기존 자연어 처리에서 사용되던 Transformer라는 모델을 Vision에 적용
      • Image Patch -> Transformer Encoder

      Tags: ViT

      Categories: Paper

      Updated: February 19, 2021

      Twitter Facebook LinkedIn
      Previous Next

      You May Also Enjoy

      RemixMatch, FixMatch

      less than 1 minute read

      Semi-Supervised Learning Supervised Learning은 Labeled data만을 이용하여 만들어진다. But, 현실세계에서는 Unlabeled data가 훨씬 많고 Labeled data와 Unlabeled data를 같이 학습시키는 것이 Se...

      Multi Task, Multi Modal

      less than 1 minute read

      Multi Task & Multi Modal Single Task : Input X —> Output Y (ex. Object Classification) Multi Task : Input X —> Output Y1, Y2, Y3 (ex. Object ...

      Deep Compression

      less than 1 minute read

      Deep Compression: Compressing Deep Neural Networks with Pruning, Trained Quantization and Huffman Coding

      Auto-ML

      less than 1 minute read

      Auto ML - Machine Learning 모델 학습에서 가장 중요한 것은 올바른 Data의 확보 그리고 Hyperparameter Tunning 일 것이다. Auto-ML은 이 Hyperparameter Tunning을 자동으로 해주는 방법이다.

      • Feed
      © 2021 hongym7. Powered by Jekyll & Minimal Mistakes.