sji

  • 홈
  • 태그
  • 방명록

Knowledge Distiliation 1

DistilBERT (a distilled version of BERT: smaller, faster, cheaper and lighter)

arxiv.org/abs/1910.01108 DistilBERT, a distilled version of BERT: smaller, faster, cheaper and lighter As Transfer Learning from large-scale pre-trained models becomes more prevalent in Natural Language Processing (NLP), operating these large models in on-the-edge and/or under constrained computational training or inference budgets remains challenging. In t arxiv.org Knowledge Distillation에 대해 훑..

DL&ML/papers 2021.04.23
1
더보기
  • 분류 전체보기 (87)
    • intro. (0)
    • DL&ML (38)
      • concept (19)
      • papers (9)
      • code.data.tips (10)
    • http&server (7)
    • dist. system (6)
    • docker (5)
    • algorithm (24)
    • design pattern (1)
    • architecture (1)
    • trouble shooting (3)
    • better way (1)
    • problems (0)

Tag

CrossEntropy, summarization, gunicorn, stack, seq2seq, tree, gaussian mixture, MLE, Docker, Graph, flask, Likelihood, queue, React Native, Zookeeper, string, dp, docker-compose, Dynamic Programming, nginx,

최근글과 인기글

  • 최근글
  • 인기글
250x250

Calendar

«   2026/04   »
일 월 화 수 목 금 토
1 2 3 4
5 6 7 8 9 10 11
12 13 14 15 16 17 18
19 20 21 22 23 24 25
26 27 28 29 30

방문자수Total

  • Today :
  • Yesterday :

Designed by Tistory.

티스토리툴바