sji

  • 홈
  • 태그
  • 방명록

Knowledge Distiliation 1

DistilBERT (a distilled version of BERT: smaller, faster, cheaper and lighter)

arxiv.org/abs/1910.01108 DistilBERT, a distilled version of BERT: smaller, faster, cheaper and lighter As Transfer Learning from large-scale pre-trained models becomes more prevalent in Natural Language Processing (NLP), operating these large models in on-the-edge and/or under constrained computational training or inference budgets remains challenging. In t arxiv.org Knowledge Distillation에 대해 훑..

DL&ML/papers 2021.04.23
1
더보기
  • 분류 전체보기 (87)
    • intro. (0)
    • DL&ML (38)
      • concept (19)
      • papers (9)
      • code.data.tips (10)
    • http&server (7)
    • dist. system (6)
    • docker (5)
    • algorithm (24)
    • design pattern (1)
    • architecture (1)
    • trouble shooting (3)
    • better way (1)
    • problems (0)

Tag

summarization, string, Docker, queue, Likelihood, MLE, gunicorn, seq2seq, gaussian mixture, React Native, Dynamic Programming, docker-compose, Graph, stack, nginx, CrossEntropy, tree, flask, dp, Zookeeper,

최근글과 인기글

  • 최근글
  • 인기글
250x250

Calendar

«   2025/05   »
일 월 화 수 목 금 토
1 2 3
4 5 6 7 8 9 10
11 12 13 14 15 16 17
18 19 20 21 22 23 24
25 26 27 28 29 30 31

방문자수Total

  • Today :
  • Yesterday :

Designed by Tistory.

티스토리툴바