일 | 월 | 화 | 수 | 목 | 금 | 토 |
---|---|---|---|---|---|---|
1 | 2 | 3 | ||||
4 | 5 | 6 | 7 | 8 | 9 | 10 |
11 | 12 | 13 | 14 | 15 | 16 | 17 |
18 | 19 | 20 | 21 | 22 | 23 | 24 |
25 | 26 | 27 | 28 | 29 | 30 | 31 |
- error: mkl-service + intel(r)
- domain-adapted pre-training
- CPT
- cross-document attention
- langchain
- continued pre-training
- prompt
- instruct pre-training
- full fine-tuning (fft)
- glibcxx
- ubuntu
- transformer
- instruct-pt
- diffusion
- backbone
- continual pre-training
- instruction tuning (it)
- instruction tuning
- ViT
- PEFT
- gemma2
- Fine-tuning
- Mac
- continued pre-train (cpt)
- sfttrainer
- Text-to-Image
- lora+
- Lora
- llm
- llm tuning
- Today
- Total
목록backbone (2)
꾸준하게

CVPR 2023 [paper] [code]Wenhai Wang, Jifeng Dai, Zhe Chen, Zhenhang Huang, Zhiqi Li, Xizhou Zhu, Xiaowei Hu, Tong Lu, Lewei Lu, Hongsheng Li, Xiaogang Wang, Yu QiaoShanghai AI Laboratory | Tsinghua University | Nanjing University | SenseTime Research | The Chinese University of Hong KongSubmitted on 10 Nov 2022last revised 17 Apr 2023 (v4) Abstract저자는 근래 나온 ViT 모델과 달리 CNN은 아직 초기 상태에 머물러 있다고 주장하며..

" data-ke-type="html">HTML 삽입미리보기할 수 없는 소스ICLR 2021 (Oral) paper code Alexey Dosovitskiy, Lucas Beyer, Alexander Kolesnikov, Dirk Weissenborn, Xiaohua Zhai, Thomas Unterthiner, Mostafa Dehghani, Matthias Minderer, Georg Heigold, Sylvain Gelly, Jakob Uszkoreit, Neil HoulsbyGoogle Research, Brain Team 22 Oct 2020 (v1), 3 Jun 2021 (v2)Abstract주로 CNN을 이용해서 Vision Task를 해결하던 중 처음으로 Transformer를 이용한 논..