https://www.youtube.com/watch?v=BqqfQnyjmgg&list=PLo2EIpI_JMQvWfQndUesu0nPBAtZ9gP1o&index=4
Training from scratch --> Fine-tuning a pretrained model.
ImageNet is commnly used as a dataset for pretraining models in CV.
GPT-2 was pretrained on 40GB of internet text postsed by users on Reddit.
BERT was pretrained on the content of the English wikipedia and 11,000 unpublished books.
ImageNet content is mostly US and EU centred.
The tokenization pipeline (0) | 2023.09.14 |
---|---|
What happens inside the pipeline function? (TensorFlow) (0) | 2023.09.13 |
What happens inside the pipeline function? (PyTorch) (0) | 2023.09.13 |
The Transformer architecture (0) | 2023.09.13 |
The pipeline function (0) | 2023.09.13 |