I created this notebook to better understand the inner workings of Bert. I followed a lot of tutorials to try to understand the architecture, but I was never able to really understand what was happening under the hood. For me it always helps to see the actual code instead of just simple abstract diagrams that a lot of times don’t match the actual implementation. If you’re like … [Read more...] about BERT Inner Workings
GPT2 For Text Classification Using Hugging Face Transformers
This notebook is used to fine-tune GPT2 model for text classification using Hugging Face transformers library on a custom dataset. Hugging Face is very nice to us to include all the functionality needed for GPT2 to be used in classification tasks. Thank you Hugging Face! I wasn’t able to find much information on how to use GPT2 for classification so I … [Read more...] about GPT2 For Text Classification Using Hugging Face Transformers
Fine-tune Transformers in PyTorch Using Hugging Face Transformers
This notebook is designed to use a pretrained transformers model and fine-tune it on a classification task. The focus of this tutorial will be on the code itself and how to adjust it to your needs. This notebook is using the AutoClasses from transformer by Hugging Face functionality. This functionality can guess a model’s configuration, tokenizer and … [Read more...] about Fine-tune Transformers in PyTorch Using Hugging Face Transformers
Pretrain Transformers Models in PyTorch Using Hugging Face Transformers
This notebook is used to pretrain transformers models using Hugging Face on your own custom dataset. What do I mean by pretrain transformers? The definition of pretraining is to train in advance. That is exactly what I mean! Train a transformer model to use it as a pretrained transformers model which can be used to fine-tune it on a specific … [Read more...] about Pretrain Transformers Models in PyTorch Using Hugging Face Transformers