Currently, I'm exploring different methods for pre-training and downstream fine-tuning LLMs, mainly Masked Language Modelling, Sentence Pair Prediction, Next Word Prediction from context, and Multi-Task Training on decoder-only architectures like GPT2.
Actively looking for Summer 2024 internships in software engineering and machine learning
π Reach me at [email protected]