Custom GPT Implementation

Completed
2025

Developed a miniature GPT model implementing custom tokenization, self-attention, and transformer blocks. Designed neural architecture with positional embeddings, achieving coherent autoregressive text generation. Includes training pipeline, evaluation metrics, and sample generation scripts. Compared performance with open-source transformer models and visualized attention weights for interpretability.

Custom GPT Implementation - Image 1
About This Project

Developed a miniature GPT model implementing custom tokenization, self-attention, and transformer blocks. Designed neural architecture with positional embeddings, achieving coherent autoregressive text generation. Includes training pipeline, evaluation metrics, and sample generation scripts. Compared performance with open-source transformer models and visualized attention weights for interpretability.

Technologies Used
PyTorchPythonNumPyTransformers
Project Links
Nadipalli Jaswanth Portfolio - Full Stack Developer & AI Engineer