This repository demonstrates core deep learning engineering skills using PyTorch, built from first principles without relying on high-level abstractions.
The goal of this project is to show that I understand how neural networks actually work, not just how to use pre-built pipelines.
This project intentionally focuses on fundamentals that every AI engineer must master:
- Built neural networks by subclassing
torch.nn.Module - Explicitly defined layers and parameters
- Clear separation of model architecture and training logic
- Implemented forward pass using tensors and layers
- Used PyTorch autograd to compute gradients automatically
- Explained how gradients flow through the network
- Manual training loop with:
- Forward pass
- Loss computation
- Backward pass
- Optimizer step
- No training shortcuts or hidden abstractions
- Device-agnostic code (
CPU β GPU) - Safe
.to(device)usage - Seamless switching between CPU and CUDA (if available)
- Saved trained model weights using
state_dict - Reloaded models for inference or continued training
- Demonstrated reproducibility and checkpointing