Unboxing the Black Box: Mechanistic Interpretability for Algorithmic Understanding of Neural Networks