EveryDayVLA: A Vision-Language-Action Model for Affordable Robotic Manipulation