MoTVLA: A Vision-Language-Action Model with Unified Fast-Slow Reasoning