Scaling Offline Model-Based RL via Jointly-Optimized World-Action Model Pretraining