Self-Augmented Preference Optimization: Off-Policy Paradigms for Language Model Alignment