Low-Rank Prune-And-Factorize for Language Model Compression