DeServe: Towards Affordable Offline LLM Inference via Decentralization