SmilesT5: Domain-specific pretraining for molecular language models