Seemingly Plausible Distractors in Multi-Hop Reasoning: Are Large Language Models Attentive Readers?