Back Attention: Understanding and Enhancing Multi-Hop Reasoning in Large Language Models