Approximating Shapley Explanations in Reinforcement Learning