Width is Less Important than Depth in ReLU Neural Networks