Initialization Matters: On the Benign Overfitting of Two-Layer ReLU CNN with Fully Trainable Layers