Enhancing Computation Efficiency in Large Language Models through Weight and Activation Quantization