Optimizing Large Language Model (LLM) Inference for Arm CPUs