Strategies to Mitigate Hallucinations in Large Language Models (LLMs)