IEEE Access, vol.13, pp.30936-30951, 2025 (SCI-Expanded)
Large language model optimization for a particular application is both crucial and challenging in natural language processing. This paper compares two salient techniques for this Retrieve-augmented generation and fine-tuning along with a new hybrid method that combines both. In this work, we investigate the effectiveness of various methods using on SQuAD (the Stanford Question Answering Dataset), MS MARCO (Microsoft MAchine Reading Comprehension) and SQL CREATE TABLE statements. RAG is used because it enriches model responses with external data without much computational load during inference. Fine-tuning updates model parameters for better contextual accuracy. Our hybrid model balances these two techniques in terms of both accuracy and efficiency. While fine-tuning entails semantic precision, RAG is more resource efficient. The hybrid approach while it may not offer surpassing results over fine-tuning-offers a balanced solution in scenarios where the application demands both efficiency and accuracy. These findings represent the trade-off involved in the LLM optimization and offer scope for further studies and practical applications.