Skip to main content

3 posts tagged with "rag"

View All Tags

Applied AI Series - RAG Better Results with Re-ranking

· 11 min read
Niko
Software Engineer @ Naver

Re-ranking in Retrieval-Augmented Generation (RAG) refines the documents retrieved in response to a user query, ensuring that only the most relevant and contextually appropriate ones are passed to the generation model. This step enhances response accuracy, handles ambiguity, and improves overall result quality by prioritizing the best matches for the query, ultimately leading to more precise and coherent AI-generated answers.

Applied AI Series - RAG Speedup LLMs with Document chunking

· 9 min read
Niko
Software Engineer @ Naver

Document chunking is crucial for optimizing Retrieval-Augmented Generation (RAG) systems by breaking large documents into smaller, manageable pieces, which significantly speeds up retrieval and enhances the relevance of results. In RAG, where information retrieval is followed by text generation, chunking allows the system to search and process only the most relevant sections of content, improving both efficiency and accuracy. This approach ensures faster retrieval times, better handling of long-form documents, and more precise generation by focusing on contextually meaningful chunks rather than entire documents, ultimately enhancing the overall performance of LLMs in real-time applications.

Applied AI Series - What is RAG?

· 5 min read
Niko
Software Engineer @ Naver

Welcome to the latest installment of our Applied AI Series, where we delve into cutting-edge technologies and explore how they are transforming industries and everyday tasks. In this edition, we’ll unravel the intricacies of RAG (Retrieve and Generate), a powerful framework that’s making waves in the field of artificial intelligence. Whether you're a tech enthusiast, a researcher, or someone curious about AI, this post will provide a clear understanding of RAG and its applications.