RAG ARCHITECTURE AND AGENTS

Authors

DOI:

https://doi.org/10.31510/infa.v21i2.1995

Keywords:

RAG architecture, Agents, ReAct, Artificial intelligence, LLM

Abstract

This article proposes a bibliographical survey to provide a comprehensive overview of the RAG (Retrieval-Augmented Generation) Architecture and the application of agents in the current context of artificial intelligence (AI). The RAG architecture combines information retrieval techniques with generation models, providing a hybrid approach that improves the efficiency and accuracy of responses generated by AI systems. Proper architecture planning is crucial to the success of RAG-based systems, as it directly affects the agents' ability to handle large volumes of data and provide relevant answers. In this article, the fundamental principles of the RAG Architecture are explored, as well as its practical applications, advantages, disadvantages, and implementation challenges. In addition, a critical analysis of the contributions and emerging trends in the literature on the subject was carried out. The results of the research indicate that, although the RAG Architecture offers significant advances in the interaction between agents and data, there are still important challenges to be overcome for its large-scale adoption. This study seeks to contribute to understanding these challenges and provide insights for future research and implementation.

Downloads

Download data is not yet available.

References

BOSTROM, N.; YUDKOWSKY, E. The Ethics of Artificial Intelligence. 2014. DOI: https://doi.org/10.1017/CBO9781139046855.020

BROWN, T. B. et al. Language Models are Few-Shot Learners. 28 maio 2020.

FINARDI, P. et al. The Chronicles of RAG: The Retriever, the Chunk and the Generator. 15 jan. 2024.

GAO, Y. et al. Retrieval-Augmented Generation for Large Language Models: A Survey. 18 dez. 2023.

GOODFELLOW, I.; BENGIO, Y.; COURVILLE, A. Deep Learning. MIT Press, 2016.

JOHNSON, J.; DOUZE, M.; JÉGOU, H. Billion-scale similarity search with GPUs. 28 fev. 2017.

KAPA. Kapa AI Documentation. Disponível em: <https://docs.kapa.ai>. Acesso em: 27 nov. 2024.

KARPUKHIN, V. et al. Dense Passage Retrieval for Open-Domain Question Answering. 10 abr. 2020. DOI: https://doi.org/10.18653/v1/2020.emnlp-main.550

LANGCHAIN. LangChain Documentation. Disponível em: <https://www.langchain.com/>. Acesso em: 11 ago. 2024.

LEWIS, P. et al. Retrieval-Augmented Generation for Knowledge-Intensive NLP Tasks. 22 maio 2020.

LLAMAINDEX. LlamaIndex Documentation. Disponível em: <https://www.llamaindex.ai/>. Acesso em: 11 ago. 2024.

NG, A. What Artificial Intelligence Can and Can’t Do Right Now. Disponível em: <https://hbr.org/2016/11/what-artificial-intelligence-can-and-cant-do-right-now>. Acesso em: 11 ago. 2024.

NGUYEN, Z. et al. Enhancing Q&A with Domain-Specific Fine-Tuning and Iterative Reasoning: A Comparative Study. 10 abr. 2024.

OVADIA, O. et al. Fine-Tuning or Retrieval? Comparing Knowledge Injection in LLMs. 30 jan. 2024. DOI: https://doi.org/10.18653/v1/2024.emnlp-main.15

RUSSEL, S.; NORVIG, P. Artificial Intelligence: A Modern Approach. 2016.

SOUDANI, H. et al. Fine Tuning vs. Retrieval Augmented Generation for Less Popular Knowledge. 27 set. 2024. DOI: https://doi.org/10.1145/3673791.3698415

TOUVRON, H. et al. LLaMA: Open and Efficient Foundation Language Models. 27 fev. 2023.

WANG, J. et al. Milvus: A Purpose-Built Vector Data Management System. Proceedings of the 2021 International Conference on Management of Data. New York, NY, USA: ACM, 9 jun. 2021. DOI: https://doi.org/10.1145/3448016.3457550

YAO, S. et al. ReAct: Synergizing Reasoning and Acting in Language Models. 5 out. 2022.

ZENG, S. et al. The Good and The Bad: Exploring Privacy Issues in Retrieval-Augmented Generation (RAG). 23 fev. 2024. DOI: https://doi.org/10.18653/v1/2024.findings-acl.267

Published

2025-09-17

Issue

Section

Tecnologia em Informática

How to Cite

SOUZA, Antonio Henrique Nascimento Machado de; SOUZA, Isadora Mota de; DIAS, Jônatas Cerqueira. RAG ARCHITECTURE AND AGENTS. Revista Interface Tecnológica, Taquaritinga, SP, v. 21, n. 2, p. 15–26, 2025. DOI: 10.31510/infa.v21i2.1995. Disponível em: https://revista.fatectq.edu.br/interfacetecnologica/article/view/1995. Acesso em: 5 dec. 2025.