Redirecting to original paper in 30 seconds...
Click below to go immediately or wait for automatic redirect
📄 Abstract
Abstract: This paper reviews the state-of-the-art of large language models (LLM)
architectures and strategies for "complex" question-answering with a focus on
hybrid architectures. LLM based chatbot services have allowed anyone to grasp
the potential of LLM to solve many common problems, but soon discovered their
limitations for complex questions. Addressing more specific, complex questions
(e.g., "What is the best mix of power-generation methods to reduce climate
change ?") often requires specialized architectures, domain knowledge, new
skills, decomposition and multi-step resolution, deep reasoning, sensitive data
protection, explainability, and human-in-the-loop processes. Therefore, we
review: (1) necessary skills and tasks for handling complex questions and
common LLM limits to overcome; (2) dataset, cost functions and evaluation
metrics for measuring and improving (e.g. accuracy, explainability, fairness,
robustness, groundedness, faithfulness, toxicity...); (3) family of solutions
to overcome LLM limitations by (a) training and reinforcement (b)
hybridization, (c) prompting, (d) agentic-architectures (agents, tools) and
extended reasoning.
Authors (5)
Xavier Daull
Patrice Bellot
Emmanuel Bruno
Vincent Martin
Elisabeth Murisasco
Submitted
February 17, 2023