LLM-PQA

LLM-enhanced Prediction Query Answering

Conference Paper (2024)
Author(s)

Ziyu Li (TU Delft - Team Arjan Mol)

Wenjie Zhao (Student TU Delft)

A Katsifodimos (TU Delft - Data-Intensive Systems)

Rihan Hai (TU Delft - Web Information Systems)

Research Group
Team Arjan Mol
DOI related publication
https://doi.org/10.1145/3627673.3679210
More Info
expand_more
Publication Year
2024
Language
English
Research Group
Team Arjan Mol
Pages (from-to)
5239-5243
ISBN (electronic)
9798400704369
Reuse Rights

Other than for strictly personal use, it is not permitted to download, forward or distribute the text or part of it, without the consent of the author(s) and/or copyright holder(s), unless the work is under an open content license such as Creative Commons.

Abstract

The advent of Large Language Models (LLMs) provides an opportunity to change the way queries are processed, moving beyond the constraints of conventional SQL-based database systems. However, using an LLM to answer a prediction query is still challenging, since an external ML model has to be employed and inference has to be performed in order to provide an answer. This paper introduces LLM-PQA, a novel tool that addresses prediction queries formulated in natural language. LLM-PQA is the first to combine the capabilities of LLMs and retrieval-augmented mechanism for the needs of prediction queries by integrating data lakes and model zoos. This integration provides users with access to a vast spectrum of heterogeneous data and diverse ML models, facilitating dynamic prediction query answering. In addition, LLM-PQA can dynamically train models on demand, based on specific query requirements, ensuring reliable and relevant results even when no pre-trained model in a model zoo, available for the task.