EPISODE · Aug 22, 2023 · 1H 2M
FrugalGPT: Better Quality and Lower Cost for LLM Applications // Lingjiao Chen // #172
from MLOps.community · host Demetrios
MLOps Coffee Sessions #172 with Lingjiao Chen, FrugalGPT: Better Quality and Lower Cost for LLM Applications. This episode is sponsored by QuantumBlack. We are now accepting talk proposals for our next LLM in Production virtual conference on October 3rd. Apply to speak here: https://go.mlops.community/NSAX1O// AbstractThere is a rapidly growing number of large language models (LLMs) that users can query for a fee. We review the cost associated with querying popular LLM APIs, e.g. GPT-4, ChatGPT, J1-Jumbo, and find that these models have heterogeneous pricing structures, with fees that can differ by two orders of magnitude. In particular, using LLMs on large collections of queries and text can be expensive. Motivated by this, we outline and discuss three types of strategies that users can exploit to reduce the inference cost associated with using LLMs: 1) prompt adaptation, 2) LLM approximation, and 3) LLM cascade. As an example, we propose FrugalGPT, a simple yet flexible instantiation of LLM cascade that learns which combinations of LLMs to use for different queries in order to reduce cost and improve accuracy. Our experiments show that FrugalGPT can match the performance of the best individual LLM (e.g., GPT-4) with up to 98% cost reduction or improve the accuracy over GPT-4 by 4% with the same cost. The ideas and findings presented here lay a foundation for using LLMs sustainably and efficiently.// BioLingjiao Chen is a Ph.D. candidate in the Computer Science department at Stanford University. He is broadly interested in machine learning, data management, and optimization. Working with Matei Zaharia and James Zou, he is currently exploring the fast-growing marketplaces of artificial intelligence and data. His work has been published at premier conferences and journals such as ICML, NeurIPS, SIGMOD, and PVLDB, and partially supported by a Google fellowship.// MLOps Jobs board jobs.mlops.community// MLOps Swag/Merchhttps://mlops-community.myshopify.com/// Related LinksWebsite: https://lchen001.github.io/FrugalGPT: How to Use Large Language Models While Reducing Cost and Improving Performance paper: https://arxiv.org/abs/2305.05176--------------- ✌️Connect With Us ✌️ -------------Join our Slack community: https://go.mlops.community/slackFollow us on Twitter: @mlopscommunitySign up for the next meetup: https://go.mlops.community/registerCatch all episodes, blogs, newsletters, and more: https://mlops.community/Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/Connect with Lingjiao on LinkedIn: Timestamps:[00:00] Lingjiao's preferred coffee[00:35] Takeaways[02:41] Sponsor Ad: Nayur Khan of QuantumBlack[05:27] Lingjiao's research at Stanford[07:51] Day-to-day research overview[10:11] Inventing data management inspired abstractions research[13:58] Agnostic Approach to Data Management[15:56] Frugal GPT[18:59] Just another data provider[19:51] Frugal GPT breakdown[26:33] First step of optimizing the prompts[28:04] Prompt overlap[29:06] Query Concatenation[32:30] Money saving[35:04] Economizing the prompts[38:52] Questions to accommodate[41:33] LLM Cascade[47:25] Frugal GPT saves cost and improves performance[51:37] End-user implementation[52:31] Completion Cache[56:33] Using a vector store[1:00:51] Wrap up
NOW PLAYING
FrugalGPT: Better Quality and Lower Cost for LLM Applications // Lingjiao Chen // #172
No transcript for this episode yet
Similar Episodes
Apr 21, 2026 ·13m
Apr 19, 2026 ·16m
Apr 17, 2026 ·13m
Apr 13, 2026 ·11m
Apr 11, 2026 ·16m