Skip to main content

2 posts tagged with "FastAPI"

View All Tags

· 19 min read

Off-the-shelf LLMs are excellent at manipulating and generating text, but they only know general facts about the world and probably very little about your use case. Retrieval augmented generation (RAG) refers not to a single algorithm, but rather a broad approach to provide relevant context to an LLM. As industry applications mature, RAG strategies will be tailored case-by-case to optimize relevance, business outcomes, and operational concerns.

Alt text

crosspost from https://blog.dagworks.io/p/retrieval-augmented-generation-reference-arch

· 15 min read

Skip learning convoluted LLM-specific frameworks and write your first LLM application using regular Python functions and Hamilton! In this post, we’ll present a containerized PDF summarizer powered by the OpenAI API. Its flow is encoded in Hamilton, which the FastAPI backend runs and exposes as an inference endpoint. The lightweight frontend uses Streamlit and exercises the backend. (GitHub repo)

Alt text

crosspost from https://blog.dagworks.io/p/containerized-pdf-summarizer-with