Skip to content

Latest commit

 

History

History
 
 

dspy-rag-fastapi

Folders and files

NameName
Last commit message
Last commit date

parent directory

..
 
 
 
 
 
 
 
 
 
 
 
 

Full-Stack DSPy Application with FastAPI and Streamlit

Introduction

This project is a full-stack application designed to leverage natural language processing capabilities entirely locally and to integrate with the DSPy framework developed by StanfordNLP. It features a FastAPI backend for processing and a Streamlit frontend for interactive user interfaces. This implementation utilizes OpenAI or Cohere for language and embedding models, Weaviate for vector storage, and Arize Phoenix for observability.

Features

  • OpenAI or Cohere: Leveraged for language and embedding models.
  • Weaviate DB Vector Storage: Utilizes Weaviate DB for efficient, scalable vector storage, enabling quick and precise information retrieval.
  • Arize Phoenix Observability: Integrates Arize Phoenix for real-time monitoring and analytics, aiding in performance improvement and system health tracking.
  • FastAPI Backend: Offers robust and scalable API endpoints for interacting with the NLP models and performing various queries and compilations.
  • Streamlit Frontend: Provides an intuitive and interactive UI for users to easily interact with the backend services, improving the overall user experience.

Architecture

This full-stack application combines the DSPy Framework with OpenAI, Cohere, Arize Phoenix, and Weaviate DB in a cohesive ecosystem. Here's a brief overview of the system components:

  • DSPy Framework: Serves as the core for language model interactions, offering advanced NLP capabilities.
  • OpenAI/Cohere: Acts as the backend engine for language understanding and generation.
  • Weaviate: Provides efficient vector storage solutions, essential for NLP tasks like semantic search.
  • Arize Phoenix: Enhances visibility into the application's performance and health.
  • FastAPI: Facilitates the backend logic, handling API requests and responses.
  • Streamlit: Creates the frontend interface, enabling users to engage with the backend services visually.

Installation

Prerequisites

  • Docker and Docker-Compose

Getting Started with Local Development

Backend setup

First, navigate to the backend directory:

cd backend/

Second, setup the environment:

poetry config virtualenvs.in-project true
poetry install
poetry shell

Specify your environment variables in an .env file in backend directory. Example .env file:

ENVIRONMENT=<your_environment_value>
INSTRUMENT_DSPY=<true or false>
COLLECTOR_ENDPOINT=<your_arize_phoenix_endpoint>
OPENAI_API_KEY=<your_openai_api_key>
CO_API_KEY=<your_cohere_api_key>

Third, run this command to create embeddings of data located in data/example folder:

python app/utils/load.py

Then run this command to start the FastAPI server:

python main.py

Frontend setup

First, navigate to the frontend directory:

cd frontend/

Second, setup the environment:

poetry config virtualenvs.in-project true
poetry install
poetry shell

Specify your environment variables in an .env file in backend directory. Example .env file:

FASTAPI_BACKEND_URL = <your_fastapi_address>

Then run this command to start the Streamlit application:

streamlit run about.py

Getting Started with Docker-Compose

This project now supports Docker Compose for easier setup and deployment, including backend services and Arize Phoenix for query tracing.

  1. Configure your environment variables in the .env file or modify the compose file directly.
  2. Ensure that Docker is installed and running.
  3. This project uses OpenAI to embed data, so you will need to create the embeddings first. Run the command python -m app.utils.load from the backend folder to create embeddings for the data located in the data/example folder.
  4. Run the command docker-compose -f compose.yml up to spin up services for the backend, and Phoenix.
  5. Backend docs can be viewed using the OpenAPI Spec.
  6. Frontend can be viewed using Streamlit
  7. Traces can be viewed using the Phoenix UI.
  8. When you're finished, run docker compose down to spin down the services.

Usage

The FastAPI and Streamlit integration allows for seamless interaction between the user and the NLP backend. Utilize the FastAPI endpoints for NLP tasks and visualize results and interact with the system through the Streamlit frontend.

Acknowledgements

This example is a fork of dspy-rag-fastapi by @diicellman and credit for the implementation goes to them.