Skip to content

LLM-powered RAG Question Answering Slack bot for DataTalksClub Zoomcamps

Notifications You must be signed in to change notification settings

aaalexlit/faq-slack-bot

Repository files navigation

This is a repo for a human-interface LLM-powered QA Slack chatbot for answering questions related to DataTalks.Club Zoomcamps

Current version

Please follow this report that explains in detail how the bot works

First version

Behind the scenes

Course FAQ Google Document and the Course repo get indexed to the Pinecone vector store. Then semantic search retrieves the most similar (and hopefully most relevant) pieces to the question asked. Then this information is passed as a context to a conversational LLM to form the final answer.

Diagram

Before you start

Use any python virtual environment manager of your preference and install the dependencies from requirements.txt

Fill Pinecone index

  1. Set PINECONE_API_KEY and PINECONE_ENV environmental variables accordingly

  2. run ingest/ingest.py

    python ingest_mlops_old.py

Test QA pipeline locally

  1. Set OPENAI_API_KEY, PINECONE_API_KEY, and PINECONE_ENV environmental variables accordingly

  2. Run test.py

    python test.py

Launch the bot

  1. Set OPENAI_API_KEY, SLACK_APP_TOKEN, SLACK_BOT_TOKEN, PINECONE_API_KEY, and PINECONE_ENV environmental variables accordingly
  2. To launch the bot on the CLI run slack_bot/main.py
    python main.py

Alternatively it can be launched with Docker, please follow this README

About

LLM-powered RAG Question Answering Slack bot for DataTalksClub Zoomcamps

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published