This is an AI-driven platform that retrieves, analyzes, and surfaces accurate information from YouTube videos related to Disaster & Emergency Management (DEM) and Health, Safety & Environment (HSE). It semantically evaluates video content using Large Language Models (LLMs) to answer user queries using Retrieval Augmented Generation (RAG) with context-rich responses while displaying the relevant videos within the application.
In fields like DEM and HSE, timely and accurate access to information can be critical. However, valuable insights are often locked inside unstructured video content. This project was created to address the gap by:
- Automatically identifying and retrieving relevant videos from YouTube.
- Extracting and semantically validating their transcripts for alignment with user queries.
- Chunking the transcripts intelligently and storing them in a vector database for fast, accurate retrieval.
- Providing an integrated conversational AI assistant to surface insights interactively and visually.
- Navigate to the frontend directory and install dependencies using the provided package-lock.toml.
cd frontend npm install - Build and run the development server.
npm run build npm start
- Ensure Docker and Docker Compose are installed.
- Run the backend services using Docker Compose.
This will start:
docker-compose up --build
- The backend service (Flask)
- MongoDB (for storing metadata and sentiment/topic analysis)
- Qdrant (for vector storage of chunked transcripts)
- Ollama (for embeddings and LLM responses)
- Before running the backend, ensure:
- The Ollama container has the Nomic embedding model downloaded.
- A Groq API Key is available for chat model inference.
