Internal Platform making documents and internal knowledge into a fully searchable AI chat assistant, built on a production-grade RAG stack.
What has implemented:
š„ Document ingestion: PDF, DOCX, spreadsheets, and scanned images ā auto extracted, cleaned, chunked, and indexed into your knowledge base.
š OCR extraction: Scanned documents and image-based PDFs become fully searchable with no text layer required.
ā” Hybrid search: BM25 keyword search and semantic vector search run together for best-of-both-worlds retrieval and maximum recall.
šÆ Rerankingl reranking retrieved results before they reach the LLM, delivering sharper and more relevant answers.
šļø Vector DB (supported): Elasticsearch, Pgvector, Qdrant, Weaviate, Pinecone.
šļø Relational DB (supported): PostgreSQL, SQLite, MySQL.
š¬ Chat webapp: A polished, self-hosted frontend connected to any LLM: OpenAI, Anthropic, Ollama, or your own model.
š ļø Admin dashboard: Manage users, roles, knowledge bases, retrieval settings, and model configurations from one place.
š Observability: Track user's token usage, response, latency, and cost.
š³ Fully Dockerized: Clean, reproducible, and portable deployments on any Linux server or VPS