
Services
Technologies
OpenWebUI Configuration for Proprietary Document Databases
Configuration and optimization of OpenWebUI to enable intelligent conversations with proprietary document databases. Comprehensive technical audit and advanced setup of embedding models, vector search, and reranking.

Project Context
MPH1865, industrial leader, needed to configure OpenWebUI to intelligently interact with a proprietary document database. The goal was to enable conversational AI that could understand and retrieve relevant information from complex technical documents.
However, the initial configuration prevented the system from working effectively. For example, when users asked questions about specific technical concepts, the system failed to retrieve the most relevant document sections. The search results were often irrelevant or incomplete, making it difficult for users to find the information they needed.
The main challenges included:
Queries bypassed the vector database entirely, relying only on keyword matching
Documents were not properly segmented, leading to poor context extraction
The embedding model was too limited to capture semantic relationships between concepts
Search results were slow and lacked the precision needed for technical documentation
Solution
We conducted a comprehensive technical audit of the OpenWebUI configuration and implemented a series of optimizations to transform the system into an effective conversational AI tool for document databases.
Our approach focused on three key areas:
1. Enabling Semantic Search
We disabled the "Bypass Embedding & Retrieval" mode, allowing queries to leverage the full power of vector search. This enabled the system to understand the semantic meaning of questions, not just match keywords. For example, a query about "data processing methods" now correctly retrieves documents discussing "data transformation techniques" or "information processing approaches".
2. Optimizing Document Processing
We refined the document chunking strategy to 400 / 50, ensuring that each document segment contains enough context while remaining focused. This means that when users ask about a specific concept, the system can retrieve the exact paragraph or section that contains the relevant information, rather than returning entire documents.
We upgraded the embedding model from a 384-dimensional model to OpenAI's text-embedding-3-small (1536 dimensions). This significantly improved the system's ability to understand nuanced relationships between concepts. For instance, it can now distinguish between "API authentication" and "user authentication" even when both terms appear in similar contexts.
3. Enhancing Search Performance
We implemented hybrid search combining BM25 (keyword-based) and vector search (semantic), with BM25 weight set to 0.5. This dual approach ensures that both exact matches and semantically similar content are considered. For example, when searching for "error handling", the system finds documents with exact matches, but also identifies related content about "exception management" or "fault tolerance".
We replaced the heavy reranker (bge-reranker-v2-m3) with OpenAI's optimized reranker. This change dramatically improved response times while maintaining result quality. Users now receive answers in seconds rather than waiting for lengthy processing.
Results
The optimizations transformed OpenWebUI into a powerful tool for interacting with proprietary document databases. The system now delivers precise, relevant results that help users quickly find the information they need.
Key improvements:
Semantic understanding: Users can now ask questions in natural language, and the system understands the intent behind their queries, not just keyword matches. For example, asking "How do I handle errors?" retrieves relevant sections about exception handling, error management, and troubleshooting procedures.
Improved relevance: Hybrid search (BM25 + vectors) ensures that both exact matches and semantically related content appear in results. This means users find comprehensive answers even when exact terminology differs between their question and the documentation.
Faster responses: The optimized reranker reduced response times significantly. What previously took 30-60 seconds now completes in under 5 seconds, improving the user experience dramatically.
Better context extraction: Optimized chunking (400/50) allows the system to retrieve precisely the right document sections. Instead of returning entire documents, users receive focused excerpts that directly answer their questions.
Higher quality semantic matching: The upgraded embedding model (1536 dimensions) better captures relationships between technical concepts, enabling more accurate retrieval of related information.