没有市场词。这页讲清楚 WeChaser / 时光脉络的代码、数据、调用链是怎么搭起来的——技术决策人可以直接判断这套东西能不能进你机房。 No marketing words. This page lays out how WeChaser / Shiguangmailuo is actually built — code, data, call paths — so technical decision-makers can judge directly whether it fits your stack.
FileParser · PDF / Word / Excel / PPT / Image
DocumentParser · MinerU · PaddleOCR
DeepSeek-Reasoner / GLM-4.5 · prompts/ 全部专门设计
Validator · file_hash 唯一约束 · schema check
Neo4jLoader · Qdrant VectorStore · 1024-d embedding
quick / standard / deep · 不同档位走不同 pipeline
vector + graph + sparse · 并行
rerank fusion + redundancy removal
BAAI/bge-reranker-base · 最终 top-k
DeepSeek-Chat · grounded RAG · 100% citation
port 1000 · Uvicornport 7860 + Ingest UI :7861提升整体准确度有两条路:让用户表达更精确(不可控),或让 AI 更懂这家企业(我们的事)。后者是知识库的核心价值。
To raise accuracy you can either tighten user expression (out of our hands) or deepen AI's understanding of this specific company (our job). The latter is the entire value of a knowledge base.
它让其他 AI 工具变得能解决问题。边界要清晰——我们不做生成,不抢调用方的活。这条原则保护我们不被功能蔓延拖垮。
It makes other AI tools capable of solving them. The boundary stays sharp — we don't generate, we don't compete with the calling agent. This is what protects us from feature creep.
知识库只做后者,前者交给调用方 Agent。这避免了"知识库幻觉"——大多数 RAG 灾难都来自把生成揉进检索环节。
The KB does only retrieval; generation belongs to the calling agent. This avoids the "knowledge-base hallucination" failure mode — most RAG disasters come from blending generation into retrieval.
未来要介入工业生产决策。决策辅助系统级别的责任 ≠ 聊天玩具级别的容忍度。100% 溯源是底线,不知道就拒答。
We are headed toward supporting industrial production decisions. Decision-support liability ≠ chat-toy tolerance. 100% citation is the floor; if we don't know, we refuse.
不被花哨 UI 或未来场景拖住。第一阶段:能简单安装、操作、有 app 图标、有 WebUI。等真用起来,反馈来推迭代方向。
Don't be paralysed by polished UI or hypothetical scenarios. Phase one: install, operate, app icon, WebUI. Let real usage drive what to build next.
如果你是技术负责人,我们可以走更深的会议——演示 MCP 接入流程、跑一遍 docker-compose 到客户机房的部署。 If you're a technical lead, we can do a deeper session — live demo of MCP integration and a docker-compose deployment dry-run on your hardware.