AI stack built on Node.js and powered by RAG Retrieval-Augmented Generation
The below table is describing the different layer abstractions
| Layer | Node.js Tool/Lib | Notes |
|---|---|---|
| Document loader | fs |
- |
| Document Parser | langchain |
File format supported txt |
| RAG Framework | langchain |
Chain together: Retrieval + Generation |
| RAG Embeddings | Xenova/all-MiniLM-L6-v2 |
Fast 384-dim sentence embeddings |
| LLM | EleutherAI/gpt-neo-125M |
Lightweight Model |
Initally based on chatGPT discussion, 27-06-2025
The current instance is running a generative chat supplying basic response on a specific theme
- Python
3.9, later version are not yet covered by torch prebuilt - Nodejs version >=
18.0(not tested) - English documentation is only supported by LLM
Perform global installation
Use the installation script which is performing the following sub scripts,
_js-env_installscript dedicated to node environment_py-env_installscript dedicated to python environment
Note,
py-env_install script is creating a .venv virtual environment and imports the following packages (fastapi, uvicorn, torch, transformers, accelerate, pydantic)
- Dispose your documentation into the
/docfolder. Note: Currently only 'txt' files are supported - Run both
nodejsandmodelscript instances withquick-startornpm run start-dev:all - Send a POST HTTP JSON message to
http://localhost:3000/generatei.e : {"query" : "What is (your scope) ?"} - Expecting for a HTTP '201' JSON message response (json anwser response : Context... > Question... > Answer...)