Skip to main content
Version: 2.3.1 (Latest)

FAQ

Can I use this with my own vector database?

Yes! You can implement the Retriever interface and register it:

registry.register("retriever", "mydb", new MyRetriever());

Then use it via CLI or .ragrc.json.


Is streaming LLM response supported?

Yes. LLM runners can yield async output tokens. This is handled internally via AsyncIterable in compatible models.


How do I evaluate performance?

Use rag-pipeline evaluate dataset.json or open the dashboard at http://localhost:3000 to visualize BLEU/ROUGE and pass rates.


What file types can I ingest?

Currently supported:

  • .pdf (mocked)
  • .md
  • .html
  • .csv
  • Full directory ingestion

Can I rerank results using a local model?

Yes, as long as your local LLM implements:

llm.generate(prompt, context): Promise<string>

Use it to power LLMReranker.


How do I contribute?

Fork the repo → add your plugin or CLI feature → test → submit PR.


Still stuck? Open an issue at GitHub.


Next → Back to Introduction