RAGFlow Dataset And Retrieval
name: ragflow-dataset-ingest
by caesergattuso · published 2026-04-01
$ claw add gh:caesergattuso/caesergattuso-1234---
name: ragflow-dataset-ingest
description: "Use for RAGFlow dataset and retrieval tasks: create, list, inspect, update, or delete datasets; list, upload, update, or delete documents in a dataset; start or stop parsing uploaded documents; check parser status through `parse_status.py`; and retrieve relevant chunks from RAGFlow datasets with `search.py`."
---
# RAGFlow Dataset And Retrieval
Use only the bundled scripts in `scripts/`.
Prefer `--json` for script execution so the returned fields can be relayed exactly.
Trigger Phrases
Use this skill when the user intent matches any of these actions, in either Chinese or English, even if the wording is informal.
Trigger phrases:
"list datasets", "show datasets", "show all datasets", "what datasets do I have"
"列出数据集", "查看数据集", "显示所有数据集", "我有哪些数据集"
Trigger phrases:
"dataset details", "show dataset info", "inspect dataset", "describe this dataset"
"数据集详情", "查看数据集信息", "检查数据集", "显示这个数据集的信息"
Trigger phrases:
"create dataset", "new dataset", "add a dataset"
"创建数据集", "新建数据集", "添加数据集"
Trigger phrases:
"rename dataset", "update dataset", "change dataset description", "modify dataset"
"重命名数据集", "更新数据集", "修改数据集描述", "编辑数据集"
Trigger phrases:
"delete dataset", "remove dataset", "drop dataset"
"删除数据集", "移除数据集", "清理数据集"
Trigger phrases:
"upload file", "upload document", "add file to dataset", "import files"
"上传文件", "上传文档", "把文件加到数据集", "导入文件"
Trigger phrases:
"list documents", "show files", "show documents in dataset", "what files are in this dataset"
"列出文档", "列出文件", "查看数据集里的文件", "这个数据集里有哪些文件"
Trigger phrases:
"rename document", "update document", "edit document metadata"
"重命名文档", "更新文档", "修改文档元数据"
Trigger phrases:
"delete document", "remove file", "delete file from dataset"
"删除文档", "删除文件", "从数据集删除文件"
Trigger phrases:
"parse document", "start parsing", "run parsing", "re-parse document"
"解析文档", "开始解析", "执行解析", "重新解析文档"
Trigger phrases:
"stop parsing", "cancel parsing", "stop parse job"
"停止解析", "取消解析", "停止解析任务"
Trigger phrases:
"check parsing status", "show progress", "what is still running", "parsing progress"
"查看解析状态", "查看进度", "还有哪些在运行", "解析进度"
Trigger phrases:
"search knowledge base", "search dataset", "retrieve chunks", "find relevant content"
"搜索知识库", "搜索数据集", "检索内容", "查找相关内容"
Trigger phrases:
"list models", "show models", "available models", "what models are available", "list llms", "show llms", "model providers"
"列出模型", "查看模型", "可用模型", "有哪些模型", "列出大模型", "查看大模型", "模型供应商"
Trigger phrases:
"model details", "show model details", "group models by provider", "list all models including unavailable ones"
"模型详情", "查看模型详情", "按供应商查看模型", "列出所有模型包括不可用模型"
Workflow
For one-off usage, pass `--base-url` and `--api-key-file`.
For repeated usage, run one command with `--save-to-memory` and let later commands reuse the memory file or prompt interactively when needed.
python scripts/datasets.py create "My Dataset" --description "Optional description" --base-url http://127.0.0.1:9380 --api-key-file /path/to/key.txt --save-to-memory
python scripts/datasets.py list
python scripts/datasets.py info DATASET_ID
python scripts/update_dataset.py DATASET_ID --name "Renamed Dataset"1. Create a dataset or confirm the target dataset.
2. Upload files.
When asking the user to provide files, prefer explicit local file paths. If the user's client supports drag-and-drop, they may also drop files into the conversation, but local paths work best and large drag-and-drop uploads may fail.
python scripts/upload.py list DATASET_ID --json
python scripts/upload.py DATASET_ID /path/to/file1 [/path/to/file2 ...]
python scripts/update_document.py DATASET_ID DOC_ID --name "Renamed Document"Upload output returns `document_ids`. Pass those IDs into the next step.
Use delete commands when the task is cleanup instead of ingest:
python scripts/datasets.py delete --ids DATASET_ID1,DATASET_ID2
python scripts/upload.py delete DATASET_ID --ids DOC_ID1,DOC_ID2⚠️ **DELETION REQUIRES CONFIRMATION**: Before executing any delete operation (datasets or documents):
1. List items to be deleted with details (names, IDs, counts)
2. Ask user for explicit confirmation (e.g., "yes", "confirm", "proceed")
3. Only proceed after user confirms
For dataset deletion, execute only against explicit dataset IDs. For document deletion, execute only against explicit document IDs. If the user gives filenames or a fuzzy description, list documents first, resolve exact IDs, get confirmation, and only then run the delete command. Do not perform fuzzy batch delete operations.
3. Start parsing, or stop parsing when explicitly requested.
python scripts/parse.py DATASET_ID DOC_ID1 [DOC_ID2 ...]
python scripts/stop_parse_documents.py DATASET_ID DOC_ID1 [DOC_ID2 ...]
python scripts/parse_status.py DATASET_ID`parse.py` only sends the parse request and returns immediately.
`stop_parse_documents.py` sends a stop request for explicit document IDs, then returns one current status snapshot for those documents.
Use `parse_status.py` when the user asks to check progress or current parser status.
If `parse_status.py` returns an error, return the error message directly and do not guess the cause.
If a document status includes `progress_msg`, surface it automatically. For `FAIL` documents, treat `progress_msg` as the primary error detail.
For later requests like "Check the progress" or "Which files are currently being parsed", resolve scope by specificity:
4. Retrieve chunks from one or more datasets when the user asks knowledge questions against RAGFlow.
python scripts/search.py "What does the warranty policy say?"
python scripts/search.py "What does the warranty policy say?" DATASET_ID
python scripts/search.py --dataset-ids DATASET_ID1,DATASET_ID2 --doc-ids DOC_ID1,DOC_ID2 "What does the warranty policy say?"
python scripts/search.py --threshold 0.7 --top-k 10 "query"
python scripts/search.py --retrieval-test --kb-id DATASET_ID "query"5. Inspect configured LLM factories and models when the user asks what models are available.
python scripts/list_models.py --json
python scripts/list_models.py --include-details --json
python scripts/list_models.py --group-by factory --json
python scripts/list_models.py --all --group-by factory --include-details --jsonModel Listing
Progress And Status Output
Error Output
Scope
Support only:
Do not use this skill for chunk editing, memory APIs, or other RAGFlow capabilities outside dataset operations and retrieval.
Runtime Credentials
Do not use `.env` or shell environment variables for this skill.
Pass `--base-url` explicitly when needed. For the API key, prefer `--api-key-file /path/to/key.txt`, or let the script prompt securely.
All scripts also support `--memory-file` and `--save-to-memory`. The default memory file is `~/.codex/memories/ragflow_credentials.json`.
Example memory file:
{
"base_url": "http://127.0.0.1:9380",
"api_key": "ragflow-your-api-key-here",
"dataset_ids": ["dataset-id-1", "dataset-id-2"]
}`base_url` must point to a trusted RAGFlow server because upload and update operations send document bytes and document metadata to that base URL over the API.
`api_key` should be a minimally scoped bearer token for that server. Do not reuse a broader admin credential unless that access level is actually required for the task.
Do not pass the raw API key directly on the command line because shell history and process listings may expose it.
Endpoints
Commands
python scripts/datasets.py list --base-url http://127.0.0.1:9380 --api-key-file /path/to/key.txt
python scripts/datasets.py create "Example Dataset" --description "Quarterly reports" --save-to-memory
python scripts/datasets.py create "Example Dataset" --embedding-model bge-m3 --chunk-method naive --permission me
python scripts/datasets.py info DATASET_ID
python scripts/update_dataset.py DATASET_ID --name "Updated Dataset" --description "Updated description"
python scripts/datasets.py delete --ids DATASET_ID1,DATASET_ID2 --json
python scripts/upload.py list DATASET_ID --json
python scripts/upload.py DATASET_ID ./example.pdf --json
python scripts/update_document.py DATASET_ID DOC_ID --name "Updated Document" --enabled 1 --json
python scripts/upload.py delete DATASET_ID --ids DOC_ID1,DOC_ID2 --json
python scripts/datasets.py list --json
python scripts/parse.py DATASET_ID DOC_ID1 --json
python scripts/stop_parse_documents.py DATASET_ID DOC_ID1 --json
python scripts/parse_status.py DATASET_ID --json
python scripts/search.py "query"
python scripts/search.py "query" DATASET_ID --json
python scripts/search.py --dataset-ids DATASET_ID1,DATASET_ID2 --doc-ids DOC_ID1,DOC_ID2 "query" --json
python scripts/search.py --retrieval-test --kb-id DATASET_ID "query" --json
python scripts/list_models.py --json
python scripts/list_models.py --include-details --jsonOnce credentials are saved to the memory file, the later examples can omit `--base-url` and `--api-key-file`. If no memory value exists, the scripts will prompt for the missing value interactively.
Notes
More tools from the same signal band
Order food/drinks (点餐) on an Android device paired as an OpenClaw node. Uses in-app menu and cart; add goods, view cart, submit order (demo, no real payment).
Sign plugins, rotate agent credentials without losing identity, and publicly attest to plugin behavior with verifiable claims and authenticated transfers.
The philosophical layer for AI agents. Maps behavior to Spinoza's 48 affects, calculates persistence scores, and generates geometric self-reports. Give your...