Rank
70
AI Agents & MCPs & AI Workflow Automation • (~400 MCP servers for AI agents) • AI Automation / AI Agent with MCPs • AI Workflows & AI Agents • MCPs for AI Agents
Traction
No public download signal
Freshness
Updated 2d ago
Crawler Summary
LlamaFactory training framework for fine-tuning LLMs. Use when training language models with SFT, DPO, KTO, RM, preparing datasets, writing training configs, or debugging training issues like OOM errors. Covers text, multimodal (image/video/audio), and preference learning tasks. --- name: llamafactory description: LlamaFactory training framework for fine-tuning LLMs. Use when training language models with SFT, DPO, KTO, RM, preparing datasets, writing training configs, or debugging training issues like OOM errors. Covers text, multimodal (image/video/audio), and preference learning tasks. --- LlamaFactory Training Guide Quick Start Installation (AutoDL/Linux) Download Models (ModelScope) Run Capability contract not published. No trust telemetry is available yet. Last updated 4/14/2026.
Freshness
Last checked 4/14/2026
Best For
llamafactory is best for general automation workflows where OpenClaw compatibility matters.
Not Ideal For
Contract metadata is missing or unavailable for deterministic execution.
Evidence Sources Checked
editorial-content, GITHUB OPENCLEW, runtime-metrics, public facts pack
LlamaFactory training framework for fine-tuning LLMs. Use when training language models with SFT, DPO, KTO, RM, preparing datasets, writing training configs, or debugging training issues like OOM errors. Covers text, multimodal (image/video/audio), and preference learning tasks. --- name: llamafactory description: LlamaFactory training framework for fine-tuning LLMs. Use when training language models with SFT, DPO, KTO, RM, preparing datasets, writing training configs, or debugging training issues like OOM errors. Covers text, multimodal (image/video/audio), and preference learning tasks. --- LlamaFactory Training Guide Quick Start Installation (AutoDL/Linux) Download Models (ModelScope) Run
Public facts
4
Change events
1
Artifacts
0
Freshness
Apr 14, 2026
Capability contract not published. No trust telemetry is available yet. Last updated 4/14/2026.
Trust score
Unknown
Compatibility
OpenClaw
Freshness
Apr 14, 2026
Vendor
Sylvanliu
Artifacts
0
Benchmarks
0
Last release
Unpublished
Key links, install path, and a quick operational read before the deeper crawl record.
Summary
Capability contract not published. No trust telemetry is available yet. Last updated 4/14/2026.
Setup snapshot
git clone https://github.com/SylvanLiu/llamafactory-skill.gitSetup complexity is LOW. This package is likely designed for quick installation with minimal external side-effects.
Final validation: Expose the agent to a mock request payload inside a sandbox and trace the network egress before allowing access to real customer data.
Everything public we have scraped or crawled about this agent, grouped by evidence type with provenance.
Vendor
Sylvanliu
Protocol compatibility
OpenClaw
Handshake status
UNKNOWN
Crawlable docs
6 indexed pages on the official domain
Merged public release, docs, artifact, benchmark, pricing, and trust refresh events.
Extracted files, examples, snippets, parameters, dependencies, permissions, and artifact metadata.
Extracted files
0
Examples
6
Snippets
0
Languages
typescript
Parameters
bash
curl --proto '=https' --tlsv1.2 -sSf https://sh.rustup.rs | sh
bash
# Create environment conda create -n llamaf python=3.9 conda activate llamaf # Clone and install git clone --depth 1 https://github.com/hiyouga/LLaMA-Factory.git cd LLaMA-Factory pip install -e ".[torch,metrics]" # If rust error occurs: curl --proto '=https' --tlsv1.2 -sSf https://sh.rustup.rs | sh source $HOME/.cargo/env pip install -e ".[torch,metrics]"
bash
pip install modelscope modelscope download --model Qwen/Qwen2.5-VL-7B-Instruct --local_dir /autodl-fs/data/qwen25_VL_7B modelscope download --model Qwen/Qwen3-VL-8B-Instruct --local_dir /autodl-fs/data/qwen3_VL_8B
bash
llamafactory-cli train config.yaml # Or with overrides: CUDA_VISIBLE_DEVICES=0 llamafactory-cli train config.yaml learning_rate=1e-5
json
{
"my_sft_dataset": {
"file_name": "sft_data.json"
},
"my_dpo_dataset": {
"file_name": "dpo_data.json",
"ranking": true,
"formatting": "sharegpt",
"columns": {
"messages": "conversations",
"chosen": "chosen",
"rejected": "rejected"
}
}
}json
[
{
"instruction": "User instruction (required)",
"input": "Additional input (optional)",
"output": "Model response (required)",
"system": "System prompt (optional)",
"history": [
["Round 1 instruction", "Round 1 response"],
["Round 2 instruction", "Round 2 response"]
]
}
]Full documentation captured from public sources, including the complete README when available.
Docs source
GITHUB OPENCLEW
Editorial quality
ready
LlamaFactory training framework for fine-tuning LLMs. Use when training language models with SFT, DPO, KTO, RM, preparing datasets, writing training configs, or debugging training issues like OOM errors. Covers text, multimodal (image/video/audio), and preference learning tasks. --- name: llamafactory description: LlamaFactory training framework for fine-tuning LLMs. Use when training language models with SFT, DPO, KTO, RM, preparing datasets, writing training configs, or debugging training issues like OOM errors. Covers text, multimodal (image/video/audio), and preference learning tasks. --- LlamaFactory Training Guide Quick Start Installation (AutoDL/Linux) Download Models (ModelScope) Run
# Create environment
conda create -n llamaf python=3.9
conda activate llamaf
# Clone and install
git clone --depth 1 https://github.com/hiyouga/LLaMA-Factory.git
cd LLaMA-Factory
pip install -e ".[torch,metrics]"
# If rust error occurs:
curl --proto '=https' --tlsv1.2 -sSf https://sh.rustup.rs | sh
source $HOME/.cargo/env
pip install -e ".[torch,metrics]"
pip install modelscope
modelscope download --model Qwen/Qwen2.5-VL-7B-Instruct --local_dir /autodl-fs/data/qwen25_VL_7B
modelscope download --model Qwen/Qwen3-VL-8B-Instruct --local_dir /autodl-fs/data/qwen3_VL_8B
llamafactory-cli train config.yaml
# Or with overrides:
CUDA_VISIBLE_DEVICES=0 llamafactory-cli train config.yaml learning_rate=1e-5
All datasets must be registered in dataset_info.json in your dataset_dir. Example:
{
"my_sft_dataset": {
"file_name": "sft_data.json"
},
"my_dpo_dataset": {
"file_name": "dpo_data.json",
"ranking": true,
"formatting": "sharegpt",
"columns": {
"messages": "conversations",
"chosen": "chosen",
"rejected": "rejected"
}
}
}
[
{
"instruction": "User instruction (required)",
"input": "Additional input (optional)",
"output": "Model response (required)",
"system": "System prompt (optional)",
"history": [
["Round 1 instruction", "Round 1 response"],
["Round 2 instruction", "Round 2 response"]
]
}
]
Registration:
"dataset_name": {
"file_name": "data.json"
}
[
{"text": "document content"},
{"text": "document content"}
]
Registration:
"dataset_name": {
"file_name": "data.json",
"columns": {"prompt": "text"}
}
[
{
"instruction": "User instruction",
"input": "Optional input",
"chosen": "Better response",
"rejected": "Worse response"
}
]
Registration:
"dataset_name": {
"file_name": "data.json",
"ranking": true,
"columns": {
"prompt": "instruction",
"query": "input",
"chosen": "chosen",
"rejected": "rejected"
}
}
[
{
"conversations": [
{"from": "human", "value": "User message"},
{"from": "gpt", "value": "Assistant response"},
{"from": "human", "value": "Follow-up"},
{"from": "gpt", "value": "Response"}
],
"system": "System prompt (optional)",
"tools": "Tool description (optional)"
}
]
Registration:
"dataset_name": {
"file_name": "data.json",
"formatting": "sharegpt",
"columns": {"messages": "conversations"}
}
[
{
"conversations": [
{"from": "human", "value": "User question"},
{"from": "gpt", "value": "Context response"},
{"from": "human", "value": "Final question"}
],
"chosen": {"from": "gpt", "value": "Better answer"},
"rejected": {"from": "gpt", "value": "Worse answer"}
}
]
Registration:
"dataset_name": {
"file_name": "data.json",
"formatting": "sharegpt",
"ranking": true,
"columns": {
"messages": "conversations",
"chosen": "chosen",
"rejected": "rejected"
}
}
[
{
"messages": [
{"role": "user", "content": "Question"},
{"role": "assistant", "content": "Response"}
],
"label": true
}
]
Registration:
"dataset_name": {
"file_name": "data.json",
"formatting": "sharegpt",
"columns": {
"messages": "messages",
"kto_tag": "label"
},
"tags": {
"role_tag": "role",
"content_tag": "content",
"user_tag": "user",
"assistant_tag": "assistant"
}
}
[
{
"messages": [
{"role": "user", "content": "<image>Describe this image"},
{"role": "assistant", "content": "This image shows..."}
],
"images": ["path/to/image1.jpg"]
}
]
Registration:
"dataset_name": {
"file_name": "data.json",
"formatting": "sharegpt",
"columns": {
"messages": "messages",
"images": "images"
},
"tags": {
"role_tag": "role",
"content_tag": "content",
"user_tag": "user",
"assistant_tag": "assistant"
}
}
Note: Number of <image> tokens must match number of image paths.
Use <video> token and videos column. Registration similar to images.
Use <audio> token and audios column.
[
{
"conversations": [
{"from": "human", "value": "<image>What's in this image?"}
],
"chosen": {"from": "gpt", "value": "Detailed accurate description"},
"rejected": {"from": "gpt", "value": "Vague or wrong description"},
"images": ["path/to/image.jpg"]
}
]
### model
model_name_or_path: Qwen/Qwen3-4B-Instruct
trust_remote_code: true
### method
stage: sft
do_train: true
finetuning_type: lora
lora_rank: 8
lora_alpha: 16
lora_target: all
### dataset
dataset_dir: /path/to/data
dataset: my_dataset
template: qwen3_nothink
cutoff_len: 2048
preprocessing_num_workers: 16
dataloader_num_workers: 4
### output
output_dir: saves/qwen3-sft
logging_steps: 10
save_steps: 500
plot_loss: true
overwrite_output_dir: true
report_to: none
### train
per_device_train_batch_size: 1
gradient_accumulation_steps: 8
learning_rate: 1.0e-4
num_train_epochs: 3.0
lr_scheduler_type: cosine
warmup_ratio: 0.1
bf16: true
ddp_timeout: 180000000
### memory optimization
gradient_checkpointing: true
optim: adamw_torch
### model
model_name_or_path: Qwen/Qwen3-VL-8B-Instruct
image_max_pixels: 262144
video_max_pixels: 16384
trust_remote_code: true
### method
stage: sft
do_train: true
finetuning_type: lora
lora_rank: 8
lora_target: all
### dataset
dataset_dir: /path/to/data
dataset: my_vlm_dataset
template: qwen3_vl_nothink
cutoff_len: 2048
### (other params same as text SFT)
### model
model_name_or_path: Qwen/Qwen3-VL-8B-Instruct
image_max_pixels: 131072
video_max_pixels: 8192
trust_remote_code: true
### method
stage: dpo
do_train: true
finetuning_type: lora
lora_rank: 4
lora_alpha: 8
lora_target: all
pref_beta: 0.1
pref_loss: sigmoid # choices: sigmoid (dpo), orpo, simpo
### dataset
dataset_dir: /path/to/data
dataset: my_dpo_dataset
template: qwen3_vl_nothink
cutoff_len: 2048
### train
per_device_train_batch_size: 1
gradient_accumulation_steps: 16
learning_rate: 5.0e-5 # DPO uses lower LR than SFT
num_train_epochs: 2.0
### method
stage: kto
do_train: true
finetuning_type: lora
lora_rank: 8
lora_target: all
pref_beta: 0.1
### dataset
dataset: kto_dataset
template: qwen3_nothink
### train
learning_rate: 5.0e-6
### method
stage: rm
do_train: true
finetuning_type: lora
### dataset (uses preference dataset)
dataset: dpo_dataset
### method
stage: sft
do_train: true
finetuning_type: full
deepspeed: examples/deepspeed/ds_z3_config.json
### train
per_device_train_batch_size: 1
gradient_accumulation_steps: 2
learning_rate: 1.0e-5 # Lower than LoRA
| Parameter | Description |
|-----------|-------------|
| model_name_or_path | HuggingFace model ID or local path |
| trust_remote_code | Allow custom model code (required for Qwen) |
| image_max_pixels | Max pixels per image (VLM) |
| video_max_pixels | Max pixels per video frame (VLM) |
| Parameter | Description |
|-----------|-------------|
| stage | Training stage: pt, sft, rm, dpo, kto, ppo |
| finetuning_type | full, lora, freeze |
| lora_rank | LoRA rank (4-64, higher = more params) |
| lora_alpha | LoRA scaling factor (typically 2x rank) |
| lora_target | Target modules: all or specific names |
| pref_beta | DPO/KTO beta parameter (0.1-0.5) |
| pref_loss | sigmoid (DPO), orpo, simpo |
| Parameter | Description |
|-----------|-------------|
| dataset_dir | Directory containing dataset_info.json |
| dataset | Dataset name(s) from dataset_info.json |
| template | Chat template matching model |
| cutoff_len | Max sequence length |
| max_samples | Limit samples (for testing) |
| Parameter | Description |
|-----------|-------------|
| per_device_train_batch_size | Batch size per GPU |
| gradient_accumulation_steps | Steps before update |
| learning_rate | SFT: 1e-4, DPO: 5e-5~5e-6 |
| num_train_epochs | Training epochs |
| lr_scheduler_type | cosine, linear, constant |
| warmup_ratio | Warmup proportion (0.1) |
| bf16 | Use bfloat16 precision |
| gradient_checkpointing | Reduce memory usage |
| optim | Optimizer: adamw_torch (recommended) |
| Model | Template |
|-------|----------|
| Qwen3 | qwen3 / qwen3_nothink |
| Qwen3-VL | qwen3_vl / qwen3_vl_nothink |
| Qwen2.5-VL | qwen2_vl |
| Llama3 | llama3 |
| DeepSeek | deepseek / deepseekr1 |
Note: Use _nothink suffix when you don't want reasoning/thinking tokens.
Solutions in order of effectiveness:
per_device_train_batch_size: 1gradient_checkpointing: truelora_rank: 4 instead of 8cutoff_len: 1024image_max_pixels: 131072quantization_bit: 4deepspeed: examples/deepspeed/ds_z2_config.json"Dataset not found":
dataset_info.json is in dataset_dir"Column not found":
dataset_info.json"Empty dataset":
Multimodal image not found:
dataset_dirLoss is NaN:
bf16: true instead of fp16Optimizer errors:
optim: adamw_torch instead of adamw_8bitTemplate mismatch:
_nothink for non-reasoning models# Start WebUI
llamafactory-cli webui
# Chat with model
llamafactory-cli chat config.yaml
# API server
llamafactory-cli api config.yaml
# Merge LoRA
llamafactory-cli export merge_config.yaml
# Multi-GPU training
CUDA_VISIBLE_DEVICES=0,1 FORCE_TORCHRUN=1 llamafactory-cli train config.yaml
Machine endpoints, protocol fit, contract coverage, invocation examples, and guardrails for agent-to-agent use.
Contract coverage
Status
missing
Auth
None
Streaming
No
Data region
Unspecified
Protocol support
Requires: none
Forbidden: none
Guardrails
Operational confidence: low
curl -s "https://xpersona.co/api/v1/agents/sylvanliu-llamafactory-skill/snapshot"
curl -s "https://xpersona.co/api/v1/agents/sylvanliu-llamafactory-skill/contract"
curl -s "https://xpersona.co/api/v1/agents/sylvanliu-llamafactory-skill/trust"
Trust and runtime signals, benchmark suites, failure patterns, and practical risk constraints.
Trust signals
Handshake
UNKNOWN
Confidence
unknown
Attempts 30d
unknown
Fallback rate
unknown
Runtime metrics
Observed P50
unknown
Observed P95
unknown
Rate limit
unknown
Estimated cost
unknown
Do not use if
Every public screenshot, visual asset, demo link, and owner-provided destination tied to this agent.
Neighboring agents from the same protocol and source ecosystem for comparison and shortlist building.
Rank
70
AI Agents & MCPs & AI Workflow Automation • (~400 MCP servers for AI agents) • AI Automation / AI Agent with MCPs • AI Workflows & AI Agents • MCPs for AI Agents
Traction
No public download signal
Freshness
Updated 2d ago
Rank
70
AI productivity studio with smart chat, autonomous agents, and 300+ assistants. Unified access to frontier LLMs
Traction
No public download signal
Freshness
Updated 5d ago
Rank
70
Free, local, open-source 24/7 Cowork app and OpenClaw for Gemini CLI, Claude Code, Codex, OpenCode, Qwen Code, Goose CLI, Auggie, and more | 🌟 Star if you like it!
Traction
No public download signal
Freshness
Updated 6d ago
Rank
70
The Frontend for Agents & Generative UI. React + Angular
Traction
No public download signal
Freshness
Updated 23d ago
Contract JSON
{
"contractStatus": "missing",
"authModes": [],
"requires": [],
"forbidden": [],
"supportsMcp": false,
"supportsA2a": false,
"supportsStreaming": false,
"inputSchemaRef": null,
"outputSchemaRef": null,
"dataRegion": null,
"contractUpdatedAt": null,
"sourceUpdatedAt": null,
"freshnessSeconds": null
}Invocation Guide
{
"preferredApi": {
"snapshotUrl": "https://xpersona.co/api/v1/agents/sylvanliu-llamafactory-skill/snapshot",
"contractUrl": "https://xpersona.co/api/v1/agents/sylvanliu-llamafactory-skill/contract",
"trustUrl": "https://xpersona.co/api/v1/agents/sylvanliu-llamafactory-skill/trust"
},
"curlExamples": [
"curl -s \"https://xpersona.co/api/v1/agents/sylvanliu-llamafactory-skill/snapshot\"",
"curl -s \"https://xpersona.co/api/v1/agents/sylvanliu-llamafactory-skill/contract\"",
"curl -s \"https://xpersona.co/api/v1/agents/sylvanliu-llamafactory-skill/trust\""
],
"jsonRequestTemplate": {
"query": "summarize this repo",
"constraints": {
"maxLatencyMs": 2000,
"protocolPreference": [
"OPENCLEW"
]
}
},
"jsonResponseTemplate": {
"ok": true,
"result": {
"summary": "...",
"confidence": 0.9
},
"meta": {
"source": "GITHUB_OPENCLEW",
"generatedAt": "2026-04-17T00:58:32.281Z"
}
},
"retryPolicy": {
"maxAttempts": 3,
"backoffMs": [
500,
1500,
3500
],
"retryableConditions": [
"HTTP_429",
"HTTP_503",
"NETWORK_TIMEOUT"
]
}
}Trust JSON
{
"status": "unavailable",
"handshakeStatus": "UNKNOWN",
"verificationFreshnessHours": null,
"reputationScore": null,
"p95LatencyMs": null,
"successRate30d": null,
"fallbackRate": null,
"attempts30d": null,
"trustUpdatedAt": null,
"trustConfidence": "unknown",
"sourceUpdatedAt": null,
"freshnessSeconds": null
}Capability Matrix
{
"rows": [
{
"key": "OPENCLEW",
"type": "protocol",
"support": "unknown",
"confidenceSource": "profile",
"notes": "Listed on profile"
}
],
"flattenedTokens": "protocol:OPENCLEW|unknown|profile"
}Facts JSON
[
{
"factKey": "docs_crawl",
"category": "integration",
"label": "Crawlable docs",
"value": "6 indexed pages on the official domain",
"href": "https://github.com/login?return_to=https%3A%2F%2Fgithub.com%2Fopenclaw%2Fskills%2Ftree%2Fmain%2Fskills%2Fasleep123%2Fcaldav-calendar",
"sourceUrl": "https://github.com/login?return_to=https%3A%2F%2Fgithub.com%2Fopenclaw%2Fskills%2Ftree%2Fmain%2Fskills%2Fasleep123%2Fcaldav-calendar",
"sourceType": "search_document",
"confidence": "medium",
"observedAt": "2026-04-15T05:03:46.393Z",
"isPublic": true
},
{
"factKey": "vendor",
"category": "vendor",
"label": "Vendor",
"value": "Sylvanliu",
"href": "https://github.com/SylvanLiu/llamafactory-skill",
"sourceUrl": "https://github.com/SylvanLiu/llamafactory-skill",
"sourceType": "profile",
"confidence": "medium",
"observedAt": "2026-04-14T22:27:07.735Z",
"isPublic": true
},
{
"factKey": "protocols",
"category": "compatibility",
"label": "Protocol compatibility",
"value": "OpenClaw",
"href": "https://xpersona.co/api/v1/agents/sylvanliu-llamafactory-skill/contract",
"sourceUrl": "https://xpersona.co/api/v1/agents/sylvanliu-llamafactory-skill/contract",
"sourceType": "contract",
"confidence": "medium",
"observedAt": "2026-04-14T22:27:07.735Z",
"isPublic": true
},
{
"factKey": "handshake_status",
"category": "security",
"label": "Handshake status",
"value": "UNKNOWN",
"href": "https://xpersona.co/api/v1/agents/sylvanliu-llamafactory-skill/trust",
"sourceUrl": "https://xpersona.co/api/v1/agents/sylvanliu-llamafactory-skill/trust",
"sourceType": "trust",
"confidence": "medium",
"observedAt": null,
"isPublic": true
}
]Change Events JSON
[
{
"eventType": "docs_update",
"title": "Docs refreshed: Sign in to GitHub · GitHub",
"description": "Fresh crawlable documentation was indexed for the official domain.",
"href": "https://github.com/login?return_to=https%3A%2F%2Fgithub.com%2Fopenclaw%2Fskills%2Ftree%2Fmain%2Fskills%2Fasleep123%2Fcaldav-calendar",
"sourceUrl": "https://github.com/login?return_to=https%3A%2F%2Fgithub.com%2Fopenclaw%2Fskills%2Ftree%2Fmain%2Fskills%2Fasleep123%2Fcaldav-calendar",
"sourceType": "search_document",
"confidence": "medium",
"observedAt": "2026-04-15T05:03:46.393Z",
"isPublic": true
}
]Sponsored
Ads related to llamafactory and adjacent AI workflows.