Work with Obsidian vaults (plain Markdown notes) and automate via obsidian-cli.
npx skills add benchflow-ai/skillsbench --skill "skillsbench"
Install specific skill from multi-skill repository
# Description
SkillsBench contribution workflow. Use when: (1) Creating benchmark tasks, (2) Understanding repo structure, (3) Preparing PRs for task submission.
# SKILL.md
name: skillsbench
description: "SkillsBench contribution workflow. Use when: (1) Creating benchmark tasks, (2) Understanding repo structure, (3) Preparing PRs for task submission."
SkillsBench
Benchmark evaluating how well AI agents use skills.
Official Resources
- GitHub: https://github.com/benchflow-ai/skillsbench
- Harbor Docs: https://harborframework.com/docs
- Terminal-Bench Task Guide: https://www.tbench.ai/docs/task-quickstart
Repo Structure
tasks/<task-id>/ # Benchmark tasks
.claude/skills/ # Contributor skills (harbor, skill-creator)
CONTRIBUTING.md # Full contribution guide
Local Workspace & API Keys
.local-workspace/- Git-ignored directory for cloning PRs, temporary files, external repos, etc..local-workspace/.env- May containANTHROPIC_API_KEYand other API credentials. Check and use when running harbor with API access.
Quick Workflow
# 1. Install harbor and initialize a task
uv tool install harbor
harbor tasks init tasks/<task-name>
# 2. Write files (see harbor skill for templates)
# 3. Validate
harbor tasks check tasks/my-task
harbor run -p tasks/my-task -a oracle # Must pass 100%
# 4. Test with agent
# you can test agents with your claude code/cursor/codex/vscode/GitHub copilot subscriptions, etc.
# you can also trigger harbor run with your own API keys.
# for PR we only need some idea of how sota level models and agents perform on the task with and without skills.
harbor run -p tasks/my-task -a claude-code -m 'anthropic/claude-opus-4-5'
# 5. Submit PR
# please see our PR template for more details.
Task Requirements
- Things people actually do
- Agent should be more performant with skills than without skills.
- task.toml and solve.sh must be human-authored
- Prioritize verifiable tasks. LLM-as-judge or agent-as-a-judge is not priority right now. But we are open to these ideas and will brainstorm if we can make these tasks easier to verify.
References
- Full guide: CONTRIBUTING.md
- Harbor commands: See harbor skill
- Task ideas: references/task-ideas.md
# Supported AI Coding Agents
This skill is compatible with the SKILL.md standard and works with all major AI coding agents:
Learn more about the SKILL.md standard and how to use these skills with your preferred AI coding agent.