diff --git a/README.md b/README.md index 28363a6..c066c08 100644 --- a/README.md +++ b/README.md @@ -728,27 +728,62 @@ The coordinator model has access to these tools: **Task tracking:** `todo_write`, `todo_read` -## Skills +## Agent Skills -Skills are Python modules loaded from `~/.team-harness/skills/` and `/skills/`. Each skill exports `name`, `description`, `parameters_schema`, and an async `execute(**args, ctx)` function. +team-harness supports the [Agent Skills](https://agentskills.io) standard — a cross-tool format for giving AI agents specialized knowledge and workflows. -Example (`skills/summarise.py`): +A skill is a directory containing a `SKILL.md` file with YAML frontmatter (name + description) and markdown instructions. The coordinator sees skill metadata at startup and can read the full instructions via its `read_file` tool when a task calls for it. -```python -name = "summarise_file" -description = "Summarise a file using the coordinator model." -parameters_schema = { - "type": "object", - "properties": {"path": {"type": "string"}}, - "required": ["path"], -} - -async def execute(path: str, ctx): - content = await ctx.read_file(path) - # ctx.client gives access to the coordinator model - return f"Summary of {path}: {len(content)} chars" +### Skill directories + +| Location | Scope | +|----------|-------| +| `/.agents/skills/` | Project-local (also searched in parent directories up to root) | +| `~/.agents/skills/` | User-global | + +Project skills override user-global skills of the same name. The `.agents/skills/` path matches the Codex CLI convention, so skills written for Codex work in team-harness without changes. + +### Creating a skill + +```bash +mkdir -p .agents/skills/my-skill +cat > .agents/skills/my-skill/SKILL.md << 'EOF' +--- +name: my-skill +description: Summarize files and produce a brief report. Use when the user asks for a summary or overview. +--- + +# My Skill + +## Steps + +1. Read the target files using `read_file` +2. Summarize the key points +3. Write a brief report + +## Notes + +- Keep summaries under 500 words +- Focus on actionable insights +EOF ``` +### Skill naming rules + +- 1-64 characters, lowercase letters, digits, and hyphens only +- Must not start or end with a hyphen, no consecutive hyphens +- Directory name is the canonical skill name + +### Optional subdirectories + +| Directory | Purpose | +|-----------|---------| +| `scripts/` | Executable code the agent can run | +| `references/` | Additional documentation loaded on demand | +| `assets/` | Templates, data files, schemas | + +The agent reads these files on demand via its file tools — they are not loaded at startup. + ## Run logs Each run creates a directory under `~/.team-harness/runs//` containing: diff --git a/pyproject.toml b/pyproject.toml index 5342bad..ba49f18 100644 --- a/pyproject.toml +++ b/pyproject.toml @@ -44,6 +44,7 @@ dependencies = [ "openai>=1.78", "pydantic>=2.11", "prompt_toolkit>=3.0.43", + "pyyaml>=6.0", "rich>=14.0", "click>=8.1", ] diff --git a/src/team_harness/cli.py b/src/team_harness/cli.py index 35ca4c1..fd48e83 100644 --- a/src/team_harness/cli.py +++ b/src/team_harness/cli.py @@ -28,8 +28,7 @@ from team_harness.harness import _show_no_config_hint from team_harness.harness import _warn_provider_startup from team_harness.harness import TeamHarness -from team_harness.skills.loader import load_skills -from team_harness.skills.loader import SkillContext +from team_harness.skills.loader import load_skill_metadata from team_harness.tools import agent_tools from team_harness.tools import fs_tools from team_harness.tools import todo_tools @@ -267,7 +266,7 @@ async def _repl(**kwargs: Any) -> None: ctx = ContextTracker(model_id=config.model, model_limit=model_limit) ui = make_console(ctx=ctx, manager=manager, run_dir=run_dir, mode="auto") _warn_provider_startup(config, ui=ui) - skills = load_skills(cwd=config.cwd) + skills = load_skill_metadata(cwd=config.cwd) allowed_types = get_allowed_types(config) validate_templates(config=config, allowed_types=allowed_types) agent_tools.setup( @@ -279,11 +278,8 @@ async def _repl(**kwargs: Any) -> None: ) todo_tools.setup(run_dir=run_dir) fs_tools.setup_fs() - skill_ctx = SkillContext(client=client, config=config) registry = _build_registry( allowed_types=allowed_types, - skills=skills, - skill_ctx=skill_ctx, manager=manager, run_log=run_log, config=config, diff --git a/src/team_harness/config.py b/src/team_harness/config.py index 6201931..f063c10 100644 --- a/src/team_harness/config.py +++ b/src/team_harness/config.py @@ -18,7 +18,7 @@ LOCAL_CONFIG_DIR_NAME = ".team-harness" CONFIG_PATH = Path.home() / ".team-harness" / "config.toml" RUNS_DIR = Path.home() / ".team-harness" / "runs" -SKILLS_USER_DIR = Path.home() / ".team-harness" / "skills" +SKILLS_USER_DIR = Path.home() / ".agents" / "skills" PROMPT_FILE_MAX_BYTES = 100 * 1024 diff --git a/src/team_harness/coordinator/system_prompt.py b/src/team_harness/coordinator/system_prompt.py index f102a8a..8f5863a 100644 --- a/src/team_harness/coordinator/system_prompt.py +++ b/src/team_harness/coordinator/system_prompt.py @@ -225,6 +225,29 @@ ---""".strip() +def _render_skills_section(skills: list) -> str: + """Render the skills section for the system prompt.""" + if not skills: + return "" + lines = [ + "## Skills", + "", + "Skills are local instruction sets that provide specialized knowledge.", + "Available skills:", + "", + ] + for skill in skills: + lines.append(f"- **{skill.name}**: {skill.description} (file: {skill.path})") + lines.extend( + [ + "", + "To use a skill: read its SKILL.md file to get full instructions,", + "then follow those instructions. Load referenced files only as needed.", + ] + ) + return "\n".join(lines) + + def build_system_prompt( config: object, allowed_types: list[str], skills: list, session_output_dir: str ) -> str: @@ -248,11 +271,9 @@ def build_system_prompt( if extension: parts.append(extension) - if skills: - parts.append( - "Additional tools (skills) available:\n" - + "\n".join(f"- {skill.name}: {skill.description}" for skill in skills) - ) + skills_section = _render_skills_section(skills) + if skills_section: + parts.append(skills_section) worker_suffix = getattr(config, "worker_suffix", "") if worker_suffix: diff --git a/src/team_harness/harness.py b/src/team_harness/harness.py index b911f94..2ea012f 100644 --- a/src/team_harness/harness.py +++ b/src/team_harness/harness.py @@ -23,9 +23,7 @@ from team_harness.coordinator.loop import run from team_harness.coordinator.protocols import CoordinatorLike from team_harness.coordinator.system_prompt import build_system_prompt -from team_harness.skills.loader import load_skills -from team_harness.skills.loader import Skill -from team_harness.skills.loader import SkillContext +from team_harness.skills.loader import load_skill_metadata from team_harness.tools import shell_tools from team_harness.tools.agent_tools import build_agent_tool_bindings from team_harness.tools.fs_tools import build_fs_tool_bindings @@ -128,14 +126,11 @@ async def run(self, task: str) -> TeamHarnessResult: ) _show_no_config_hint(config, ui=ui) _warn_provider_startup(config, ui=ui) - skills = load_skills(cwd=config.cwd) + skills = load_skill_metadata(cwd=config.cwd) allowed_types = get_allowed_types(config) validate_templates(config=config, allowed_types=allowed_types) - skill_ctx = SkillContext(client=client, config=config) registry = _build_registry( allowed_types=allowed_types, - skills=skills, - skill_ctx=skill_ctx, manager=manager, run_log=run_log, config=config, @@ -415,20 +410,9 @@ def _make_run_id() -> str: ) -def _make_skill_wrapper(skill: Skill, ctx: SkillContext) -> Any: - """Create an async wrapper that invokes a skill with its context.""" - - async def _wrapper(**args: object) -> str: - return await skill.execute(ctx=ctx, **args) - - return _wrapper - - def _build_registry( *, allowed_types: list[str], - skills: list[Skill], - skill_ctx: SkillContext, manager: AgentManager, run_log: RunLogWriter, config: Config, @@ -464,17 +448,4 @@ def _build_registry( for schema, fn in todo_bindings: registry.register(schema=schema, fn=fn) - # Skills - for skill in skills: - registry.register( - schema={ - "type": "function", - "function": { - "name": skill.name, - "description": skill.description, - "parameters": skill.parameters_schema, - }, - }, - fn=_make_skill_wrapper(skill=skill, ctx=skill_ctx), - ) return registry diff --git a/src/team_harness/skills/__init__.py b/src/team_harness/skills/__init__.py index 8b13789..99196d4 100644 --- a/src/team_harness/skills/__init__.py +++ b/src/team_harness/skills/__init__.py @@ -1 +1,4 @@ +from team_harness.skills.loader import load_skill_metadata +from team_harness.skills.loader import SkillMetadata +__all__ = ["SkillMetadata", "load_skill_metadata"] diff --git a/src/team_harness/skills/loader.py b/src/team_harness/skills/loader.py index 72d1ca5..754c326 100644 --- a/src/team_harness/skills/loader.py +++ b/src/team_harness/skills/loader.py @@ -1,77 +1,177 @@ +"""Agent Skills standard loader. + +Discovers SKILL.md files from skill roots, parses YAML frontmatter, +and returns metadata for the coordinator system prompt. +""" + from dataclasses import dataclass -import importlib.util from pathlib import Path -from typing import Any -from typing import Callable -from typing import TYPE_CHECKING import warnings -from team_harness.config import SKILLS_USER_DIR +import yaml -if TYPE_CHECKING: - from team_harness.config import Config - from team_harness.coordinator.protocols import CoordinatorLike +_MAX_DEPTH = 6 +_MAX_DIRS_PER_ROOT = 2000 @dataclass -class SkillContext: - client: "CoordinatorLike" - config: "Config" +class SkillMetadata: + name: str + description: str + path: Path + skill_dir: Path - async def read_file(self, path: str) -> str: - from team_harness.tools.fs_tools import read_file - return await read_file(path) +def load_skill_metadata(*, cwd: str | Path | None = None) -> list[SkillMetadata]: + """Scan skill roots and return metadata for all discovered skills.""" + roots = _resolve_skill_roots(cwd) + seen_names: dict[str, SkillMetadata] = {} + for root in roots: + if not root.exists(): + continue + for skill in _scan_root(root): + # Project skills override user/global skills of the same name + if skill.name not in seen_names: + seen_names[skill.name] = skill + return list(seen_names.values()) - async def write_file(self, path: str, content: str) -> str: - from team_harness.tools.fs_tools import write_file - return await write_file(path, content) +def _resolve_skill_roots(cwd: str | Path | None) -> list[Path]: + """Return skill root directories in priority order (project first, then global).""" + resolved_cwd = Path(cwd).resolve() if cwd else Path.cwd().resolve() + roots: list[Path] = [] + # Walk parent directories from cwd up to filesystem root + # looking for .agents/skills/ directories (project-local) + current = resolved_cwd + visited: set[Path] = set() + while True: + candidate = current / ".agents" / "skills" + if candidate.is_dir() and candidate not in visited: + roots.append(candidate) + visited.add(candidate) + parent = current.parent + if parent == current: + break + current = parent -@dataclass -class Skill: - name: str - description: str - parameters_schema: dict - execute: Callable[..., Any] + # User-global skills + user_global = Path.home() / ".agents" / "skills" + if user_global not in visited: + roots.append(user_global) + return roots -def load_skills( - *, cwd: str | Path | None = None, extra_dirs: list[Path] | None = None -) -> list[Skill]: - root = Path(cwd).resolve() if cwd else Path.cwd() - skill_dirs = [SKILLS_USER_DIR, root / "skills"] - if extra_dirs: - skill_dirs.extend(extra_dirs) - skills: list[Skill] = [] - for directory in skill_dirs: - if not directory.exists(): +def _scan_root(root: Path) -> list[SkillMetadata]: + """BFS scan a skill root for SKILL.md files, max depth _MAX_DEPTH.""" + skills: list[SkillMetadata] = [] + dirs_scanned = 0 + queue: list[tuple[Path, int]] = [(root, 0)] + + while queue and dirs_scanned < _MAX_DIRS_PER_ROOT: + directory, depth = queue.pop(0) + dirs_scanned += 1 + + skill_md = directory / "SKILL.md" + if skill_md.is_file(): + skill = _parse_skill(skill_md) + if skill is not None: + skills.append(skill) + continue # Don't recurse into skill directories + + if depth >= _MAX_DEPTH: + continue + + try: + children = sorted(directory.iterdir()) + except OSError: continue - for path in sorted(directory.glob("*.py")): - module_name = f"team_harness_skill_{path.stem}_{abs(hash(path))}" - try: - spec = importlib.util.spec_from_file_location(module_name, path) - if spec is None or spec.loader is None: - raise ImportError(f"Unable to load spec for {path}") - module = importlib.util.module_from_spec(spec) - spec.loader.exec_module(module) - except Exception as exc: - warnings.warn(f"Failed to load skill {path}: {exc}", stacklevel=2) - continue - required = ("name", "description", "parameters_schema", "execute") - if not all(hasattr(module, attr) for attr in required): - warnings.warn( - f"Skipping skill {path}: missing one of {required}", stacklevel=2 - ) - continue - skills.append( - Skill( - name=module.name, - description=module.description, - parameters_schema=module.parameters_schema, - execute=module.execute, - ) - ) + + for child in children: + if child.is_dir() and not child.name.startswith("."): + queue.append((child, depth + 1)) + return skills + + +def _parse_skill(skill_md: Path) -> SkillMetadata | None: + """Parse a SKILL.md file and return metadata, or None if invalid.""" + try: + text = skill_md.read_text(encoding="utf-8") + except (OSError, UnicodeDecodeError) as exc: + warnings.warn(f"Cannot read {skill_md}: {exc}", stacklevel=2) + return None + + frontmatter = _extract_frontmatter(text) + if frontmatter is None: + warnings.warn(f"No YAML frontmatter in {skill_md}", stacklevel=2) + return None + + # Directory name is canonical (per review feedback) + dir_name = skill_md.parent.name + fm_name = frontmatter.get("name") + if fm_name and fm_name != dir_name: + warnings.warn( + f"Skill {skill_md}: frontmatter name '{fm_name}' does not match " + f"directory name '{dir_name}'; using directory name", + stacklevel=2, + ) + name = dir_name + + # Validate name + if not _is_valid_name(name): + warnings.warn(f"Invalid skill name '{name}' in {skill_md}", stacklevel=2) + return None + + raw_description = frontmatter.get("description", "") + if not raw_description: + warnings.warn(f"Missing description in {skill_md}", stacklevel=2) + return None + + description = str(raw_description) + if len(description) > 1024: + description = description[:1024] + + return SkillMetadata( + name=name, description=description, path=skill_md, skill_dir=skill_md.parent + ) + + +def _extract_frontmatter(text: str) -> dict | None: + """Extract YAML frontmatter from between --- delimiters.""" + lines = text.split("\n") + if not lines or lines[0].strip() != "---": + return None + + end_idx = None + for i, line in enumerate(lines[1:], 1): + if line.strip() == "---": + end_idx = i + break + + if end_idx is None: + return None + + yaml_text = "\n".join(lines[1:end_idx]) + try: + data = yaml.safe_load(yaml_text) + except yaml.YAMLError as exc: + warnings.warn(f"Invalid YAML frontmatter: {exc}", stacklevel=2) + return None + + if not isinstance(data, dict): + return None + + return data + + +def _is_valid_name(name: str) -> bool: + """Check if a skill name follows the naming convention.""" + if not name or len(name) > 64: + return False + if name.startswith("-") or name.endswith("-"): + return False + if "--" in name: + return False + return all(c.isalnum() or c == "-" for c in name) and name == name.lower() diff --git a/src/tests/test_cli.py b/src/tests/test_cli.py index bebb825..be19e30 100644 --- a/src/tests/test_cli.py +++ b/src/tests/test_cli.py @@ -243,7 +243,7 @@ async def fake_run(messages, config, run_log, ui, tool_registry, client, ctx): "team_harness.harness.resolve_model_limit", fake_resolve_model_limit ) monkeypatch.setattr("team_harness.harness.make_console", lambda **_: FakeConsole()) - monkeypatch.setattr("team_harness.harness.load_skills", lambda cwd=None: []) + monkeypatch.setattr("team_harness.harness.load_skill_metadata", lambda cwd=None: []) monkeypatch.setattr( "team_harness.harness.validate_templates", lambda *args, **kwargs: None ) @@ -346,7 +346,7 @@ def fake_build_system_prompt(*, config, allowed_types, skills, session_output_di "team_harness.cli.resolve_model_limit", fake_resolve_model_limit ) monkeypatch.setattr("team_harness.cli.make_console", lambda **_: FakeConsole()) - monkeypatch.setattr("team_harness.cli.load_skills", lambda cwd=None: []) + monkeypatch.setattr("team_harness.cli.load_skill_metadata", lambda cwd=None: []) monkeypatch.setattr( "team_harness.cli.validate_templates", lambda *args, **kwargs: None ) @@ -452,7 +452,7 @@ def fake_build_system_prompt(*, config, allowed_types, skills, session_output_di "team_harness.cli.resolve_model_limit", fake_resolve_model_limit ) monkeypatch.setattr("team_harness.cli.make_console", lambda **_: FakeConsole()) - monkeypatch.setattr("team_harness.cli.load_skills", lambda cwd=None: []) + monkeypatch.setattr("team_harness.cli.load_skill_metadata", lambda cwd=None: []) monkeypatch.setattr( "team_harness.cli.validate_templates", lambda *args, **kwargs: None ) @@ -598,7 +598,7 @@ async def fake_run_one_turn(**kwargs): "team_harness.cli.resolve_model_limit", fake_resolve_model_limit ) monkeypatch.setattr("team_harness.cli.make_console", lambda **_: fake_console) - monkeypatch.setattr("team_harness.cli.load_skills", lambda cwd=None: []) + monkeypatch.setattr("team_harness.cli.load_skill_metadata", lambda cwd=None: []) monkeypatch.setattr( "team_harness.cli.validate_templates", lambda *args, **kwargs: None ) @@ -719,7 +719,7 @@ async def fake_run_one_turn(**kwargs): "team_harness.cli.resolve_model_limit", fake_resolve_model_limit ) monkeypatch.setattr("team_harness.cli.make_console", lambda **_: fake_console) - monkeypatch.setattr("team_harness.cli.load_skills", lambda cwd=None: []) + monkeypatch.setattr("team_harness.cli.load_skill_metadata", lambda cwd=None: []) monkeypatch.setattr( "team_harness.cli.validate_templates", lambda *args, **kwargs: None ) @@ -852,7 +852,7 @@ async def fake_perform_manual_compaction(*, messages, client, ctx, ui, focus_tex "team_harness.cli.resolve_model_limit", fake_resolve_model_limit ) monkeypatch.setattr("team_harness.cli.make_console", lambda **_: fake_console) - monkeypatch.setattr("team_harness.cli.load_skills", lambda cwd=None: []) + monkeypatch.setattr("team_harness.cli.load_skill_metadata", lambda cwd=None: []) monkeypatch.setattr( "team_harness.cli.validate_templates", lambda *args, **kwargs: None ) @@ -1009,7 +1009,7 @@ async def fake_perform_manual_compaction(*, messages, client, ctx, ui, focus_tex "team_harness.cli.resolve_model_limit", fake_resolve_model_limit ) monkeypatch.setattr("team_harness.cli.make_console", lambda **_: fake_console) - monkeypatch.setattr("team_harness.cli.load_skills", lambda cwd=None: []) + monkeypatch.setattr("team_harness.cli.load_skill_metadata", lambda cwd=None: []) monkeypatch.setattr( "team_harness.cli.validate_templates", lambda *args, **kwargs: None ) @@ -1148,7 +1148,7 @@ async def fake_perform_manual_compaction(*, messages, client, ctx, ui, focus_tex "team_harness.cli.resolve_model_limit", fake_resolve_model_limit ) monkeypatch.setattr("team_harness.cli.make_console", lambda **_: fake_console) - monkeypatch.setattr("team_harness.cli.load_skills", lambda cwd=None: []) + monkeypatch.setattr("team_harness.cli.load_skill_metadata", lambda cwd=None: []) monkeypatch.setattr( "team_harness.cli.validate_templates", lambda *args, **kwargs: None ) @@ -1272,7 +1272,7 @@ async def fake_perform_manual_compaction(*, messages, client, ctx, ui, focus_tex "team_harness.cli.resolve_model_limit", fake_resolve_model_limit ) monkeypatch.setattr("team_harness.cli.make_console", lambda **_: fake_console) - monkeypatch.setattr("team_harness.cli.load_skills", lambda cwd=None: []) + monkeypatch.setattr("team_harness.cli.load_skill_metadata", lambda cwd=None: []) monkeypatch.setattr( "team_harness.cli.validate_templates", lambda *args, **kwargs: None ) @@ -1391,7 +1391,7 @@ async def fake_run_one_turn(**kwargs): "team_harness.cli.resolve_model_limit", fake_resolve_model_limit ) monkeypatch.setattr("team_harness.cli.make_console", lambda **_: FakeConsole()) - monkeypatch.setattr("team_harness.cli.load_skills", lambda cwd=None: []) + monkeypatch.setattr("team_harness.cli.load_skill_metadata", lambda cwd=None: []) monkeypatch.setattr( "team_harness.cli.validate_templates", lambda *args, **kwargs: None ) diff --git a/src/tests/test_config.py b/src/tests/test_config.py index 93e5786..298aa11 100644 --- a/src/tests/test_config.py +++ b/src/tests/test_config.py @@ -729,7 +729,7 @@ def test_default_config_text_contains_verified_flag_tokens(): def test_config_paths_remain_under_team_harness_dir(): assert CONFIG_PATH == config_module.Path.home() / ".team-harness" / "config.toml" assert RUNS_DIR == config_module.Path.home() / ".team-harness" / "runs" - assert SKILLS_USER_DIR == config_module.Path.home() / ".team-harness" / "skills" + assert SKILLS_USER_DIR == config_module.Path.home() / ".agents" / "skills" def test_provider_aware_codex_defaults(tmp_path, monkeypatch): diff --git a/src/tests/test_harness.py b/src/tests/test_harness.py index da0cf6d..67325c8 100644 --- a/src/tests/test_harness.py +++ b/src/tests/test_harness.py @@ -280,7 +280,7 @@ def fake_build_system_prompt(*, config, allowed_types, skills, session_output_di monkeypatch.setattr( "team_harness.harness.resolve_model_limit", fake_resolve_model_limit ) - monkeypatch.setattr("team_harness.harness.load_skills", lambda cwd=None: []) + monkeypatch.setattr("team_harness.harness.load_skill_metadata", lambda cwd=None: []) monkeypatch.setattr( "team_harness.harness.build_system_prompt", fake_build_system_prompt ) diff --git a/src/tests/test_integration.py b/src/tests/test_integration.py index f6700d5..dee7f09 100644 --- a/src/tests/test_integration.py +++ b/src/tests/test_integration.py @@ -70,14 +70,16 @@ async def aclose(self): ) captured: dict[str, str | None] = {"cwd": None} - def fake_load_skills(*, cwd=None): + def fake_load_skill_metadata(*, cwd=None): captured["cwd"] = str(cwd) if cwd is not None else None return [] other_cwd = tmp_path / "elsewhere" other_cwd.mkdir() monkeypatch.chdir(other_cwd) - monkeypatch.setattr("team_harness.harness.load_skills", fake_load_skills) + monkeypatch.setattr( + "team_harness.harness.load_skill_metadata", fake_load_skill_metadata + ) await _run( task="hello", task_file=None, diff --git a/src/tests/test_skills.py b/src/tests/test_skills.py index e866a2a..e6bbc69 100644 --- a/src/tests/test_skills.py +++ b/src/tests/test_skills.py @@ -1,117 +1,282 @@ # pyright: reportMissingParameterType=false, reportArgumentType=false -import textwrap +from pathlib import Path import pytest -from team_harness.config import Config -from team_harness.coordinator.client import ChatResponse -from team_harness.coordinator.client import ChoiceRecord -from team_harness.coordinator.client import MessageRecord -from team_harness.skills.loader import load_skills -from team_harness.skills.loader import SkillContext -from team_harness.tools import fs_tools -from team_harness.tools.registry import ToolRegistry - - -class SkillClient: - model = "test/model" - api_base = "http://localhost:9999" - provider = "openai_compat" - - async def chat(self, messages, tools=None, stream=False, token_callback=None): - return ChatResponse( - choices=[ChoiceRecord(message=MessageRecord(content="summary"))] - ) - - async def get_models(self): - return {"data": []} - - async def aclose(self): - return None - - -@pytest.mark.asyncio -async def test_load_skills_and_ctx(tmp_path): - valid = tmp_path / "valid.py" - valid.write_text( - textwrap.dedent( - """ - name = "demo" - description = "demo skill" - parameters_schema = {"type": "object", "properties": {"path": {"type": "string"}}, "required": ["path"]} - async def execute(path: str, ctx): - return await ctx.read_file(path) - """ - ) +from team_harness.skills.loader import load_skill_metadata + + +@pytest.fixture(autouse=True) +def _isolate_home(tmp_path, monkeypatch): + """Prevent discovery of real ~/.agents/skills/ during tests.""" + fake_home = tmp_path / "fake-home" + fake_home.mkdir() + monkeypatch.setattr(Path, "home", classmethod(lambda cls: fake_home)) + + +def _make_skill( + base: Path, + name: str, + *, + description: str = "A test skill", + fm_name: str | None = None, +) -> Path: + """Create a SKILL.md in base/name/ with YAML frontmatter.""" + skill_dir = base / name + skill_dir.mkdir(parents=True, exist_ok=True) + fm_name_line = f"name: {fm_name}\n" if fm_name else "" + (skill_dir / "SKILL.md").write_text( + f"---\n{fm_name_line}description: {description}\n---\n\n# {name}\n\nInstructions here.\n", + encoding="utf-8", + ) + return skill_dir + + +def test_discover_skills_from_project_dir(tmp_path): + project = tmp_path / "project" + project.mkdir() + skills_root = project / ".agents" / "skills" + _make_skill(skills_root, "my-skill", description="My skill does things") + + skills = load_skill_metadata(cwd=str(project)) + + assert len(skills) == 1 + assert skills[0].name == "my-skill" + assert skills[0].description == "My skill does things" + assert skills[0].path == skills_root / "my-skill" / "SKILL.md" + assert skills[0].skill_dir == skills_root / "my-skill" + + +def test_frontmatter_parsing(tmp_path): + project = tmp_path / "project" + project.mkdir() + skills_root = project / ".agents" / "skills" + _make_skill( + skills_root, "parser-skill", description="Parses things", fm_name="parser-skill" + ) + + skills = load_skill_metadata(cwd=str(project)) + + assert len(skills) == 1 + assert skills[0].name == "parser-skill" + assert skills[0].description == "Parses things" + + +def test_missing_name_falls_back_to_dir_name(tmp_path): + project = tmp_path / "project" + project.mkdir() + skills_root = project / ".agents" / "skills" + # No fm_name => no name field in frontmatter + _make_skill(skills_root, "fallback-skill", description="Has no name field") + + skills = load_skill_metadata(cwd=str(project)) + + assert len(skills) == 1 + assert skills[0].name == "fallback-skill" + + +def test_missing_description_skips_with_warning(tmp_path): + project = tmp_path / "project" + project.mkdir() + skill_dir = project / ".agents" / "skills" / "no-desc" + skill_dir.mkdir(parents=True) + (skill_dir / "SKILL.md").write_text( + "---\nname: no-desc\n---\n\nBody.\n", encoding="utf-8" + ) + + with pytest.warns(UserWarning, match="Missing description"): + skills = load_skill_metadata(cwd=str(project)) + + assert len(skills) == 0 + + +def test_invalid_name_skipped(tmp_path): + project = tmp_path / "project" + project.mkdir() + skills_root = project / ".agents" / "skills" + # Uppercase name directory + skill_dir = skills_root / "BadName" + skill_dir.mkdir(parents=True) + (skill_dir / "SKILL.md").write_text( + "---\ndescription: Bad name skill\n---\n\nBody.\n", encoding="utf-8" + ) + + with pytest.warns(UserWarning, match="Invalid skill name"): + skills = load_skill_metadata(cwd=str(project)) + + assert len(skills) == 0 + + +def test_no_skill_dirs_returns_empty(tmp_path): + project = tmp_path / "project" + project.mkdir() + + skills = load_skill_metadata(cwd=str(project)) + + assert skills == [] + + +def test_non_utf8_file_produces_warning_and_is_skipped(tmp_path): + project = tmp_path / "project" + project.mkdir() + skill_dir = project / ".agents" / "skills" / "binary-skill" + skill_dir.mkdir(parents=True) + (skill_dir / "SKILL.md").write_bytes(b"\xff\xfe\x00\x00invalid") + + with pytest.warns(UserWarning, match="Cannot read"): + skills = load_skill_metadata(cwd=str(project)) + + assert len(skills) == 0 + + +def test_dotfile_dirs_skipped(tmp_path): + project = tmp_path / "project" + project.mkdir() + hidden_dir = project / ".agents" / "skills" / ".hidden" + hidden_dir.mkdir(parents=True) + (hidden_dir / "SKILL.md").write_text( + "---\ndescription: Hidden skill\n---\n\nBody.\n", encoding="utf-8" ) - invalid = tmp_path / "invalid.py" - invalid.write_text("x = 1") - with pytest.warns(UserWarning): - skills = load_skills(extra_dirs=[tmp_path]) - names = [skill.name for skill in skills] - assert "demo" in names - - fs_tools.setup_fs() - target = tmp_path / "note.txt" - target.write_text("hello") - ctx = SkillContext(client=SkillClient(), config=Config()) - registry = ToolRegistry() - - def _make_wrapper(skill, skill_ctx): - async def _wrapper(**args: object) -> str: - return await skill.execute(ctx=skill_ctx, **args) - - return _wrapper - - skill = next(skill for skill in skills if skill.name == "demo") - registry.register( - schema={ - "type": "function", - "function": { - "name": skill.name, - "description": skill.description, - "parameters": skill.parameters_schema, - }, - }, - fn=_make_wrapper(skill=skill, skill_ctx=ctx), + + skills = load_skill_metadata(cwd=str(project)) + + assert len(skills) == 0 + + +def test_name_mismatch_produces_warning_but_uses_dir_name(tmp_path): + project = tmp_path / "project" + project.mkdir() + skills_root = project / ".agents" / "skills" + _make_skill( + skills_root, "real-name", description="Mismatched", fm_name="wrong-name" + ) + + with pytest.warns(UserWarning, match="does not match"): + skills = load_skill_metadata(cwd=str(project)) + + assert len(skills) == 1 + assert skills[0].name == "real-name" + + +def test_project_skills_override_user_skills(tmp_path): + # _isolate_home already patches Path.home to fake_home + fake_home = Path.home() + project = tmp_path / "project" + project.mkdir() + + project_skills = project / ".agents" / "skills" + _make_skill(project_skills, "shared-skill", description="Project version") + + user_skills = fake_home / ".agents" / "skills" + _make_skill(user_skills, "shared-skill", description="User version") + + skills = load_skill_metadata(cwd=str(project)) + + assert len(skills) == 1 + assert skills[0].description == "Project version" + + +def test_no_frontmatter_produces_warning(tmp_path): + project = tmp_path / "project" + project.mkdir() + skill_dir = project / ".agents" / "skills" / "no-fm" + skill_dir.mkdir(parents=True) + (skill_dir / "SKILL.md").write_text( + "# Just a heading\n\nNo frontmatter.\n", encoding="utf-8" + ) + + with pytest.warns(UserWarning, match="No YAML frontmatter"): + skills = load_skill_metadata(cwd=str(project)) + + assert len(skills) == 0 + + +def test_discovers_repo_root_skills_from_nested_cwd(tmp_path): + repo = tmp_path / "repo" + nested = repo / "src" / "deep" / "dir" + nested.mkdir(parents=True) + skills_root = repo / ".agents" / "skills" + _make_skill(skills_root, "root-skill", description="Found from nested cwd") + + skills = load_skill_metadata(cwd=str(nested)) + + assert len(skills) == 1 + assert skills[0].name == "root-skill" + + +def test_invalid_yaml_frontmatter_warns_and_skips(tmp_path): + project = tmp_path / "project" + project.mkdir() + skill_dir = project / ".agents" / "skills" / "bad-yaml" + skill_dir.mkdir(parents=True) + (skill_dir / "SKILL.md").write_text( + "---\n: invalid: yaml: [broken\n---\n\nBody.\n", encoding="utf-8" ) - assert ( - await registry.execute(name="demo", arguments={"path": str(target)}) == "hello" + + with pytest.warns(UserWarning, match="Invalid YAML frontmatter"): + skills = load_skill_metadata(cwd=str(project)) + + assert len(skills) == 0 + + +def test_non_mapping_frontmatter_skips(tmp_path): + project = tmp_path / "project" + project.mkdir() + skill_dir = project / ".agents" / "skills" / "list-yaml" + skill_dir.mkdir(parents=True) + (skill_dir / "SKILL.md").write_text( + "---\n- item1\n- item2\n---\n\nBody.\n", encoding="utf-8" ) + with pytest.warns(UserWarning, match="No YAML frontmatter"): + skills = load_skill_metadata(cwd=str(project)) + + assert len(skills) == 0 -def test_load_skills_uses_passed_cwd_for_project_skills(monkeypatch, tmp_path): - user_dir = tmp_path / "user-skills" - cwd_one = tmp_path / "project-one" - cwd_two = tmp_path / "project-two" - (cwd_one / "skills").mkdir(parents=True) - (cwd_two / "skills").mkdir(parents=True) - (cwd_one / "skills" / "one.py").write_text( - textwrap.dedent( - """ - name = "one" - description = "project one" - parameters_schema = {"type": "object", "properties": {}} - async def execute(ctx): - return "one" - """ - ) + +def test_discover_skills_from_user_dir(tmp_path): + fake_home = Path.home() + project = tmp_path / "project" + project.mkdir() + # No project skills — only user global + user_skills = fake_home / ".agents" / "skills" + _make_skill(user_skills, "global-skill", description="User global skill") + + skills = load_skill_metadata(cwd=str(project)) + + assert len(skills) == 1 + assert skills[0].name == "global-skill" + + +@pytest.mark.parametrize( + "dirname", ["-leading", "trailing-", "double--hyphen", "has_underscore", "A" * 65] +) +def test_invalid_name_variants_rejected(tmp_path, dirname): + project = tmp_path / "project" + project.mkdir() + skill_dir = project / ".agents" / "skills" / dirname + skill_dir.mkdir(parents=True) + (skill_dir / "SKILL.md").write_text( + "---\ndescription: Bad name variant\n---\n\nBody.\n", encoding="utf-8" ) - (cwd_two / "skills" / "two.py").write_text( - textwrap.dedent( - """ - name = "two" - description = "project two" - parameters_schema = {"type": "object", "properties": {}} - async def execute(ctx): - return "two" - """ - ) + + with pytest.warns(UserWarning, match="Invalid skill name"): + skills = load_skill_metadata(cwd=str(project)) + + assert len(skills) == 0 + + +def test_non_string_description_coerced(tmp_path): + project = tmp_path / "project" + project.mkdir() + skill_dir = project / ".agents" / "skills" / "int-desc" + skill_dir.mkdir(parents=True) + (skill_dir / "SKILL.md").write_text( + "---\ndescription: 123\n---\n\nBody.\n", encoding="utf-8" ) - monkeypatch.setattr("team_harness.skills.loader.SKILLS_USER_DIR", user_dir) - monkeypatch.chdir(cwd_two) - skills = load_skills(cwd=cwd_one) + skills = load_skill_metadata(cwd=str(project)) - assert [skill.name for skill in skills] == ["one"] + assert len(skills) == 1 + assert skills[0].description == "123" diff --git a/src/tests/test_system_prompt.py b/src/tests/test_system_prompt.py index ec5901b..82af18c 100644 --- a/src/tests/test_system_prompt.py +++ b/src/tests/test_system_prompt.py @@ -1,11 +1,12 @@ # pyright: reportMissingParameterType=false -from types import SimpleNamespace +from pathlib import Path from team_harness.config import Config from team_harness.coordinator.system_prompt import build_system_prompt from team_harness.coordinator.system_prompt import COORDINATOR_PROMPT from team_harness.coordinator.system_prompt import DEFAULT_WORKER_FOOTER +from team_harness.skills.loader import SkillMetadata def test_system_prompt_contains_coordinator_identity(): @@ -68,16 +69,27 @@ def test_system_prompt_appends_extensions_and_skills(): config=Config(cwd="/repo", system_prompt_extension="Extra config rules"), allowed_types=["codex"], skills=[ - SimpleNamespace(name="skill-a", description="First skill"), - SimpleNamespace(name="skill-b", description="Second skill"), + SkillMetadata( + name="skill-a", + description="First skill", + path=Path("/repo/.agents/skills/skill-a/SKILL.md"), + skill_dir=Path("/repo/.agents/skills/skill-a"), + ), + SkillMetadata( + name="skill-b", + description="Second skill", + path=Path("/repo/.agents/skills/skill-b/SKILL.md"), + skill_dir=Path("/repo/.agents/skills/skill-b"), + ), ], session_output_dir="/repo/_outputs/run_123", ) assert "Extra config rules" in prompt - assert "Additional tools (skills) available:" in prompt - assert "- skill-a: First skill" in prompt - assert "- skill-b: Second skill" in prompt + assert "## Skills" in prompt + assert "**skill-a**: First skill" in prompt + assert "**skill-b**: Second skill" in prompt + assert "To use a skill: read its SKILL.md file" in prompt def test_build_system_prompt_uses_config_coordinator_prompt(): diff --git a/uv.lock b/uv.lock index 942526b..dd081d1 100644 --- a/uv.lock +++ b/uv.lock @@ -409,6 +409,52 @@ wheels = [ { url = "https://files.pythonhosted.org/packages/e5/35/f8b19922b6a25bc0880171a2f1a003eaeb93657475193ab516fd87cac9da/pytest_asyncio-1.3.0-py3-none-any.whl", hash = "sha256:611e26147c7f77640e6d0a92a38ed17c3e9848063698d5c93d5aa7aa11cebff5", size = 15075 }, ] +[[package]] +name = "pyyaml" +version = "6.0.3" +source = { registry = "https://pypi.org/simple" } +sdist = { url = "https://files.pythonhosted.org/packages/05/8e/961c0007c59b8dd7729d542c61a4d537767a59645b82a0b521206e1e25c2/pyyaml-6.0.3.tar.gz", hash = "sha256:d76623373421df22fb4cf8817020cbb7ef15c725b9d5e45f17e189bfc384190f", size = 130960 } +wheels = [ + { url = "https://files.pythonhosted.org/packages/d1/33/422b98d2195232ca1826284a76852ad5a86fe23e31b009c9886b2d0fb8b2/pyyaml-6.0.3-cp312-cp312-macosx_10_13_x86_64.whl", hash = "sha256:7f047e29dcae44602496db43be01ad42fc6f1cc0d8cd6c83d342306c32270196", size = 182063 }, + { url = "https://files.pythonhosted.org/packages/89/a0/6cf41a19a1f2f3feab0e9c0b74134aa2ce6849093d5517a0c550fe37a648/pyyaml-6.0.3-cp312-cp312-macosx_11_0_arm64.whl", hash = "sha256:fc09d0aa354569bc501d4e787133afc08552722d3ab34836a80547331bb5d4a0", size = 173973 }, + { url = "https://files.pythonhosted.org/packages/ed/23/7a778b6bd0b9a8039df8b1b1d80e2e2ad78aa04171592c8a5c43a56a6af4/pyyaml-6.0.3-cp312-cp312-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:9149cad251584d5fb4981be1ecde53a1ca46c891a79788c0df828d2f166bda28", size = 775116 }, + { url = "https://files.pythonhosted.org/packages/65/30/d7353c338e12baef4ecc1b09e877c1970bd3382789c159b4f89d6a70dc09/pyyaml-6.0.3-cp312-cp312-manylinux2014_s390x.manylinux_2_17_s390x.manylinux_2_28_s390x.whl", hash = "sha256:5fdec68f91a0c6739b380c83b951e2c72ac0197ace422360e6d5a959d8d97b2c", size = 844011 }, + { url = "https://files.pythonhosted.org/packages/8b/9d/b3589d3877982d4f2329302ef98a8026e7f4443c765c46cfecc8858c6b4b/pyyaml-6.0.3-cp312-cp312-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:ba1cc08a7ccde2d2ec775841541641e4548226580ab850948cbfda66a1befcdc", size = 807870 }, + { url = "https://files.pythonhosted.org/packages/05/c0/b3be26a015601b822b97d9149ff8cb5ead58c66f981e04fedf4e762f4bd4/pyyaml-6.0.3-cp312-cp312-musllinux_1_2_aarch64.whl", hash = "sha256:8dc52c23056b9ddd46818a57b78404882310fb473d63f17b07d5c40421e47f8e", size = 761089 }, + { url = "https://files.pythonhosted.org/packages/be/8e/98435a21d1d4b46590d5459a22d88128103f8da4c2d4cb8f14f2a96504e1/pyyaml-6.0.3-cp312-cp312-musllinux_1_2_x86_64.whl", hash = "sha256:41715c910c881bc081f1e8872880d3c650acf13dfa8214bad49ed4cede7c34ea", size = 790181 }, + { url = "https://files.pythonhosted.org/packages/74/93/7baea19427dcfbe1e5a372d81473250b379f04b1bd3c4c5ff825e2327202/pyyaml-6.0.3-cp312-cp312-win32.whl", hash = "sha256:96b533f0e99f6579b3d4d4995707cf36df9100d67e0c8303a0c55b27b5f99bc5", size = 137658 }, + { url = "https://files.pythonhosted.org/packages/86/bf/899e81e4cce32febab4fb42bb97dcdf66bc135272882d1987881a4b519e9/pyyaml-6.0.3-cp312-cp312-win_amd64.whl", hash = "sha256:5fcd34e47f6e0b794d17de1b4ff496c00986e1c83f7ab2fb8fcfe9616ff7477b", size = 154003 }, + { url = "https://files.pythonhosted.org/packages/1a/08/67bd04656199bbb51dbed1439b7f27601dfb576fb864099c7ef0c3e55531/pyyaml-6.0.3-cp312-cp312-win_arm64.whl", hash = "sha256:64386e5e707d03a7e172c0701abfb7e10f0fb753ee1d773128192742712a98fd", size = 140344 }, + { url = "https://files.pythonhosted.org/packages/d1/11/0fd08f8192109f7169db964b5707a2f1e8b745d4e239b784a5a1dd80d1db/pyyaml-6.0.3-cp313-cp313-macosx_10_13_x86_64.whl", hash = "sha256:8da9669d359f02c0b91ccc01cac4a67f16afec0dac22c2ad09f46bee0697eba8", size = 181669 }, + { url = "https://files.pythonhosted.org/packages/b1/16/95309993f1d3748cd644e02e38b75d50cbc0d9561d21f390a76242ce073f/pyyaml-6.0.3-cp313-cp313-macosx_11_0_arm64.whl", hash = "sha256:2283a07e2c21a2aa78d9c4442724ec1eb15f5e42a723b99cb3d822d48f5f7ad1", size = 173252 }, + { url = "https://files.pythonhosted.org/packages/50/31/b20f376d3f810b9b2371e72ef5adb33879b25edb7a6d072cb7ca0c486398/pyyaml-6.0.3-cp313-cp313-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:ee2922902c45ae8ccada2c5b501ab86c36525b883eff4255313a253a3160861c", size = 767081 }, + { url = "https://files.pythonhosted.org/packages/49/1e/a55ca81e949270d5d4432fbbd19dfea5321eda7c41a849d443dc92fd1ff7/pyyaml-6.0.3-cp313-cp313-manylinux2014_s390x.manylinux_2_17_s390x.manylinux_2_28_s390x.whl", hash = "sha256:a33284e20b78bd4a18c8c2282d549d10bc8408a2a7ff57653c0cf0b9be0afce5", size = 841159 }, + { url = "https://files.pythonhosted.org/packages/74/27/e5b8f34d02d9995b80abcef563ea1f8b56d20134d8f4e5e81733b1feceb2/pyyaml-6.0.3-cp313-cp313-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:0f29edc409a6392443abf94b9cf89ce99889a1dd5376d94316ae5145dfedd5d6", size = 801626 }, + { url = "https://files.pythonhosted.org/packages/f9/11/ba845c23988798f40e52ba45f34849aa8a1f2d4af4b798588010792ebad6/pyyaml-6.0.3-cp313-cp313-musllinux_1_2_aarch64.whl", hash = "sha256:f7057c9a337546edc7973c0d3ba84ddcdf0daa14533c2065749c9075001090e6", size = 753613 }, + { url = "https://files.pythonhosted.org/packages/3d/e0/7966e1a7bfc0a45bf0a7fb6b98ea03fc9b8d84fa7f2229e9659680b69ee3/pyyaml-6.0.3-cp313-cp313-musllinux_1_2_x86_64.whl", hash = "sha256:eda16858a3cab07b80edaf74336ece1f986ba330fdb8ee0d6c0d68fe82bc96be", size = 794115 }, + { url = "https://files.pythonhosted.org/packages/de/94/980b50a6531b3019e45ddeada0626d45fa85cbe22300844a7983285bed3b/pyyaml-6.0.3-cp313-cp313-win32.whl", hash = "sha256:d0eae10f8159e8fdad514efdc92d74fd8d682c933a6dd088030f3834bc8e6b26", size = 137427 }, + { url = "https://files.pythonhosted.org/packages/97/c9/39d5b874e8b28845e4ec2202b5da735d0199dbe5b8fb85f91398814a9a46/pyyaml-6.0.3-cp313-cp313-win_amd64.whl", hash = "sha256:79005a0d97d5ddabfeeea4cf676af11e647e41d81c9a7722a193022accdb6b7c", size = 154090 }, + { url = "https://files.pythonhosted.org/packages/73/e8/2bdf3ca2090f68bb3d75b44da7bbc71843b19c9f2b9cb9b0f4ab7a5a4329/pyyaml-6.0.3-cp313-cp313-win_arm64.whl", hash = "sha256:5498cd1645aa724a7c71c8f378eb29ebe23da2fc0d7a08071d89469bf1d2defb", size = 140246 }, + { url = "https://files.pythonhosted.org/packages/9d/8c/f4bd7f6465179953d3ac9bc44ac1a8a3e6122cf8ada906b4f96c60172d43/pyyaml-6.0.3-cp314-cp314-macosx_10_13_x86_64.whl", hash = "sha256:8d1fab6bb153a416f9aeb4b8763bc0f22a5586065f86f7664fc23339fc1c1fac", size = 181814 }, + { url = "https://files.pythonhosted.org/packages/bd/9c/4d95bb87eb2063d20db7b60faa3840c1b18025517ae857371c4dd55a6b3a/pyyaml-6.0.3-cp314-cp314-macosx_11_0_arm64.whl", hash = "sha256:34d5fcd24b8445fadc33f9cf348c1047101756fd760b4dacb5c3e99755703310", size = 173809 }, + { url = "https://files.pythonhosted.org/packages/92/b5/47e807c2623074914e29dabd16cbbdd4bf5e9b2db9f8090fa64411fc5382/pyyaml-6.0.3-cp314-cp314-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:501a031947e3a9025ed4405a168e6ef5ae3126c59f90ce0cd6f2bfc477be31b7", size = 766454 }, + { url = "https://files.pythonhosted.org/packages/02/9e/e5e9b168be58564121efb3de6859c452fccde0ab093d8438905899a3a483/pyyaml-6.0.3-cp314-cp314-manylinux2014_s390x.manylinux_2_17_s390x.manylinux_2_28_s390x.whl", hash = "sha256:b3bc83488de33889877a0f2543ade9f70c67d66d9ebb4ac959502e12de895788", size = 836355 }, + { url = "https://files.pythonhosted.org/packages/88/f9/16491d7ed2a919954993e48aa941b200f38040928474c9e85ea9e64222c3/pyyaml-6.0.3-cp314-cp314-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:c458b6d084f9b935061bc36216e8a69a7e293a2f1e68bf956dcd9e6cbcd143f5", size = 794175 }, + { url = "https://files.pythonhosted.org/packages/dd/3f/5989debef34dc6397317802b527dbbafb2b4760878a53d4166579111411e/pyyaml-6.0.3-cp314-cp314-musllinux_1_2_aarch64.whl", hash = "sha256:7c6610def4f163542a622a73fb39f534f8c101d690126992300bf3207eab9764", size = 755228 }, + { url = "https://files.pythonhosted.org/packages/d7/ce/af88a49043cd2e265be63d083fc75b27b6ed062f5f9fd6cdc223ad62f03e/pyyaml-6.0.3-cp314-cp314-musllinux_1_2_x86_64.whl", hash = "sha256:5190d403f121660ce8d1d2c1bb2ef1bd05b5f68533fc5c2ea899bd15f4399b35", size = 789194 }, + { url = "https://files.pythonhosted.org/packages/23/20/bb6982b26a40bb43951265ba29d4c246ef0ff59c9fdcdf0ed04e0687de4d/pyyaml-6.0.3-cp314-cp314-win_amd64.whl", hash = "sha256:4a2e8cebe2ff6ab7d1050ecd59c25d4c8bd7e6f400f5f82b96557ac0abafd0ac", size = 156429 }, + { url = "https://files.pythonhosted.org/packages/f4/f4/a4541072bb9422c8a883ab55255f918fa378ecf083f5b85e87fc2b4eda1b/pyyaml-6.0.3-cp314-cp314-win_arm64.whl", hash = "sha256:93dda82c9c22deb0a405ea4dc5f2d0cda384168e466364dec6255b293923b2f3", size = 143912 }, + { url = "https://files.pythonhosted.org/packages/7c/f9/07dd09ae774e4616edf6cda684ee78f97777bdd15847253637a6f052a62f/pyyaml-6.0.3-cp314-cp314t-macosx_10_13_x86_64.whl", hash = "sha256:02893d100e99e03eda1c8fd5c441d8c60103fd175728e23e431db1b589cf5ab3", size = 189108 }, + { url = "https://files.pythonhosted.org/packages/4e/78/8d08c9fb7ce09ad8c38ad533c1191cf27f7ae1effe5bb9400a46d9437fcf/pyyaml-6.0.3-cp314-cp314t-macosx_11_0_arm64.whl", hash = "sha256:c1ff362665ae507275af2853520967820d9124984e0f7466736aea23d8611fba", size = 183641 }, + { url = "https://files.pythonhosted.org/packages/7b/5b/3babb19104a46945cf816d047db2788bcaf8c94527a805610b0289a01c6b/pyyaml-6.0.3-cp314-cp314t-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:6adc77889b628398debc7b65c073bcb99c4a0237b248cacaf3fe8a557563ef6c", size = 831901 }, + { url = "https://files.pythonhosted.org/packages/8b/cc/dff0684d8dc44da4d22a13f35f073d558c268780ce3c6ba1b87055bb0b87/pyyaml-6.0.3-cp314-cp314t-manylinux2014_s390x.manylinux_2_17_s390x.manylinux_2_28_s390x.whl", hash = "sha256:a80cb027f6b349846a3bf6d73b5e95e782175e52f22108cfa17876aaeff93702", size = 861132 }, + { url = "https://files.pythonhosted.org/packages/b1/5e/f77dc6b9036943e285ba76b49e118d9ea929885becb0a29ba8a7c75e29fe/pyyaml-6.0.3-cp314-cp314t-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:00c4bdeba853cc34e7dd471f16b4114f4162dc03e6b7afcc2128711f0eca823c", size = 839261 }, + { url = "https://files.pythonhosted.org/packages/ce/88/a9db1376aa2a228197c58b37302f284b5617f56a5d959fd1763fb1675ce6/pyyaml-6.0.3-cp314-cp314t-musllinux_1_2_aarch64.whl", hash = "sha256:66e1674c3ef6f541c35191caae2d429b967b99e02040f5ba928632d9a7f0f065", size = 805272 }, + { url = "https://files.pythonhosted.org/packages/da/92/1446574745d74df0c92e6aa4a7b0b3130706a4142b2d1a5869f2eaa423c6/pyyaml-6.0.3-cp314-cp314t-musllinux_1_2_x86_64.whl", hash = "sha256:16249ee61e95f858e83976573de0f5b2893b3677ba71c9dd36b9cf8be9ac6d65", size = 829923 }, + { url = "https://files.pythonhosted.org/packages/f0/7a/1c7270340330e575b92f397352af856a8c06f230aa3e76f86b39d01b416a/pyyaml-6.0.3-cp314-cp314t-win_amd64.whl", hash = "sha256:4ad1906908f2f5ae4e5a8ddfce73c320c2a1429ec52eafd27138b7f1cbe341c9", size = 174062 }, + { url = "https://files.pythonhosted.org/packages/f1/12/de94a39c2ef588c7e6455cfbe7343d3b2dc9d6b6b2f40c4c6565744c873d/pyyaml-6.0.3-cp314-cp314t-win_arm64.whl", hash = "sha256:ebc55a14a21cb14062aa4162f906cd962b28e2e9ea38f9b4391244cd8de4ae0b", size = 149341 }, +] + [[package]] name = "rich" version = "14.3.3" @@ -458,7 +504,7 @@ wheels = [ [[package]] name = "team-harness" -version = "0.1.0" +version = "0.2.0" source = { editable = "." } dependencies = [ { name = "click" }, @@ -466,6 +512,7 @@ dependencies = [ { name = "openai" }, { name = "prompt-toolkit" }, { name = "pydantic" }, + { name = "pyyaml" }, { name = "rich" }, ] @@ -487,6 +534,7 @@ requires-dist = [ { name = "pyright", marker = "extra == 'dev'", specifier = ">=1.1" }, { name = "pytest", marker = "extra == 'dev'", specifier = ">=8.3" }, { name = "pytest-asyncio", marker = "extra == 'dev'", specifier = ">=0.25" }, + { name = "pyyaml", specifier = ">=6.0" }, { name = "rich", specifier = ">=14.0" }, { name = "ruff", marker = "extra == 'dev'", specifier = ">=0.4" }, ]