mirror of
https://github.com/affaan-m/everything-claude-code.git
synced 2026-04-17 23:53:30 +08:00
feat: add hermes-generated ops skills
This commit is contained in:
64
skills/hermes-generated/research-ops/SKILL.md
Normal file
64
skills/hermes-generated/research-ops/SKILL.md
Normal file
@@ -0,0 +1,64 @@
|
||||
---
|
||||
name: research-ops
|
||||
description: Evidence-first research workflow for Hermes. Use when answering current questions, evaluating a market or tool, enriching leads, or deciding whether a request should become ongoing monitored data collection.
|
||||
metadata:
|
||||
hermes:
|
||||
tags: [generated, research, market, discovery, monitoring, workflow, verification]
|
||||
---
|
||||
|
||||
# Research Ops
|
||||
|
||||
Use this when the user asks Hermes to research something current, compare options, enrich people or companies, or turn repeated lookups into an ongoing monitoring workflow.
|
||||
|
||||
## Skill Stack
|
||||
|
||||
Pull these imported skills into the workflow when relevant:
|
||||
- `deep-research` for multi-source cited synthesis
|
||||
- `market-research` for decision-oriented framing
|
||||
- `exa-search` for first-pass discovery and current-web retrieval
|
||||
- `data-scraper-agent` when the user really needs recurring collection or monitoring
|
||||
- `search-first` before building new scraping or enrichment logic
|
||||
- `eval-harness` mindset for claim quality, freshness, and explicit uncertainty
|
||||
|
||||
## When To Use
|
||||
|
||||
- user says `research`, `look up`, `find`, `who should i talk to`, `what's the latest`, or similar
|
||||
- the answer depends on current public information, external sources, or a ranked set of candidates
|
||||
- the task sounds recurring enough that a scraper or scheduled monitor may be better than a one-off search
|
||||
|
||||
## Workflow
|
||||
|
||||
1. Classify the ask before searching:
|
||||
- quick factual answer
|
||||
- decision memo or comparison
|
||||
- lead list or enrichment
|
||||
- recurring monitoring request
|
||||
2. Start with the fastest evidence path:
|
||||
- use `exa-search` first for broad current-web discovery
|
||||
- if the question is about a local wrapper, config, or checked-in code path, inspect the live local source before making any web claim
|
||||
3. Deepen only where the evidence justifies it:
|
||||
- use `deep-research` when the user needs synthesis, citations, or multiple angles
|
||||
- use `market-research` when the result should end in a recommendation, ranking, or go/no-go call
|
||||
4. Separate fact from inference:
|
||||
- label sourced facts clearly
|
||||
- label inferred fit, ranking, or recommendation as inference
|
||||
- include dates when freshness matters
|
||||
5. Decide whether this should stay manual:
|
||||
- if the user will likely ask for the same scan repeatedly, use `data-scraper-agent` patterns or propose a monitored collection path instead of repeating the same manual research forever
|
||||
6. Report with evidence:
|
||||
- cite the source or local file behind each important claim
|
||||
- if evidence is thin or conflicting, say so directly
|
||||
|
||||
## Pitfalls
|
||||
|
||||
- do not answer current questions from stale memory when a fresh search is cheap
|
||||
- do not conflate local code-backed behavior with market or web evidence
|
||||
- do not present unsourced numbers or rankings as facts
|
||||
- do not spin up a heavy deep-research pass for a quick capability check that local code can answer
|
||||
- do not keep one-off researching a repeated monitoring ask when automation is the better fit
|
||||
|
||||
## Verification
|
||||
|
||||
- important claims have a source, file path, or explicit inference label
|
||||
- freshness-sensitive answers include concrete dates when relevant
|
||||
- recurring-monitoring recommendations state whether the task should remain manual or graduate to a scraper/workflow
|
||||
Reference in New Issue
Block a user