Files
everything-claude-code/skills/hermes-generated/research-ops/SKILL.md
2026-03-25 02:41:08 -07:00

3.2 KiB

name, description, metadata
name description metadata
research-ops Evidence-first research workflow for Hermes. Use when answering current questions, evaluating a market or tool, enriching leads, or deciding whether a request should become ongoing monitored data collection.
hermes
tags
generated
research
market
discovery
monitoring
workflow
verification

Research Ops

Use this when the user asks Hermes to research something current, compare options, enrich people or companies, or turn repeated lookups into an ongoing monitoring workflow.

Skill Stack

Pull these imported skills into the workflow when relevant:

  • deep-research for multi-source cited synthesis
  • market-research for decision-oriented framing
  • exa-search for first-pass discovery and current-web retrieval
  • data-scraper-agent when the user really needs recurring collection or monitoring
  • search-first before building new scraping or enrichment logic
  • eval-harness mindset for claim quality, freshness, and explicit uncertainty

When To Use

  • user says research, look up, find, who should i talk to, what's the latest, or similar
  • the answer depends on current public information, external sources, or a ranked set of candidates
  • the task sounds recurring enough that a scraper or scheduled monitor may be better than a one-off search

Workflow

  1. Classify the ask before searching:
    • quick factual answer
    • decision memo or comparison
    • lead list or enrichment
    • recurring monitoring request
  2. Start with the fastest evidence path:
    • use exa-search first for broad current-web discovery
    • if the question is about a local wrapper, config, or checked-in code path, inspect the live local source before making any web claim
  3. Deepen only where the evidence justifies it:
    • use deep-research when the user needs synthesis, citations, or multiple angles
    • use market-research when the result should end in a recommendation, ranking, or go/no-go call
  4. Separate fact from inference:
    • label sourced facts clearly
    • label inferred fit, ranking, or recommendation as inference
    • include dates when freshness matters
  5. Decide whether this should stay manual:
    • if the user will likely ask for the same scan repeatedly, use data-scraper-agent patterns or propose a monitored collection path instead of repeating the same manual research forever
  6. Report with evidence:
    • cite the source or local file behind each important claim
    • if evidence is thin or conflicting, say so directly

Pitfalls

  • do not answer current questions from stale memory when a fresh search is cheap
  • do not conflate local code-backed behavior with market or web evidence
  • do not present unsourced numbers or rankings as facts
  • do not spin up a heavy deep-research pass for a quick capability check that local code can answer
  • do not keep one-off researching a repeated monitoring ask when automation is the better fit

Verification

  • important claims have a source, file path, or explicit inference label
  • freshness-sensitive answers include concrete dates when relevant
  • recurring-monitoring recommendations state whether the task should remain manual or graduate to a scraper/workflow