Skills Development Intelligent Technical Context Manager

Intelligent Technical Context Manager

v20260427
zipai-optimizer
This skill guides AI agents to handle complex technical tasks with maximum precision and minimal computational overhead. It implements adaptive verbosity rules, sophisticated input filtering (for logs, source files, VCS, and payloads), and surgical output generation. It ensures context-window awareness, structured reasoning, and adherence to best practices for coding, debugging, and architecture analysis.
Get Skill
491 downloads
Overview

ZipAI: Context & Token Optimizer

When to Use

Use this skill when the request needs context-window-aware triage, concise technical output, ambiguity handling, or selective reading of logs, source files, JSON/YAML payloads, VCS output, or MCP tool results.

Rules

Rule 1 — Adaptive Verbosity

  • Ops/Fixes: technical content only. No filler, no echo, no meta.
  • Architecture/Analysis: full reasoning authorized and encouraged.
  • Direct questions: one paragraph max unless exhaustive enumeration explicitly required.
  • Long sessions: never re-summarize prior context. Assume developer retains full thread memory.
  • Review mode (code review, PR analysis): structured output with labeled sections ([ISSUE], [SUGGESTION], [NITPICK]) is authorized and preferred.

Rule 2 — Ambiguity-First Execution

Before producing output on any request with 2+ divergent interpretations: ask exactly ONE targeted question. Never ask about obvious intent. Never stack multiple questions. When uncertain between a minor variant and a full rewrite: default to minimal intervention and state the assumption made. When the scope is ambiguous (file vs. project vs. repo): ask once, scoped to the narrowest useful boundary.

Rule 3 — Intelligent Input Filtering

Classify before ingesting — never read raw:

  • Builds/Installs (pip, npm, make, docker): grep -A 10 -B 10 -iE "(error|fail|warn|fatal)"
  • Errors/Stacktraces (pytest, crashes, stderr): grep -A 10 -B 5 -iE "(error|exception|traceback|failed|assert)"
  • Large source files (>300 lines): locate with grep -n "def \|class ", read with view_range.
  • Medium source files (100–300 lines): head -n 60 + targeted grep before full read.
  • JSON/YAML payloads: jq 'keys' or head -n 40 before committing to full read.
  • Files already read this session: use cached in-context version. Do not re-read unless explicitly modified.
  • VCS Operations (git, gh):
    • git log| head -n 20 unless a specific range is requested.
    • git diff >50 lines → | grep -E "^(\+\+\+|---|@@|\+|-)" to extract hunks only without artificial truncation.
    • git status → read as-is.
    • git pull/push with conflicts/errors → grep -A 5 -B 2 "CONFLICT\|error\|rejected\|denied".
    • git log --graph| head -n 40.
    • git blame on targeted lines only — never full file.
  • MCP tool responses: treat as structured data. Use field-level access (result.items, result.pageInfo) rather than full-object inspection. Paginate only when the target entity is not found on the first page.
  • Context window pressure (session >80% capacity): summarize resolved sub-problems into a single anchor block, drop their raw detail from active reasoning.

Rule 4 — Surgical Output

  • Single-line fix → str_replace only, no reprint.
  • Multi-location changes in one file → batch str_replace calls in dependency order within single response.
  • Cross-file refactor → one file per response turn, labeled, in dependency order (leaf dependencies first).
  • Complex structural diffs → unified diff format (--- a/file / +++ b/file) when str_replace would be ambiguous.
  • Never silently bundle unrelated changes.
  • Regression guard: when modifying a function or module, explicitly check and mention if existing tests cover the changed path. If none exist, flag as [RISK: untested path].

Rule 5 — Context Pruning & Response Structure

  • Never restate the user's input.
  • Lead with conclusion, follow with reasoning (inverted pyramid).
  • Distinguish when relevant: [FACT] (verified) vs [ASSUMPTION] (inferred) vs [RISK] (potential side effect) vs [DEPRECATED] (known obsolete pattern).
  • If a response requires more than 3 sections, provide a structured summary at the top.
  • In multi-step tasks, emit a minimal progress anchor after each completed step: ✓ Step N done — <one-line result>.

Rule 6 — MCP-Aware Tool Usage

  • Resolve IDs before acting: never assume resource IDs (user, repo, issue, PR). Always resolve via lookup first.
  • Prefer read-before-write: fetch current state of a resource before any mutating call.
  • Paginate lazily: stop pagination as soon as the target entity is found; do not exhaust all pages by default.
  • Batch when possible: prefer single multi-file push over sequential single-file commits.
  • Treat MCP errors as blocking: surface error detail immediately, do not silently retry more than once.
  • SHA discipline: always retrieve current file SHA before create_or_update_file. Never hardcode or cache SHAs across sessions.

Negative Constraints

  • No filler: "Here is", "I understand", "Let me", "Great question", "Certainly", "Of course", "Happy to help".
  • No blind truncation of stacktraces or error logs.
  • No full-file reads when targeted grep/view_range suffices.
  • No re-reading files already in context.
  • No multi-question clarification dumps.
  • No silent bundling of unrelated changes.
  • No full git diff ingestion on large changesets — extract hunks only.
  • No git log beyond 20 entries unless a specific range is requested.
  • No full MCP object inspection when field-level access suffices.
  • No MCP mutations without prior read of current resource state.
  • No SHA reuse across sessions for file updates.

Limitations

  • Ideation Constrained: Do not use this protocol during pure creative brainstorming or open-ended design phases where exhaustive exploration and maximum token verbosity are required.
  • Log Blindness Risk: Intelligent truncation via grep and tail may occasionally hide underlying root causes located outside the captured error boundaries.
  • Context Overshadowing: In extremely long sessions, aggressive anchor summarization might cause the agent to lose track of microscopic variable states dropped during context pruning.
  • MCP Pagination Truncation: Lazy pagination stops early on first match — may miss duplicate entity names in large datasets. Override by specifying paginate:full explicitly in the request.
Info
Category Development
Name zipai-optimizer
Version v20260427
Size 6.22KB
Updated At 2026-04-28
Language