Skip to content

feat(linear): add Linear integration, with Slack review fixes#63

Merged
krokoko merged 25 commits intoaws-samples:mainfrom
isadeks:feat/linear-integration
May 7, 2026
Merged

feat(linear): add Linear integration, with Slack review fixes#63
krokoko merged 25 commits intoaws-samples:mainfrom
isadeks:feat/linear-integration

Conversation

@isadeks
Copy link
Copy Markdown
Contributor

@isadeks isadeks commented May 6, 2026

Summary

  • Linear integration (v1) — label-triggered task submission from Linear issues. Webhook → HMAC-verified receiver → async processor → standard task pipeline, with channel_source='linear'. Agent updates the originating Linear issue inline (comments, state transition, reactions) via Linear's hosted MCP server + a direct GraphQL client for reactions. Live demo: https://isadeks.github.io/abca-linear-demo/ (7 agent-authored PRs merged, site deployed per merge).
  • Slack review fixes (addresses PR feat(slack): add Slack integration with @mention-based task submission #42 comments) — review-feedback commits for items 1-20, handler tests for all 7 Slack Lambdas, doc sync, softened checkChannelAccess. Originally on a separate local branch; folded into this PR per your preference for a single combined review surface.
  • Upstream reconciliation — merged aws-samples/main to absorb Sam's PR feat(interactive-agents): async-only background task UX + Cedar HITL design #52 (FanOutConsumer) and adjacent changes. Resolved 13 conflicts across schema additions, construct wiring, agent pipeline ordering, and a flaky-test fix. Full test suite green post-merge (CDK 1121, agent 526, CLI 76).

Response to your #42 review concerns

Being honest about what Linear addresses vs. sidesteps:

Concern Linear integration's relationship to it
DDB Streams max 2 consumers per table — 3rd adapter hits AWS limit Didn't solve, didn't worsen. Linear adds zero stream consumers (uses agent-side MCP for outbound). Post-#52 state: SlackNotifyFn + FanOutConsumer = exactly 2/2 at the AWS limit. Structural fix still needs to happen, not in this PR — migrating SlackNotifyFn (from #42) into FanOutConsumer as a dispatcher subscriber would drop the count back to 1 and restore headroom. We deliberately held the line and didn't add a third.
Channel-filter logic duplicated inside each consumer Didn't solve, didn't worsen. No Linear consumer exists to duplicate filter logic in. channelSource === 'linear' appears only in linear-webhook-processor.ts (one call site, one docstring). If a fourth channel ships as a stream consumer in the future, the duplication concern returns — but the #52 dispatcher pattern is the intended long-term shape for that; future channels should subscribe to it rather than add another filter-in-each-consumer.
Dedup logic duplicated across adapters Addressed for Linear. Webhook-level dedup lives in linear-webhook.ts only, with key ${issueId}#${action}#${webhookTimestamp} (8h TTL, exceeds Linear's 7h retry horizon). No TaskRecord-level dedup extensions.
channel_metadata accumulates lifecycle state (e.g. slack_session_msg_ts) Genuinely sidestepped. Zero linear_*_msg_ts fields on TaskRecord. All per-issue state (comments, state id, reaction id) lives in Linear itself, accessed at runtime via MCP + GraphQL. The container is the only place that touches Linear state; nothing is stashed on the record.

Architectural notes

Why MCP-based outbound instead of a FanOutConsumer subscriber. Linear ships a first-class MCP server (https://mcp.linear.app/mcp), so the agent can update the issue from inside the container using tools — same place the work is happening, no extra Lambda hop, no duplicated state. The fanout pattern is a better fit for ephemeral notifications (Slack thread updates, email) than stateful back-and-forth with a ticket (comment threads, state transitions, reactions that need the originating issue id).

Operationally: linear-integration.ts has zero stream consumers and zero outbound Lambdas. The only Lambdas are the webhook receiver, the async processor, and linear-link (CLI account linking). Everything user-visible happens in the agent container.

WAF scope-down on /v1/linear/webhook. Linear Issue webhook payloads routinely exceed the 8 KB body-size rule in AWSManagedRulesCommonRuleSet (observed 8614 bytes in smoke). Scoped the CRS out for this one path via scopeDownStatement. The route retains HMAC verification, webhookTimestamp replay guard (60s), JSON schema validation, webhook dedup, and the priority-3 rate-limit rule. AWSManagedRulesKnownBadInputsRuleSet remains unchanged.

Auto-link on bgagent linear setup. Linear identifies the API-key owner via viewer { id }, so we look up the caller's Cognito sub from the cached id_token and write the LinearUserMapping row directly. Fail-open on auth / network errors (secrets still get stored; user is told to re-run bgagent login). This closes the "now go paste bgagent link in a comment" dance for the common single-admin case.

v1 multi-user scope. Only the API-token owner can submit tasks from Linear in v1. Tasks triggered by other Linear users are dropped by the processor with "Linear actor has no linked platform user — skipping task creation". bgagent linear link <code> is registered as a CLI command but the Linear-side code generator isn't in v1 — OAuth bot install is the proper multi-user UX and will subsume any link-code bridge, so shipping the bridge didn't seem worth it. Constraint is documented in the setup guide.

In scope / out of scope

In this PR (v1 / shipped):

  • Linear webhook + processor + link handler + two mapping tables (LinearProjectMappingTable, LinearUserMappingTable) + dedup table
  • LinearIntegration CDK construct (zero stream consumers, zero outbound Lambdas)
  • Agent-side Linear MCP loader (configure_channel_mcp, gated on channel_source='linear')
  • Agent-side 👀/✅/❌ reactions via Linear GraphQL (retire when Linear MCP ships create_reaction)
  • CLI: bgagent linear setup (+ auto-link), bgagent linear link, bgagent linear list-projects, bgagent linear onboard-project (+ UUID validation)
  • Prompt addendum for the Linear MCP tools (save_comment, save_issue with graceful "no matching state" fallback)
  • WAF scope-down + 8h dedup TTL + flaky test_server race fix
  • Slack review fixes from your PR feat(slack): add Slack integration with @mention-based task submission #42 comments + handler tests for all 7 Slack Lambdas

Out of scope (planned follow-ups):

Band What Why parked
v1.1 Processor posts a Linear comment + ❌ when task creation fails pre-container (guardrail block, user concurrency cap, unmapped project) — today these drop silently on the Linear side. Observed 3 distinct silent-failure cases during smoke; all share a single ~30 LOC fix. Worth shipping as a polish pass.
v1.1 Attachment pre-fetch to S3 for Linear tickets with images / mockups. Linear MCP has attachment size limits; PMs attach screenshots routinely.
v2 Comment-triggered retry UX (/bgagent retry in a Linear comment) for partial-PR recovery. Needs Comment webhook subscription + slash-command parser in processor.
v2 Auto-close Linear issue on PR merge. Open design question first — merging ≠ solution accepted; post-merge tests may reveal issues. May be better done as a user-side GitHub Action than ABCA-side infrastructure.
v3 Cancel-from-Linear terminal feedback. bgagent cancel SIGKILLs the container, so Linear-side ❌ never fires from the agent's except. Needs an orchestrator-side Linear client.
v3 OAuth bot install + multi-workspace support — replaces personal API key, enables any Linear user in the workspace to trigger tasks. Current v1 uses a single personal API key (single submitter). The right fix when multi-user UX becomes a priority.

Validation

  • Live site: https://isadeks.github.io/abca-linear-demo/ — 7 agent-authored PRs merged and deployed during validation.
  • Test counts: CDK 1121 passing (was 853 pre-merge), agent 526 passing (was 306), CLI 76 passing. All green post-merge.
  • Smoke tests run on the deployed stack:
    • WAF scope-down returns 401 (Lambda HMAC reject), not 403 (WAF block)
    • Dedup TTL ≥ 7h verified, distinct webhookTimestamp keys → no delivery collapse
    • UUID validation on onboard-project rejects the Linear-URL-slug form
    • Create-with-label-at-creation happy path (ABCA-3 → ABCA-22, 15+ runs)
    • Concurrent tasks don't cross-pollinate reactions
    • Unmapped-project silent drop (documents the v1.1 gap)
    • bgagent cancel mid-run (documents the v3 gap)
    • Post-merge Linear + Slack pipelines both working
  • Architectural assertions on the deployed stack:
    • aws dynamodb describe-table --query 'Table.StreamSpecification' on TaskEventsTable — 2 consumers (SlackNotifyFn + FanOutConsumer); Linear adds 0.
    • grep -rn "channelSource === 'linear'" cdk/src/handlers/ — only in linear-webhook-processor.ts.
    • TaskRecord has no linear_*_msg_ts fields.

Test plan for reviewer

  1. mise //cdk:deploy — CFN diff should add only LinearIntegration/* resources on top of current aws-samples/main. Slack beyond Lambda bundle refreshes should not change.
  2. Generate a Linear personal API key at https://linear.app/settings/account/security.
  3. Run bgagent linear setup. It prints the webhook URL, pauses at the signing-secret prompt. Create a webhook in Linear at that URL (subscribe to Issue), copy the signing secret, return to the terminal and paste secret + API key. Auto-link fires for the token owner; success message: ✓ Linked Linear user … → platform user ….
  4. bgagent linear list-projects → copy the target project UUID.
  5. bgagent linear onboard-project <uuid> --repo <your-fork>/<demo-repo> --label abca.
  6. Create a Linear issue with the abca label applied at creation time.
  7. On the Linear issue, within ~30s: 👀 reaction, 🤖 Starting… comment, PR opens, state → In Review, 👀 swaps for ✅ when the PR URL comment is posted and the task finishes.
  8. Merge the PR — site / deploy workflow runs normally.

Closes #42.

isadeks and others added 24 commits April 20, 2026 13:25
Adds full Slack integration enabling users to submit coding tasks by
mentioning @shoof in any channel or DM, with real-time emoji reactions
and threaded notifications showing task progress.

Key features:
- @mention task submission with natural language repo extraction
- Emoji reaction progression: 👀 → ⌛ → ✅
- Threaded notifications (created, started, completed/failed)
- Cancel button with instant feedback
- DM support for private task submissions
- OAuth multi-workspace install flow
- `bgagent slack setup` CLI wizard for zero-friction onboarding
- Account linking via /bgagent link

New files:
- CDK constructs: SlackIntegration, SlackInstallationTable, SlackUserMappingTable
- Lambda handlers: events, commands, command-processor, interactions,
  oauth-callback, link, notify
- Shared utilities: slack-verify, slack-blocks
- CLI: slack.ts (setup, link commands)
- Docs: SLACK_SETUP_GUIDE.md with screenshots
# Conflicts:
#	docs/guides/DEVELOPER_GUIDE.md
#	docs/guides/USER_GUIDE.md
#	docs/src/content/docs/developer-guide/Installation.md
#	docs/src/content/docs/developer-guide/Repository-preparation.md
#	docs/src/content/docs/user-guide/Overview.md
Add SLACK_SETUP_GUIDE.md to the explicit route map in sync-starlight.mjs,
mirror it to using/Slack-setup-guide.md, and wire it into the astro sidebar.
Without this, links to the Slack setup guide rendered as /architecture/... and
the page itself never appeared on the site.

Co-Authored-By: Claude Opus 4.7 (1M context) <noreply@anthropic.com>
…1-20)

Full response to the 20-point review on PR aws-samples#42. Grouped by theme below
rather than strictly by item number so related edits stay together.

**Signature + auth (items 2, 3, 9)**

- Item 2: url_verification challenge used to be answered before signature
  verification, exposing the endpoint as a Slack handler without auth. Now
  every request is verified first. During initial setup — before the
  signing secret is populated — url_verification is still answered so the
  Slack App can be wired up, but every other request is rejected.
- Item 3: previously any request with X-Slack-Retry-Num was acked with 200
  immediately. Transient failures on app_uninstalled or tokens_revoked
  silently dropped the revocation and left the bot token live forever.
  Security-critical event retries now reprocess; UX events still short-
  circuit on retry.
- Item 9: slack-verify's 5-minute secret cache had no invalidation. After
  a signing-secret rotation, warm Lambdas rejected every request for up
  to 5 minutes. New verifySlackRequest helper does verify-then-retry with
  forced refresh; invalidateSlackSecretCache exported for explicit eviction.
  All three authed handlers (events, commands, interactions) go through
  verifySlackRequest.

**Error surfacing (items 4, 7, 20)**

- Item 4: slack-notify caught every error per-record and logged warn, which
  made retryAttempts and bisectBatchOnError useless. Now distinguishes
  SlackApiError (swallow — not retryable via stream replay) from infra
  errors (rethrow so Lambda retries the batch).
- Item 7: six .catch(() => {}) swallows in slack-events and
  slack-command-processor replaced with a shared slackFetch helper in
  shared/slack-api.ts that logs at warn. Benign Slack errors
  (already_reacted, no_reaction) are still silenced as expected.
- Item 20: safeJsonParse in slack-notify logs the parse error and the
  truncated payload preview instead of swallowing silently.

**Revocation ordering (item 6)**

- revokeInstallation previously deleted the bot-token secret even if the
  DDB update to mark the installation revoked had failed, leaving DB
  status "active" with no backing token. Now the secret delete runs only
  if the DDB update succeeded; both steps throw on failure so Slack
  retries the event end-to-end.

**UX fallbacks (items 8, 10)**

- Item 8: when the command-processor Lambda invocation fails from
  handleAppMention, swap 👀 → ❌ and reply in-thread, mirroring the
  no-repo-found path. Previously the user was left with a stuck 👀.
- Item 10: checkChannelAccess fails closed on definitively-blocking Slack
  errors (channel_not_found, not_in_channel, missing_scope). Transient
  or unknown errors (ratelimited, internal_error, network blips) fail
  open so a Slack outage doesn't block task submission; slack-notify
  surfaces any real delivery failure downstream.

**Dedup ordering (item 17)**

- Moved the channel_source === 'slack' check before the terminal dedup
  write so we don't poke non-Slack task records' channel_metadata with a
  Slack-specific marker.

**Dead code + type safety (items 1, 13, 14)**

- Item 1: delete unused handleStatus / handleCancel / statusEmoji from
  slack-command-processor. They were fully implemented but never wired
  into the switch(subcommand). Sam's PR aws-samples#52 roadmap delivers `bgagent
  status` via the CLI; the Slack Cancel button already exists.
- Item 13: MentionPayload used to extend SlackCommandPayload and set
  phantom fields (`response_url: ''`, etc.). Refactored into a discriminated
  union `CommandProcessorEvent = SlashCommandEvent | MentionEvent`.
  slack-events.ts constructs MentionEvent directly; handleSubmit takes
  MentionEvent only.
- Item 14: SlackBlock was a single interface requiring
  `as unknown as SlackBlock` to construct actions blocks. Now
  SectionBlock | ActionsBlock discriminated union with typed button
  variants. Casts removed; tests updated with narrowing helpers.

**Shared utilities (items 15, 16, 19)**

- Items 15/16: extract ChannelSource union from shared/types.ts and use
  it on both TaskCreationContext.channelSource and TaskRecord.channel_source.
  Extract truncate() and formatDuration() into shared/slack-format.ts so
  they stop drifting between slack-blocks and slack-command-processor.
- Item 19: formatDuration now coerces via Number() for all callers, fixing
  the inconsistency where taskTimedOutMessage skipped the cast.

**CLI error surfacing (item 11)**

- cli/src/commands/slack.ts:getStackOutput used to swallow every
  CloudFormation error as "not deployed yet". Now only swallows
  ValidationError / does-not-exist; auth and permission errors bubble up
  to the user.

**Excess IAM grant (item 18)**

- slackCommandsFn had readSlackSecretsPolicy attached, but the slash-
  command acknowledger only needs the signing secret (already granted
  separately). Grant removed; comment added explaining the intent.

**OAuth CSRF (item 5)**

- Documented the tradeoff in slack-oauth-callback.ts. Slack's client_secret
  requirement on the code exchange is the primary defense; a full state
  nonce requires per-session signed storage and is tracked as a follow-up
  coordinated with PR aws-samples#52's dispatcher rework.

**Not addressed in this PR (architectural feedback for follow-up)**

Alain's "What I would like to see" items 1-5 — dispatcher pattern,
notification-state extraction from channel_metadata, ADR for the adapter
contract — are scoped to the PR that will rebase onto Sam's PR aws-samples#52
(feat(interactive-agents)) once it lands. Sam's locked plan establishes
the FanOutConsumer router that those items depend on.
Addresses review item 12 / general feedback 4: the shared utilities
(slack-verify, slack-blocks) and CDK constructs had solid coverage but
all 7 Lambda handlers had zero unit tests, a departure from the rest of
the codebase.

Added test files covering the most load-bearing paths:

- slack-events.test.ts (11 tests) — url_verification challenge with/
  without signature (item 2), retry semantics (drop non-critical,
  reprocess app_uninstalled + tokens_revoked, item 3), revoke-installation
  ordering (item 6: no secret delete on DDB failure), app_mention
  forwarding, no-repo + Lambda-failure reaction swaps (item 8)
- slack-commands.test.ts (4 tests) — signature rejection, inline help,
  async processor invocation, graceful processor failure
- slack-interactions.test.ts (6 tests) — signature rejection, cancel_task
  happy path + owner rejection + terminal-state warn, unknown action_id,
  unlinked account
- slack-notify.test.ts (8 tests) — non-Slack task short-circuit, dedup
  ordering after source check (item 17), dedup write skip on duplicate,
  SlackApiError swallow vs infra-error rethrow (item 4), non-INSERT /
  non-notifiable filter, malformed metadata logging (item 20)
- slack-link.test.ts (6 tests) — 401 without JWT, 400 without code,
  404 on missing/stale code, happy path write+delete, code normalization
- slack-oauth-callback.test.ts (5 tests) — missing code, unpopulated
  client secret, token exchange failure, create-secret-on-RNF path,
  restore-then-update path for reinstall after uninstall
- slack-command-processor.test.ts (9 tests) — legacy source-less payload
  normalization, slash-submit-rejection UX, mention link-prompt + repo
  validation + private channel guard (item 10 hard failure) + transient
  fail-open (item 10 softening), createTaskCore invocation, link code
  persistence, help text

The regression tests for items 2, 3, 4, 6, 9 all fail the buggy pre-fix
version of the code, verified by running them against the unchanged
source during review.
Addresses blind-review feedback S2 + S3.

**S2 — checkChannelAccess transient failures now fail open.**

The prior "fail closed on any unknown Slack error" was too strict: a
30-second Slack ratelimited or internal_error during conversations.info
would block every Slack task submission for that window, even though
the notification path downstream could have recovered. Now:

- Hard failures (channel_not_found, not_in_channel, missing_scope) still
  fail closed — these definitively mean the bot cannot post there, and
  silently creating an undeliverable task is the worse UX.
- Transient or unknown errors (ratelimited, internal_error, fatal_error,
  network blips) fail open. slack-notify surfaces any real delivery
  failure when the event plays through.

Regression test added: `mention submit fails open on transient Slack
errors` asserts createTaskCore IS called when conversations.info returns
`ratelimited`.

**S3 — USER_GUIDE Slack bullet drops "check status".**

handleStatus was deleted in this branch (review item 1) because Sam's
locked plan in PR aws-samples#52 delivers `bgagent status` via the CLI rather than
a Slack slash subcommand. The user-facing docs still claimed Slack could
"check status" via `/bgagent`. Rewritten to describe the shipped
behavior: @mention-based submission plus threaded progress notifications
with reaction-based status. Starlight content regenerated via
`node docs/scripts/sync-starlight.mjs`.
Trivial enabler for the Linear integration branch. No functional change:
all existing code already uses the broad ChannelSource union; adding
'linear' makes the Linear-origin task records type-check once the
inbound adapter starts writing them.
- shared/linear-verify.ts: HMAC-SHA256 hex verify over raw body,
  60s webhookTimestamp replay window, secret caching + rotation retry.
- LinearProjectMappingTable: linear_project_id → repo + label_filter.
- LinearUserMappingTable: {workspace}#{user} confirmed / pending#{code}
  with TTL and PlatformUserIndex GSI.
- linear-webhook.ts: verify + dedup (60s TTL on (issue_id, action))
  + async-invoke processor; replies fast inside Linear's 5s budget.
- linear-webhook-processor.ts: Issue label-add detection, project→repo
  + actor→platform user resolution, createTaskCore with
  channelSource 'linear' + linear_* channel metadata.
- linear-link.ts: mirror of slack-link.ts.
- LinearIntegration construct: provisions both mapping tables, the webhook
  dedup table, 3 Lambdas (webhook / processor / link), 2 Secrets Manager
  placeholders, and API Gateway routes under /v1/linear/*.
  Deliberately no DynamoDB Streams consumer — outbound updates go through
  the Linear MCP from inside the agent container.
- Stack: instantiates LinearIntegration in agent.ts and emits CfnOutputs
  for the two secret ARNs and both table names, consumed by
  `bgagent linear setup` / onboard-project.
- Orchestrator invoke payload now carries `channel_source` +
  `channel_metadata`, so the agent knows which inbound channel a task came
  from (and the Linear issue identifier for Linear-origin tasks).
- TaskConfig + build_config accept the new fields; server.py extracts them
  from the /invocations payload and forwards into pipeline.run_task.
- channel_mcp.py: new `configure_channel_mcp(repo_dir, channel_source)`.
  For channel_source=='linear' it merges a `linear-server` entry into
  `.mcp.json` (keyed so tools surface as `mcp__linear-server__*`, matching
  the verified MCP spec). For any other channel it's a no-op — the gate
  that keeps Slack/API tasks from loading Linear tools.
- config.py: `resolve_linear_api_token()` reads LinearApiTokenSecret via
  LINEAR_API_TOKEN_SECRET_ARN and caches into LINEAR_API_TOKEN so the SDK
  child process inherits it for the MCP's ${LINEAR_API_TOKEN} placeholder.
- pipeline.py: resolves the token + writes .mcp.json just before run_agent.
- prompt_builder.py: appends a short Linear-specific addendum naming
  `mcp__linear-server__save_comment` / `update_issue` by tool name.
- Stack: grants runtime read on LinearApiTokenSecret and sets the secret
  ARN in the AgentCore runtime environment via CFN override.
- run.sh: forwards LINEAR_API_TOKEN / LINEAR_API_TOKEN_SECRET_ARN into the
  local container for manual testing.
- cli/src/commands/linear.ts: new `bgagent linear` command group:
  - `setup` — prompts for the Linear webhook signing secret and personal
    API key, populates both placeholders in Secrets Manager, prints the
    next-step checklist including the webhook URL.
  - `link <code>` — calls POST /v1/linear/link with the Cognito JWT.
  - `onboard-project <linear-project-id> --repo owner/repo [--label bgagent]`
    — writes the Linear project → repo mapping row directly to
    LinearProjectMappingTable via DynamoDBDocumentClient (admin IAM path,
    matching the current onboarding UX for Slack + webhooks).
- api-client + types: LinearLinkResponse type + linearLink() method.
- bgagent.ts: registers the new command group.
- cli/package.json: adds @aws-sdk/client-dynamodb + @aws-sdk/lib-dynamodb
  so the CLI can write project mappings.
- docs/guides/LINEAR_SETUP_GUIDE.md: new end-to-end setup walkthrough
  mirroring SLACK_SETUP_GUIDE, covering API-key generation, webhook
  configuration, onboarding, linking, and troubleshooting.
- docs/scripts/sync-starlight.mjs + docs/astro.config.mjs: routing + sidebar
  entry for the new guide.
- docs/guides/USER_GUIDE.md: Linear now shows up as the 5th channel.
- cdk/test/stacks/agent.test.ts: DDB table count bumped 7 → 10 to cover
  the three new Linear tables.
New test coverage for the Linear integration:

- cdk/test/handlers/linear-webhook.test.ts — HMAC pass/fail, stale
  webhookTimestamp, non-Issue type is 200/no-op, missing data.id is 400,
  dedup-hit short-circuits processor invoke, Lambda invoke failure surfaces
  as 500 so Linear retries, malformed JSON with valid signature is 400.
- cdk/test/handlers/linear-webhook-processor.test.ts — project-not-onboarded
  and status=removed skip, label filter gates on create AND on update
  (updatedFrom.labelIds transition), unlinked Linear actor skips, happy
  path asserts channel_source='linear' + full linear_* channelMetadata,
  custom label_filter honored per project mapping.
- cdk/test/handlers/linear-link.test.ts — mirror of slack-link.test.ts:
  401/400/404 error paths + confirmed-mapping write + code normalization.
- cdk/test/constructs/linear-integration.test.ts — resource-count
  assertions (3 DDB tables, 3 Lambdas, 2 secrets, ZERO DDB Streams
  mappings), env-var wiring for both handlers, dedup-table TTL shape.
- agent/tests/test_channel_mcp.py — channel gate (slack/api/webhook/empty
  are no-ops; only linear writes), .mcp.json shape (server key
  `linear-server`, hosted URL, ${LINEAR_API_TOKEN} placeholder), merge
  semantics (preserves other servers, overwrites stale linear entry,
  tolerates malformed JSON), missing-repo_dir guard.

All 853 CDK tests and 294 agent tests pass locally.
Exempt /v1/linear/webhook from AWSManagedRulesCommonRuleSet via a
scopeDownStatement. Linear Issue webhook payloads routinely exceed the
ruleset's default 8 KB body-size limit (observed 8614 bytes), and the
full payload is needed by the webhook Lambda for HMAC verification.

The route has stronger, bespoke protection already:
- HMAC verification against LinearWebhookSecret on the raw body
- webhookTimestamp replay guard (60s window)
- Structured JSON parsing with strict field validation
- RateLimitRule (1000/IP/5min) is priority 3 and still applies

CRS continues to block on every other route (user API, Slack, etc.).
The prior dedup key was issueId#action, which collapsed distinct
label-toggle deliveries on the same issue into a single dispatch — a
label remove/reapply within 60s would be silently dropped.

Also, the prior TTL (60s) was shorter than Linear's retry horizon
(+1m, +1h, +6h). A dispatched-but-ack-lost delivery would re-dispatch
on the +1h or +6h retry after the dedup row had expired, double-
creating the task.

Fix:
- Compose the dedup key as issueId#action#webhookTimestamp. Linear
  reuses webhookTimestamp across retries of one delivery (correctly
  collapsing them) but assigns a fresh timestamp to each new delivery
  (no collapse). This is the primitive documented in linear.app's
  webhook spec; the payload's webhookId field is the webhook-config
  id, not per-delivery.
- Raise DEDUP_TTL_SECONDS to 8h, comfortably over Linear's 7h retry
  horizon.

Added a regression test that two distinct deliveries for the same
issue both dispatch, and an assertion that the TTL is >= 7h.
When shouldTrigger rejects an update event, add current_labels,
updated_from_keys, updated_from_label_ids, and current_label_ids to
the log line so we can diagnose cases where the filter drops a valid
label-added transition.

No behaviour change; diagnostic-only.
Three CLI UX improvements discovered during first-time setup:

1. Auto-link the token owner during `bgagent linear setup`.
   After storing the webhook secret and API key, query Linear's
   `viewer` + `organization` via GraphQL using the token, decode the
   caller's Cognito sub from the cached id_token, and write the
   LinearUserMapping row directly (status='active',
   link_method='auto_setup'). This skips the manual code-exchange
   ceremony for the common case where the person installing ABCA for
   their workspace is the same person who'll trigger tasks.

   Failures are fail-open: if GraphQL errors, the API key is
   malformed, or the CLI isn't logged in, setup prints a warning and
   continues. Secrets are still stored; the user can run the link
   flow manually.

2. `bgagent linear list-projects` — lists all projects visible to the
   stored API token with their full UUIDs. Solves the "Linear project
   URL contains a truncated UUID" footgun.

3. UUID validation on `onboard-project`. RFC-4122 shape check up-
   front, with a clear error message pointing at `list-projects`.
   Avoids silent mapping-row writes for slug-shaped UUIDs that would
   never match a real Linear payload.

Tests cover the auto-link branch, fail-open cases, and the GraphQL
wire shape.
Mirror the Slack integration's terminal-emoji status UX on Linear
issues: post 👀 when the agent container starts, swap it for ✅ on
success or ❌ on failure (including the pipeline's outer-except
crash path).

Implementation notes:
- Linear's MCP v1 does not expose a reactions tool (verified
  2026-05-06 against the `tools` array in the Claude Agent SDK's
  system init). We use direct GraphQL against
  `https://api.linear.app/graphql` with the same token the MCP
  server uses. Retire this module in favour of a prompt addendum
  once Linear MCP ships `create_reaction`.
- Gated on `channel_source == 'linear'` + `linear_issue_id` in
  channel_metadata. No-op for Slack/API/webhook channels.
- `react_task_started` returns the reaction id; the pipeline threads
  it into `react_task_finished` so we can `reactionDelete` the 👀
  before posting the terminal emoji. Matches Slack's behaviour of
  not leaving a lingering "watching" marker.
- All GraphQL failures are logged and swallowed — reactions are
  advisory UX, never load-bearing.
- Crash-path (pipeline.py outer except): calls
  react_task_finished(success=False, ...) as a best-effort so a
  crashed/failed task doesn't leave a dangling 👀.

Tests cover:
- Channel gate (slack/api/webhook all no-op; linear with no
  issue_id also no-op)
- GraphQL wire shape and variables
- Delete-then-create swap on terminal transition
- Fallback path when started_reaction_id is None
- All error swallowing paths (HTTP error, connection exception,
  GraphQL errors, missing token)
The Linear MCP's tool for updating an issue is
mcp__linear-server__save_issue, not update_issue. Earlier prompt
referenced the wrong name; the agent silently skipped the state
transition on small/fast tasks because the advisory framing
("use its tools to keep the reporter informed") let the model
deprioritize the calls.

Rewrite the addendum as a required three-step contract:
1. On start — save_comment 🤖 Starting
2. On PR open — save_comment with PR URL + save_issue (transition
   to In Review; fall back to In Progress; skip if neither state
   exists rather than loop list_issue_statuses)
3. On completion/failure — save_comment with final status

Added explicit "If neither exists, skip the state transition — the
PR comment alone is enough. Do not invent state names or loop on
list_issue_statuses." so an impoverished workspace state-set
doesn't trigger retry behaviour.
test_background_thread_failure_503_and_backup_terminal_write polled
/ping for _background_pipeline_failed = True, but the flag flips
several lines before write_terminal runs in the same except block.
On the first run after a fresh import the poll could win the race
and mock_write.assert_called() would fire before the call happened.

Swap the poll-the-flag pattern for a wait-until-the-background-thread-
has-exited loop, which guarantees both side effects (the flag flip
and the backup write_terminal call) have completed before we assert.

Observed flaky 2026-05-06 in a full-suite run; reliably green after
the fix across 5 consecutive runs.
Restructure the setup guide so the webhook URL is available before
the user needs to create the webhook (prior ordering was circular).

- Step 1: generate Linear API key
- Step 2: run `bgagent linear setup` — it prints the webhook URL
  and pauses at the signing-secret prompt
- Step 3: create the webhook in Linear, copy the signing secret
- Step 4: back in the terminal, paste the signing secret + API key;
  auto-link runs for the token owner
- Step 5: `bgagent linear onboard-project <uuid>` — using the
  `list-projects` command to find the full UUID (the URL slug is
  truncated)
- Step 6: manual code-exchange flow, now marked as fallback for
  linking additional Linear users beyond the token owner

Synced to docs/src/content/docs/using/Linear-setup-guide.md via
mise //docs:sync.
Merges Sam's PR aws-samples#52 (FanOutConsumer dispatcher) and adjacent upstream
changes on top of the Linear integration work.

Key reconciliations:

- `ChannelSource` union keeps all four members ('api', 'webhook',
  'slack', 'linear') with upstream's better docstring. Fixed an
  upstream test that expected 'slack' to be an invalid value.
- `TaskRecord` / `TaskConfig` / pipeline signatures now include both
  our `channel_source` + `channel_metadata` and upstream's `trace` +
  `user_id` fields (all additive).
- Agent pipeline runs both new hooks in order: Linear reactions
  (swap 👀 for ✅/❌) then upstream's `_maybe_upload_trace`, so the
  reaction fires before any trace S3 upload that could be slow.
- Crash path invokes both Linear reaction (best-effort ❌) and
  `_maybe_upload_trace` (debug artifact) so a crashed task still
  produces a meaningful Linear signal and a usable trace.
- `server.py` adopts upstream's refactored _validate_required_params /
  _spawn_background structure; our `channel_source` + `channel_metadata`
  are threaded through the params dict.
- WAF scope-down (Linear webhook path exemption from CRS body-size
  rule) and our 8h dedup TTL are preserved unchanged.
- Agent test count went from 306 to 526; CDK suite count went from 853
  to 1121. All green.

FanOutConsumer (new from aws-samples#52) and LinearIntegration sit side-by-side
in `stacks/agent.ts` — FanOutConsumer is independent of Linear's
wiring. Linear continues to route outbound updates via agent-side MCP
rather than subscribing to the fanout dispatcher (by design; keeps
per-task Linear comment/state updates in the container where the task
actually runs).

See things-to-do.md for the "evaluate linear-notify under
FanOutConsumer" follow-up. Keeping it parked: MCP-based outbound is
already shipped and validated; moving to a fanout subscriber would be
a strategy change, not a bug fix.
Roadmap doc has no Linear content; referring to it misled readers.
Replace the link with a plain explanation of the admin-assisted path.
Describe the v1 linking model accurately: ABCA ships the consumer
command (`bgagent linear link <code>`) but no Linear-side code
generator — the comment-triggered flow is a planned v1.1 follow-up.
Design around API-token-owner-as-sole-submitter until then.
Fix two `no-duplicate-imports` violations in the Slack tests by merging
`import type` + `import` pairs into single imports. ESLint/ruff also
applied cosmetic reformatting (whitespace alignment in object literals,
line-splitting long inline objects). One `# noqa: S105` added to
`LINEAR_API_TOKEN_ENV` — it's an env var *name*, not a secret value, so
ruff's hardcoded-password rule is a false positive there.

No behavioural changes. Ensures `mise run build` passes the self-
mutation guard in .github/workflows/build.yml.
@krokoko
Copy link
Copy Markdown
Contributor

krokoko commented May 6, 2026

Feedback from automated review:

Required before merge

#: 1
Issue: Dedup row persists after Lambda invoke failure — all Linear retries (+1m,+1h, +6h) fall within the 8h TTL, silently dropping the task forever if invoke keeps failing
Location: cdk/src/handlers/linear-webhook.ts ~L135
Fix: Delete the dedup row in the invoke-error catch block, or move the write to
after successful invocation
────────────────────────────────────────
#: 2
Issue: CLI/CDK ChannelSource type drift — CDK updated to 'api' | 'webhook' |
'slack' | 'linear' but cli/src/types.ts:30 still has 'api' | 'webhook'
Location: cli/src/types.ts
Fix: Add 'slack' | 'linear' to the CLI union

Recommended follow-ups (v1.1)

  • Processor feedback on pre-container failures — when createTaskCore returns
    non-201, post a Linear comment explaining why (3 silent-failure cases observed
    during smoke)
  • Escalation on persistent Linear auth failure — linear_reactions.py logs WARN
    forever on revoked tokens; add error escalation after N consecutive failures
  • Replace locals().get("linear_eyes_reaction_id") with explicit
    linear_eyes_reaction_id = None initialization before the try block
  • Narrow except Exception in resolve_linear_api_token to botocore errors; switch
    from print() to shell.log()
  • Auto-link fallback path — if auto-link fails, the suggested bgagent linear
    link requires a code the user cannot generate in v1; clarify the
    admin-assisted path

What's done well

  • MCP-based outbound is the right topology — avoids the 2-consumer DDB Streams
    limit, keeps per-issue state in Linear (not channel_metadata), eliminates a
    Lambda notification hop
  • Dedup key design (issueId#action#webhookTimestamp, 8h TTL) correctly exceeds
    Linear's 7h retry horizon and collapses retries while allowing distinct events
  • WAF scope-down is properly scoped (EXACTLY match, not prefix) and the route
    retains HMAC + replay guard + rate limiting
  • Security posture — timing-safe HMAC comparison with rotation-retry, 60s replay
    window, no secrets logged, IAM least-privilege throughout
  • Slack review response is thorough — 20 items addressed with regression tests
    that fail the pre-fix code
  • Test growth (CDK 853→1121, agent 306→526, CLI 76) with construct-level,
    handler-level, and integration-style coverage

@krokoko
Copy link
Copy Markdown
Contributor

krokoko commented May 6, 2026

Just adding as a follow up: will open an issue after to use AgentCore gateway

The current approach isn't wrong — it's a conscious shortcut for v1. But as the project adds more external MCP integrations, the .mcp.json injection pattern doesn't scale. The Gateway is the project's stated integration point for external tools, and Linear should converge on it.

1) Linear webhook: roll back dedup row on Lambda invoke failure

Previously, the dedup row was written before invoking the processor
Lambda. If InvokeCommand failed (throttle, 5xx, etc.), the handler
returned 500 and relied on Linear's retry schedule (+1m / +1h / +6h) to
redeliver — but all three retries fell inside the 8h dedup TTL, so each
one hit ConditionalCheckFailedException and silently acked as "already
processed." Net: a single transient invoke failure dropped the task
for 8 hours.

Fix: on invoke failure, delete the dedup row before returning 500 so
the next retry can reserve it again. A secondary try/catch logs (but
does not re-raise) if the delete itself fails — the primary invoke
error is the user-visible signal. Added two regression tests.

2) CLI ChannelSource sync with CDK

cli/src/types.ts:30 still had 'api' | 'webhook'; CDK had
'api' | 'webhook' | 'slack' | 'linear' since the Slack/Linear inbound
adapters landed. Any CLI switch on TaskDetail.channel_source was
narrowing against a stale union. Widened the CLI type and refreshed
the jsdoc to match the CDK side.

Co-Authored-By: Claude Opus 4.7 (1M context) <noreply@anthropic.com>
@isadeks
Copy link
Copy Markdown
Contributor Author

isadeks commented May 6, 2026

Thanks for the thorough review — both blockers fixed in 7d652a4.

#1 — Dedup row on invoke failure. Agreed this is a silent-drop trap; the 8h TTL fully swallows Linear's +1m/+1h/+6h retry horizon. Fix was the "delete on invoke-error" shape (not "write after invoke") — keeping the write before invoke preserves the dedup guarantee for the common case where a retry races an in-flight ack. Added two regression tests: one asserts the DeleteCommand fires with the exact dedup key on invoke failure, one asserts we still return 500 (not swallow the invoke error) if the rollback itself fails. Patch: cdk/src/handlers/linear-webhook.ts ~L158-L178.

#2 — CLI ChannelSource drift. Widened cli/src/types.ts:30 to 'api' | 'webhook' | 'slack' | 'linear' and refreshed the jsdoc to mirror cdk/src/handlers/shared/types.ts. CLI compile clean, 196 CLI tests green.

v1.1 follow-ups — all five parked in our internal backlog and slotted for the v1.1 pass:

  • Processor posts a Linear comment on pre-container failures (already the top v1.1 item — this is what the "parked v1.1" row in the PR body calls out)
  • Escalation on persistent Linear auth failure in linear_reactions.py
  • Explicit linear_eyes_reaction_id = None init instead of locals().get()
  • Narrow except Exception → botocore in resolve_linear_api_token + print()shell.log()
  • Admin-assisted fallback documentation for auto-link failures

AgentCore Gateway note — acknowledged. Convergence on Gateway as the MCP integration point makes sense as the project adds more external MCPs; happy to pick that up as a follow-up once the issue lands.

Full test counts post-fix: CDK 1170 (+2 rollback tests), agent 526, CLI 196 — all green.

@krokoko krokoko added this pull request to the merge queue May 7, 2026
Merged via the queue into aws-samples:main with commit 31b48d5 May 7, 2026
2 checks passed
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants