diff --git a/agent-simulations/getting-started.mdx b/agent-simulations/getting-started.mdx
index 579d91e..cf370c3 100644
--- a/agent-simulations/getting-started.mdx
+++ b/agent-simulations/getting-started.mdx
@@ -2,6 +2,10 @@
title: Getting Started
---
+
+ **Quick setup?** [Copy the scenarios prompt](/skills/code-prompts#add-scenario-tests) into your coding agent to add simulation tests automatically.
+
+
This guide will walk you through the basic setup required to run your first simulation and see the results in LangWatch.
For more in-depth information and advanced use cases, please refer to the official [`scenario` library documentation](https://github.com/langwatch/scenario).
diff --git a/agent-simulations/introduction.mdx b/agent-simulations/introduction.mdx
index c93369c..d9747c9 100644
--- a/agent-simulations/introduction.mdx
+++ b/agent-simulations/introduction.mdx
@@ -4,6 +4,10 @@ sidebarTitle: Introduction
keywords: langwatch, agent simulations, agent testing, agent development, agent development, agent testing
---
+
+ **Quick setup?** [Copy the scenarios prompt](/skills/code-prompts#add-scenario-tests) into your coding agent to add simulation tests automatically.
+
+
# What are Agent Simulations?
Agent simulations are a powerful approach to testing AI agents that goes beyond traditional evaluation methods. Unlike static input-output testing, simulations test your agent's behavior in realistic, multi-turn conversations that mimic how real users would interact with your system.
diff --git a/better-agents/overview.mdx b/better-agents/overview.mdx
index 86b1ad4..ecc3d5f 100644
--- a/better-agents/overview.mdx
+++ b/better-agents/overview.mdx
@@ -10,7 +10,7 @@ Better Agents is a CLI tool and a set of standards for building **reliable, test
Use your preferred stack—Agno, Mastra, Vercel AI, Google ADK, or anything else. Better Agents doesn't replace your stack, it stabilizes it.
-Already have a project? Add evaluations, observability, and scenarios to your existing agent project. See the [Integration Guide](/integration/overview) to get started.
+ **Already have an agent?** You don't need Better Agents -- go to [LangWatch Skills](/skills/directory) to add tracing, evaluations, scenarios, and prompt versioning to your existing project.
## Quick Start
diff --git a/datasets/overview.mdx b/datasets/overview.mdx
index 2280933..fd6b305 100644
--- a/datasets/overview.mdx
+++ b/datasets/overview.mdx
@@ -4,6 +4,10 @@ sidebarTitle: Overview
description: Create and manage datasets in LangWatch to build evaluation sets for LLMs and structured AI agent testing.
---
+
+ **Let your agent set this up.** [Copy the evaluations prompt](/skills/code-prompts#set-up-evaluations) into your coding agent to get started automatically.
+
+
## Create datasets
LangWatch allows you to create and manage datasets, with a built-in excel-like interface for collaborating with your team.
diff --git a/docs.json b/docs.json
index d7af8d2..cffd333 100644
--- a/docs.json
+++ b/docs.json
@@ -17,7 +17,7 @@
},
"favicon": "/favicon.svg",
"banner": {
- "content": "**[LangWatch MCP is live](https://langwatch.ai/docs/integration/mcp):** Your AI coding assistant can now build, version, and ship evals — no platform context-switching required.",
+ "content": "**[LangWatch MCP is live](https://langwatch.ai/docs/integration/mcp):** Your AI coding assistant can now build, version, and ship evals \u2014 no platform context-switching required.",
"dismissible": true
},
"contextual": {
@@ -58,8 +58,17 @@
"group": "Get Started",
"pages": [
"introduction",
- "better-agents/overview",
- "integration/mcp"
+ {
+ "group": "LangWatch Skills",
+ "pages": [
+ "skills/directory",
+ "skills/code-prompts",
+ "skills/platform-prompts",
+ "skills/recipes"
+ ]
+ },
+ "integration/mcp",
+ "better-agents/overview"
]
},
{
@@ -717,6 +726,26 @@
{
"source": "/hybrid-setup/sso-setup-langwatch",
"destination": "/self-hosting/sso-setup-langwatch"
+ },
+ {
+ "source": "/skills/overview",
+ "destination": "/skills/directory"
+ },
+ {
+ "source": "/skills/developers",
+ "destination": "/skills/code-prompts"
+ },
+ {
+ "source": "/skills/teams",
+ "destination": "/skills/platform-prompts"
+ },
+ {
+ "source": "/skills/platform",
+ "destination": "/skills/platform-prompts"
+ },
+ {
+ "source": "/skills/manual",
+ "destination": "/integration/quick-start"
}
]
}
diff --git a/evaluations/evaluators/overview.mdx b/evaluations/evaluators/overview.mdx
index 91aac35..eecd347 100644
--- a/evaluations/evaluators/overview.mdx
+++ b/evaluations/evaluators/overview.mdx
@@ -4,6 +4,10 @@ sidebarTitle: Overview
description: Understand evaluators - the scoring functions that assess your LLM outputs for quality, safety, and correctness.
---
+
+ **Let your agent set this up.** [Copy the evaluations prompt](/skills/code-prompts#set-up-evaluations) into your coding agent to get started automatically.
+
+
Evaluators are scoring functions that assess the quality of your LLM's outputs. They're the building blocks for [experiments](/evaluations/experiments/overview), [online evaluation](/evaluations/online-evaluation/overview), and [guardrails](/evaluations/guardrails/overview).
## Choose Your Approach
diff --git a/evaluations/experiments/overview.mdx b/evaluations/experiments/overview.mdx
index a628b6c..331b829 100644
--- a/evaluations/experiments/overview.mdx
+++ b/evaluations/experiments/overview.mdx
@@ -4,6 +4,10 @@ sidebarTitle: Overview
description: Run batch tests on your LLM applications to measure quality, compare configurations, and catch regressions before production.
---
+
+ **Let your agent set this up.** [Copy the evaluations prompt](/skills/code-prompts#set-up-evaluations) into your coding agent to get started automatically.
+
+
Experiments let you systematically test your LLM applications before deploying to production. Run your prompts, models, or agents against datasets and measure quality with evaluators.
## What is an Experiment?
diff --git a/evaluations/experiments/sdk.mdx b/evaluations/experiments/sdk.mdx
index 127a228..da85292 100644
--- a/evaluations/experiments/sdk.mdx
+++ b/evaluations/experiments/sdk.mdx
@@ -4,6 +4,10 @@ sidebarTitle: Via SDK
description: Run experiments programmatically from notebooks or scripts to batch test your LLM applications.
---
+
+ **Let your agent set this up.** [Copy the evaluations prompt](/skills/code-prompts#set-up-evaluations) into your coding agent to get started automatically.
+
+
LangWatch makes it easy to run experiments from code.
Just add a few lines to start tracking your experiments.
diff --git a/evaluations/guardrails/code-integration.mdx b/evaluations/guardrails/code-integration.mdx
index fd374b7..465123d 100644
--- a/evaluations/guardrails/code-integration.mdx
+++ b/evaluations/guardrails/code-integration.mdx
@@ -4,6 +4,10 @@ sidebarTitle: Code Integration
description: Add guardrails to your LLM application to block harmful content in real-time.
---
+
+ **Let your agent set this up.** [Copy the evaluations prompt](/skills/code-prompts#set-up-evaluations) into your coding agent to get started automatically.
+
+
This guide shows how to integrate guardrails into your application using the LangWatch SDK. Guardrails run evaluators synchronously and return results you can act on immediately.
## Basic Usage
diff --git a/evaluations/guardrails/overview.mdx b/evaluations/guardrails/overview.mdx
index d4bbbe6..705b7c8 100644
--- a/evaluations/guardrails/overview.mdx
+++ b/evaluations/guardrails/overview.mdx
@@ -4,6 +4,10 @@ sidebarTitle: Overview
description: Block or modify harmful LLM responses in real-time to enforce safety and policy constraints.
---
+
+ **Let your agent set this up.** [Copy the evaluations prompt](/skills/code-prompts#set-up-evaluations) into your coding agent to get started automatically.
+
+
Guardrails are evaluators that run in real-time and **act** on the results - blocking, modifying, or rejecting responses that violate your safety or policy rules. Unlike [monitors](/evaluations/online-evaluation/overview) which only measure and alert, guardrails actively prevent harmful content from reaching users.
## Guardrails vs Monitors
diff --git a/evaluations/online-evaluation/overview.mdx b/evaluations/online-evaluation/overview.mdx
index eead82d..99654ae 100644
--- a/evaluations/online-evaluation/overview.mdx
+++ b/evaluations/online-evaluation/overview.mdx
@@ -4,6 +4,10 @@ sidebarTitle: Overview
description: Continuously score and monitor your LLM's production traffic for quality and safety with online evaluation.
---
+
+ **Let your agent set this up.** [Copy the evaluations prompt](/skills/code-prompts#set-up-evaluations) into your coding agent to get started automatically.
+
+
Online evaluation lets you continuously score your LLM's production traffic. Unlike [experiments](/evaluations/experiments/overview) which test before deployment, online evaluation monitors your live application to catch quality issues, detect regressions, and ensure safety.
diff --git a/evaluations/overview.mdx b/evaluations/overview.mdx
index 80b43d3..f62f0f2 100644
--- a/evaluations/overview.mdx
+++ b/evaluations/overview.mdx
@@ -4,6 +4,10 @@ sidebarTitle: Overview
description: Ensure quality and safety for your LLM applications with experiments, online evaluation, guardrails, and evaluators.
---
+
+ **Let your agent set this up.** [Copy the evaluations prompt](/skills/code-prompts#set-up-evaluations) into your coding agent to get started automatically.
+
+
LangWatch provides comprehensive evaluations tools for your LLM applications. Whether you're evaluating before deployment or monitoring in production, we have you covered.
## The Agent Evaluation Lifecycle
diff --git a/integration/go/guide.mdx b/integration/go/guide.mdx
index 479af58..6815e14 100644
--- a/integration/go/guide.mdx
+++ b/integration/go/guide.mdx
@@ -7,6 +7,10 @@ keywords: LangWatch, Go, Golang, SDK, integration, guide, setup, tracing, spans,
import LLMsTxtProtip from "/snippets/llms-txt-protip.mdx";
+
+ **Quick setup?** Instead of following these steps manually, [copy a prompt](/skills/code-prompts#instrument-my-code) into your coding agent and it will set this up for you automatically.
+
+