Skip to main content

Quickstart

Choose your mode and follow the steps below.

Prerequisites

  • Node.js 18+
  • An LLM provider API key (OpenAI or Anthropic)

Step 1: Create Your Project

npm create agentmark@latest -- --cloud
The CLI walks you through selecting your language, adapter, and IDE setup. Choose AgentMark Cloud as the deployment mode.
You can skip the interactive prompts by passing flags directly:
npm create agentmark@latest -- --cloud --typescript --adapter ai-sdk --client skip
Available flags: --typescript / --python, --adapter <name> (ai-sdk, claude-agent-sdk, mastra, pydantic-ai), --cloud / --self-host, --client <ide> (claude-code, cursor, vscode, zed, skip), --path <dir>, --api-key <key>.

Step 2: Sync Your App

  1. Commit and push your project to a Git repository
  2. In the AgentMark Dashboard, navigate to your app
  3. Add your LLM provider API key in Settings > Environment Variables
  4. Connect your repository
Sync RepositoryOnce connected, the Dashboard syncs your prompt files and deploys your handler automatically.

Step 3: Run Your First Prompt

Open a prompt in the Dashboard and click Run. The platform executes it on your deployed handler and streams results back in real time.Running a prompt in the AgentMark cloud Dashboard

Step 4: Run an Experiment

Experiments test a prompt against a dataset and score the results with evaluators. Your project includes an example dataset and prompt ready to go.
  1. Navigate to the party-planner prompt in the Dashboard
  2. Open the Experiments tab
  3. Click Run Experiment
  4. Review the results — scores, pass rates, and individual outputs
Experiment results in the AgentMark cloud Dashboard showing scores, cost, and latency

Step 5: View Your Traces

Every prompt and experiment execution is automatically traced. Navigate to the Traces page to see the full execution timeline — input/output, token usage, cost, and latency.Traces list in the AgentMark cloud Dashboard

What’s in Your Project

File / DirectoryPurpose
agentmark/Prompt templates (.prompt.mdx) and test datasets (.jsonl)
agentmark.client.tsClient configuration — models, tools, and loader setup
agentmark.jsonProject configuration (models, scores, schema)
agentmark.types.tsAuto-generated TypeScript types for your prompts
handler.tsHandler for cloud deployment (Cloud mode only)
dev-entry.tsDevelopment server entry point (customizable)
index.tsExample application entry point
.envEnvironment variables (API keys)

Next Steps

Build Prompts

Create prompts with tools, structured output, and components

Evaluate

Test your prompts with datasets and automated evaluators

Observe

Monitor traces, sessions, and costs in production

Integrations

Connect with Vercel AI SDK, Pydantic AI, Mastra, and more

Have Questions?

We’re here to help! Choose the best way to reach us: