Set up Respan
Set up Respan
- Sign up — Create an account at platform.respan.ai
- Create an API key — Generate one on the API keys page
- Add credits or a provider key — Add credits on the Credits page or connect your own provider key on the Integrations page
What is prompt management?
Prompt management lets you create, version, and deploy prompt templates centrally — instead of hardcoding prompts in your application, reference them by ID.Prerequisites
Before you begin, make sure you have:- Respan API key — get one from the API keys page. See API keys for details.
- LLM provider key — add your provider credentials (e.g. OpenAI) on the Providers page. See LLM provider keys for details.
Create your first prompt
Create a new prompt
Go to the Prompts page and click Create new prompt. Name your prompt and add a description.

Configure the prompt
In the Editor tab, set parameters like model, temperature, max tokens, and top P in the right sidebar.

Write content with variables
Click + Add message to add messages. Use 
{{variable_name}} for dynamic content — see Variables for Jinja templates, JSON inputs, and more.Variable names must use underscores:
{{task_description}} not {{task description}}.
Test and commit
- Add values for each variable in the Variables tab.
- Click Run to test.
- Click Commit and write a commit message to save this version.

Deploy to production
Go to the Deployments tab and click Deploy. See Deployment & versioning for version pinning, rollbacks, and overrides.

Use your prompt in code
Find the Prompt ID in the Overview panel on the Prompts page.
OpenAI SDKs strip v2 fields like
schema_version and patch. Prompt schema v2 requires raw HTTP requests.model and messages — the prompt configuration is used automatically.
Legacy prompt schema v1 (default)
Legacy prompt schema v1 (default)
With v1, use See Prompt merge modes (v1 vs v2) for full details.
override: true to let the prompt config win over request-body parameters. This is the default when schema_version is omitted.Monitor your prompts
Filter logs by prompt name on the Logs page to track usage, response times, and token consumption. See Prompt logging for logging setup.
Next steps
- Variables — Jinja templates, conditionals, and JSON inputs
- JSON schema — structured output for consistent response formats
- Prompt composition — reference prompts inside other prompts
- Prompt merge modes (v1 vs v2) — control how prompt config and request params are merged
- Deployment & versioning — version pinning, rollbacks, and overrides
- Streaming — enable streaming for prompt responses
- Playground — test and iterate on prompts interactively