TypeScript SDK for Tracia - store, test, and trace LLM prompts.
npm install traciaIf using runLocal() or runResponses() to call LLMs directly, also install ai and your provider:
npm install ai @ai-sdk/openai # for GPT, o1, o3, etc.
npm install ai @ai-sdk/anthropic # for Claude
npm install ai @ai-sdk/google # for Geminiimport { Tracia } from 'tracia';
const tracia = new Tracia({ apiKey: process.env.TRACIA_API_KEY });
// Run a prompt stored in Tracia
const result = await tracia.prompts.run('welcome-email', { name: 'Alice' });
console.log(result.text);
// Or call LLM directly with automatic tracing
const response = await tracia.runLocal({
model: 'gpt-4o',
messages: [{ role: 'user', content: 'Hello!' }],
});
console.log(response.text);- Run prompts with variables and automatic tracing
- Manage prompts programmatically (create, update, delete)
- Full TypeScript support with exported types
- Zero dependencies (native fetch)
MIT