Prompt Config

The prompt config in Langfuse is an optional arbitrary JSON object attached to each prompt that stores structured data such as model parameters (like model name, temperature), function/tool parameters, or JSON schemas.

This config is versioned together with the prompt and allows you to manage all parameters associated with a prompt in one place, making it easier to deploy, update, and track prompt behavior and performance across different versions and environments.

!Prompt config

Use cases

  • Change the model your application uses without touching code
  • Manage and version the function/tool JSON schemas your application uses
  • Define the JSON schema for your application’s responses
  • Set and manage you model parameters such as temperature, top_p, max_tokens, etc.

Cookbooks

The following two example cookbooks demonstrate how to use the config attribute: