This template is to demonstrate how to trace the observations per execution ID in Langfuse via ingestion API.
https://cloud.langfuse.com/api/public/ingestionGeneric Credential Type with a Basic Auth: username = you_public_key, password = your_secret_key.Trigger: the workflow is executed by another workflow after an AI run finishes (input parameter execution_id).
Remove duplicates
Ensures we only process each execution_id once (optional but recommended).
Wait to get execution data
Delay (60-80 secs) so totals and per-step metrics are available.
Get execution
Fetches workflow metadata and token totals.
Code: structure execution data
Normalizes your run into an array of perModelRuns with model, tokens, latency, and text previews.
Split Out → Loop Over Items
Iterates each run step.
Code: prepare JSON for Langfuse
Builds a batch with:
HTTP Request to Langfuse
Posts the batch. Optional short Wait between sends.
parentObservationId on the generation to nest under spans.sessionId strategy (per workflow, per user, etc.). If some steps don’t produce tokens, compute and set usage yourself before sending.