Skip to main content

March 18, 2026

Deployment 1

New Features

GPT-5.4 Mini and Nano Model Support PromptLayer now supports OpenAI’s GPT-5.4 Mini and GPT-5.4 Nano models in the Playground and API.
  • Configure reasoning effort, verbosity, and response format options for both models
  • Leverage lower-cost alternatives to GPT-5.4 for appropriate use cases
  • Access prompt caching capabilities for improved performance

Improvements

  • Images now display correctly in Playground chat mode
  • Dataset columns reordered to show promptlayer_url before prompt for easier request navigation
  • Enhanced model configuration options for GPT-5.4 series models

Deployment 2

New Features

Chat History Injection for Prompt Templates Prompt template blocks in evaluations can now inject chat history messages from a dataset column directly into chat prompts.
  • Enable chat history injection in the Advanced Settings section of the prompt template block configuration
  • Select a source column containing message objects with role and content fields
  • Messages are automatically appended to the end of your prompt template
  • Supports both JSON and JSON5 formatted message lists for flexible data sources

March 17, 2026

Deployment 1

New Features

Public API for Dataset Rows New REST API endpoint to programmatically retrieve paginated rows from Datasets, enabling integration with external tools and workflows.
  • Access dataset rows via /api/public/v2/datasets/{id}/rows with support for search and pagination
  • Returns structured row data matching dataset column definitions
  • Supports up to 100 rows per request with flexible filtering
Public API for Evaluation Results New REST API endpoint to fetch evaluation results programmatically, combining dataset inputs with evaluation scores.
  • Access evaluation rows via /api/public/v2/evaluations/{id}/rows endpoint
  • Returns both dataset input variables and evaluation cell results in a unified format
  • Enables automated analysis and reporting on evaluation performance
Enhanced OpenTelemetry Tracing Support Expanded tracing instrumentation to support additional SDKs and frameworks for automatic request logging.
  • Improved compatibility with diverse instrumentation libraries
  • More reliable extraction of provider and model information from traces

Improvements

  • Enhanced input variable detection in Playground with truncation and tooltip support for long variable names
  • Improved link handling in Wrangler AI for better navigation across all resource types
  • Refined organization members table pagination for more consistent data loading
  • Updated API key input styling across provider configuration pages for better visual consistency
  • Improved deduplication logic for file annotations to prevent duplicate entries

Deployment 2

New Features

Deployment Usage Analytics Track token consumption and session activity across all your prompt deployments with new organization-level analytics.
  • View daily token usage broken down by individual deployments
  • Monitor session counts per deployment over time
  • Compare usage across public and private deployments
  • Access historical usage data for capacity planning
Prompt Remixing Enable users to create their own versions of your shared prompts directly from the deployment interface.
  • Toggle remix capability on or off for any deployment
  • Users can fork and customize prompts while maintaining attribution
  • Remixed versions are saved to the user’s own workspace
  • Great for templates and starter prompts you want others to build upon

Improvements

  • Enhanced deployment management UI with improved session visibility and controls
  • Added batch execution support for shared prompt deployments
  • Improved file upload handling and multipart processing for large media files
  • Better dataset creation flow with prompt template selection from request history
  • Enhanced permission checks and access controls across deployment endpoints

March 16, 2026

Deployment 1

New Features

Runtime Tool Variables in Prompt Templates Dynamic tool injection now supports variable substitution, enabling templates to generate tool definitions on-the-fly based on runtime context.
  • Tool schemas can include variables (e.g., {{user_id}}, {{domain}}) that resolve during template execution
  • Supports nested variable resolution within tool parameters and descriptions
  • Enables dynamic function calling patterns where tool availability adapts to request context
Chat Message Annotations Annotations can now be added directly to individual messages in chat-mode conversations for improved debugging and analysis.
  • Attach metadata, tags, or notes to specific assistant or user messages
  • Track message-level performance metrics and quality assessments

Improvements

  • Enhanced function/tool overview dialog displays complete schema details with improved formatting
  • Improved citation modal rendering with better support for complex reference structures
  • Streamlined Docker image build process reduces deployment time
  • Better visual distinction between different tool types in the functions list view
  • Optimized frontend bundle size through refined dependency management

March 15, 2026

Deployment 1

New Features

Public API Request Search Endpoint A new /api/public/v2/requests/search endpoint enables programmatic searching of request logs with structured filters.
  • Search logs using the same filtering capabilities available in the dashboard
  • Support for complex filter groups and structured queries
  • Rate-limited to 10 requests per minute with up to 25 results per page
BYOK Playground Limit Exemption Users who configure their own API keys (Bring Your Own Key) are now exempt from daily playground run limits.
  • Unlimited playground testing when using your own API credentials
  • Cost control stays with your organization while removing artificial usage caps

Improvements

  • Improved dataset column JSON parsing with better error handling for sparse or malformed data
  • Enhanced workspace member management interface with clearer permission displays
  • Fixed API key modal display to better communicate rate limits and usage policies
  • Standardized public API endpoint structure (moved GET /api/public/v2/request/<id> to /requests/<id>)
  • Added permission checks to dataset creation and editing endpoints to enforce role-based access control

March 12, 2026

Deployment 1

New Features

Hybrid Search for Registry Enhanced search across prompts, workflows, and datasets combining keyword matching with semantic understanding for more relevant results.
  • Search results now surface contextually similar items even when exact keywords don’t match
  • Improved search ranking considers both text relevance and semantic meaning
  • Background indexing keeps search up-to-date as you modify registry items
Scroll Lock in Playground Chat Chat panel now maintains your scroll position when new messages arrive, preventing automatic jumping to the bottom.
  • Toggle scroll lock on/off to control whether new messages auto-scroll
  • Manually scrolling up automatically enables scroll lock
  • Scroll to bottom re-enables auto-scroll behavior

Improvements

  • Fixed search suggestions displaying incorrect text values in autocomplete dropdowns
  • Resolved f-string variable indexing issues when searching prompt templates
  • Added language field support for Google Code Execution tool responses
  • Improved citation display by preserving original model response annotations without deduplication
  • Enhanced registry list and grid views with optimized virtualization for faster rendering of large item collections

Deployment 2

New Features

OTLP Prompt Resolution by ID and Label Enhanced OpenTelemetry trace ingestion now supports flexible prompt identification and version resolution.
  • Spans can reference prompts by promptlayer.prompt.id in addition to promptlayer.prompt.name
  • Version resolution via promptlayer.prompt.label automatically links traces to labeled prompt versions
  • Improved error handling when prompt identifiers are incomplete or not found in the workspace
Duplicate Span Handling The /spans-bulk endpoint now intelligently handles duplicate span IDs to prevent data conflicts.
  • Duplicate spans within the same batch are deduplicated before insertion (first occurrence wins)
  • Duplicate spans across separate batches use upsert logic (last write wins)
  • Ensures trace data remains consistent when the same span is reported multiple times

Improvements

  • Added flask embed_recently_used_prompts command to backfill embeddings for prompts with recent traffic
  • Added flask normalize_recently_used_prompts command to backfill normalized content for recently-used prompt versions
  • Both commands support configurable look-back windows (default 365 days) and batch sizes for gradual processing
  • Enhanced test coverage for duplicate span handling scenarios in bulk span creation

Deployment 3

New Features

Structured Search Advanced filtering interface for request logs with improved query building and autocomplete suggestions.
  • Build complex filters using fields, operators, and values with keyboard-driven autocomplete
  • Support for nested metadata filtering with key_equals, key_not_equals, and key_contains operators
  • Multi-value selection for tags, labels, and metadata fields with in and not_in operators
  • Apply date range presets like “Last 5 minutes” or shorthand like “30d” for quick filtering

Improvements

  • Enhanced date picker with single date selection mode and custom preset support
  • Added NOT_IN operator support for identifier, string, array, and nested key-value fields
  • Improved table components with better row click handling and empty state messages
  • Added IS_EMPTY and IS_NOT_EMPTY operators for nested metadata filtering
  • Expanded operator support for input/output text fields to include STARTS_WITH and ENDS_WITH

Deployment 4

Improvements

  • Fixed search filters not correctly matching boolean and numeric metadata values (e.g., false, true, 42)
  • Resolved issue where changing search filters could trigger duplicate requests and cause stale results to display
  • Improved nested field filtering to properly match metadata values regardless of type (string, boolean, or number)

March 11, 2026

Deployment 1

New Features

Nested Search Support for Outputs and Input Variables Advanced search now supports filtering by output fields and input variables, matching the existing metadata search capabilities.
  • Search for specific output values using output:key=value syntax
  • Filter requests by input variable content with input_variables:key=value
  • Use autocomplete suggestions for both output keys and input variable keys in the search bar

Improvements

  • Improved snippet handling when creating prompt versions with overrides to ensure base references are used consistently
  • Enhanced search suggestion performance for nested field queries (metadata, outputs, input variables)
  • Input variables now preserve insertion order when rendering prompt templates
  • Added structured logging context showing workspace and user IDs for better debugging and support

March 10, 2026

Deployment 1

New Features

Model Override Support in Evaluations Enhanced evaluation workflows now preserve model configuration when routing between prompt templates and agents.
  • Model override settings are now correctly passed through evaluation interfaces
  • API type and model configuration IDs are properly maintained across workflow executions

Improvements

  • Fixed tool call detection in search indexing to correctly identify assistant messages with tool calls
  • Improved “Open Prompt” button functionality in image API evaluations to use correct routing
  • Enhanced build efficiency by adding path guards to skip unnecessary backend image builds when only documentation or configuration files change

Deployment 2

New Features

Prompt Starring Users can now star important prompts for quick access and organization.
  • Star/unstar prompts directly from the prompt template page
  • View list of users who have starred a prompt
  • Filter and prioritize frequently-used prompts
Structured Search for Request Logs Advanced filtering capabilities for request logs with precise search criteria.
  • Build complex queries using field-specific filters (metadata, tags, models, etc.)
  • Get autocomplete suggestions for search fields based on your workspace data
  • Sort results by any field with flexible ascending/descending order
Enhanced Tool Rendering Native display support for Anthropic code execution and shell command tools.
  • View bash command execution results with syntax highlighting
  • See code patches applied by AI agents in a readable format
  • Improved visualization of tool use blocks in chat interfaces

Improvements

  • Filter prompt templates by tags via the API using the tags parameter
  • Fixed “Open Prompt” button behavior in image-based evaluations to correctly navigate to prompt templates
  • Resolved race condition in workflow output nodes that could cause incorrect status codes
  • Added model override routing support for evaluation workflows
  • Improved prompt template list performance with optimized tag indexing

Deployment 3

New Features

Multi-Prompt Search Filtering Advanced search now supports filtering across multiple prompts simultaneously and combining filters with logical operators.
  • Apply filters to multiple prompt templates at once for cross-prompt analysis
  • Combine search conditions using AND/OR logic for more precise queries
  • Filter suggestions now respect existing search criteria for faster query building
Inline Item Creation in Sidebar Create new items directly from empty folders in the sidebar navigation without navigating away from your current view.
  • Click “New item” buttons that appear in expanded empty folders
  • Context-aware creation automatically places items in the correct folder
  • Streamlined workflow for organizing prompts, datasets, and other resources

Improvements

  • Search autocomplete suggestions now dynamically update based on active filters
  • Added support for null/not-null operators in numeric field searches
  • Enhanced folder navigation with visual indicators for active item context
  • Improved metadata value suggestions with better handling of nested fields
  • Optimized search performance for large workspaces with complex filter combinations

Deployment 4

New Features

Anthropic Text Editor Tool Support Added support for Anthropic’s text editor built-in tool, enabling AI assistants to view, create, and edit text files programmatically.
  • Available for both Anthropic and Vertex AI (Anthropic models) providers
  • Supports commands like view, create, insert, and string-based replace operations
  • Automatically handles text editor tool results in request logs and prompt templates

Improvements

  • Enhanced subscription tracking with monthly contract value and contract duration fields for better enterprise billing management
  • Fixed real-time event listener limits to prevent connection issues when multiple components subscribe to the same channel
  • Improved tool choice handling to correctly map Text Editor tool names in API requests

Deployment 5

New Features

Plain Text Search in Structured Search You can now use plain text search alongside structured filters to find request logs more quickly.
  • Performs full-text search across request inputs and outputs while applying your structured filters
  • Enables flexible searching when you need both keyword matching and precise filtering
Tool Names Search and Filtering Search and filter request logs by the tools called during execution.
  • Search for specific tool names using the search bar with autocomplete suggestions
  • Filter requests by tool names in structured search queries
  • Helps track which tools are being used across your prompts and workflows

Improvements

  • Search results now prioritize exact matches in request inputs and outputs when using plain text search
  • Tool name suggestions appear in the search bar autocomplete for faster filtering
  • Structured search queries support filtering by tool execution status and metadata

March 09, 2026

Deployment 1

New Features

OpenTelemetry Trace Ingestion Enhancements Support for modern OpenTelemetry semantic conventions and improved compatibility with observability libraries.
  • Added support for gzip-compressed OTLP trace payloads to reduce network overhead
  • Added support for newer gen_ai.input.messages and gen_ai.output.messages JSON format used by Ruby and other emerging instrumentations
  • Improved parsing of gen_ai.system_instructions to properly handle system prompts from different providers
  • Added automatic upsert logic for duplicate span IDs to ensure trace completeness when spans are sent multiple times
Template Rendering for Tool/Function Messages Improved handling of LLM-generated tool and function call messages in prompt templates.
  • Template validation now gracefully skips tool/function messages that contain JSON responses rather than user-authored templates
  • Prevents false template rendering errors when JSON braces in tool responses are mistaken for template syntax
  • Preserves support for legitimate template variables in few-shot tool examples

Improvements

  • Added playground_session_id to request log bulk endpoint responses for better session tracking
  • Improved JSON variable parsing to optimistically parse all string values, matching frontend batch-mode behavior
  • Enhanced OTLP function name inference to support more provider-specific operation types (embeddings, text completion, content generation)
  • Fixed provider family detection for Anthropic and Google AI models in OpenTelemetry traces
  • Improved error handling for malformed Content-Type headers in trace ingestion

March 07, 2026

Deployment 1

New Features

Anthropic Code Execution Tool Support Added support for Anthropic’s native code execution tool capability, enabling AI models to write and execute Python code during conversations.
  • Models can now generate and run code snippets directly within chat sessions
  • Code execution results are displayed inline with conversation history
  • Supports dynamic data analysis and computation workflows
Enhanced Trace Filtering with Metadata Search Introduced advanced filtering for traces using custom metadata keys, making it easier to find specific traces in production systems.
  • Search and filter traces by any custom metadata key stored in span attributes
  • Autocomplete suggestions help discover available metadata keys across your workspace
  • Filter results update in real-time as you type
OpenAI Shell Tool Integration Added built-in shell tool support for OpenAI models, allowing AI assistants to execute shell commands when explicitly enabled.
  • Enables automation workflows where models can interact with system commands
  • Integrates with OpenAI’s native tool calling infrastructure

Improvements

  • Improved playground session initialization to correctly handle tool and function definitions when opening from request logs
  • Enhanced request log input variable extraction to include tool/function data for better context when replaying requests
  • Streamlined “Open in Playground” workflow to preserve all tool configurations from original requests
  • Fixed trace metadata button display issues in the span details view
  • Normalized message content format to consistently use content blocks across chat interfaces
  • Improved Vite build configuration for better development server performance

March 05, 2026

Deployment 1

New Features

OpenTelemetry Trace Ingestion Native support for industry-standard OpenTelemetry Protocol (OTLP) trace ingestion, enabling seamless integration with existing observability tooling.
  • Ingest traces via standard OTLP/HTTP endpoint at /v1/traces
  • Automatic extraction of GenAI semantic conventions for OpenAI and Anthropic providers
  • Convert OTLP spans into PromptLayer request logs with proper error mapping and metadata preservation
Multi-Message Tool Response Handling Enhanced playground chat interface now supports submitting multiple tool response messages simultaneously.
  • Import and replay conversations with parallel tool calls from request logs
  • Maintain correct message ordering when tools are invoked across conversation turns
  • Proper hydration of chat history with multiple tool responses per assistant turn
Chat History Import from Request Logs Import conversation history directly from request logs into playground chat sessions.
  • Reset and re-seed chat from any logged request with one click
  • Automatically diff request messages against current template to extract conversation context
  • Per-variable-set chat history support for testing multiple scenarios simultaneously

Improvements

  • Fixed playground chat crashes when trace metadata contains non-string values during URL sharing
  • Resolved 500 errors when reading prompts that use legacy LangChain message format
  • Fixed “No response” display issue for template render errors in request logs
  • Improved image evaluation algorithm accuracy for visual content comparison
  • Enhanced workspace member invitation dialog with better field validation
  • Fixed chat message ordering when importing request logs with tool calls

March 04, 2026

Deployment 1

New Features

Google File Search Tool Support Native integration with Google’s File Search tool for Gemini models, enabling document-based context retrieval.
  • Create and manage file search stores directly in the PromptLayer UI
  • Upload documents to stores and associate them with prompts in the playground
  • Documents are automatically indexed for semantic search during conversations
  • Grounding metadata shows which documents were referenced in responses
OpenAI MCP (Model Context Protocol) Tool Support for OpenAI’s Model Context Protocol tools in prompt templates and playground.
  • Configure MCP servers and tools through the built-in tools dialog
  • Available for OpenAI models that support function calling
  • Tool responses appear inline in conversation history
User Attribution Tracking Track which team member created or modified resources across the platform.
  • Author information displayed for prompts, datasets, evaluations, and notifications
  • Filter resources by creator in the unified registry
  • “Open Original Session” button on run requests links back to the source playground session

Improvements

  • Added support for Claude Sonnet 4.5 on Amazon Bedrock
  • Added support for Gemini 3.1 Flash Lite model
  • Debounced playground input variable parsing to reduce API calls during typing
  • Fixed issue where deleted file stores could still be selected in the UI
  • Improved search indexing with deduplication to prevent duplicate results
  • Redesigned settings navigation with clearer organization and visual hierarchy
  • Enhanced vector store management with delete store capability
  • Improved file preview URLs for local storage backends with HMAC-signed streaming

March 03, 2026

Deployment 1

New Features

Anthropic Structured Output Support Added JSON Schema support for Anthropic models to enforce structured responses.
  • Configure response_format with JSON Schema in prompt templates for Claude models
  • Automatically converts to Anthropic’s output_config format
  • Also supported for Claude models running on AWS Bedrock
Organization Members Management Enhanced organization members page with improved filtering and detailed member views.
  • View all workspaces and roles for each organization member in a detailed side panel
  • Filter members by workspace, role, or search by name/email
  • Members can now remove themselves from organizations without owner permissions

Improvements

  • Fixed score slider to properly handle integer-only scores
  • Added workspace search by name in workspace listing
  • Improved autocomplete components with better keyboard navigation and multi-select support
  • Enhanced request display to show error_type and error_message fields when present
  • Added validation for error_type field in /track-request endpoint to match /log-request behavior
  • Fixed memory leak in scheduled job processing

March 01, 2026

Deployment 1

Improvements

  • Conversation simulator now surfaces errors from follow-up turns instead of silently ending conversations, making it easier to diagnose multi-turn evaluation failures
  • Request logs with warning status now display partial responses when available, providing visibility into requests that partially succeeded
  • Fixed display logic to correctly identify the final assistant response in multi-turn conversations, ensuring request context and actual output are properly distinguished
  • Reduced backend test parallelization to improve test stability and reliability

February 28, 2026

Deployment 1

New Features

Public API Request Payload Endpoint New /api/public/v2/request-payload endpoint allows you to retrieve complete request details including prompt blueprints, token usage, and latency metrics.
  • Returns full prompt blueprint structure for easy reproduction
  • Includes comprehensive metadata: provider, model, tokens, pricing, and timing
  • Supports API key authentication

Improvements

  • Improved Playground reliability on slow network connections by buffering early messages to prevent UI stalls
  • Enhanced error handling for WebSocket token refresh failures with better logging for troubleshooting
  • Fixed race condition in report cell generation that could cause false failures under high concurrency
  • Improved WebSocket connection stability by returning cached tokens when refresh attempts fail
  • Enhanced error reporting for messaging service failures with clearer error messages and categorization

February 27, 2026

Deployment 1

New Features

OpenAI Images API Support Full support for OpenAI’s image generation models including gpt-image-1, gpt-image-1-mini, gpt-image-1.5, dall-e-3, and dall-e-2.
  • Configure quality, size, background, output format, and moderation settings directly in the Playground
  • Generate multiple images in a single request with n parameter control
  • View generated images with revised prompts in dedicated accordion sections
Google Gemini Image Generation Added gemini-3.1-flash-image-preview model for AI-generated images via Google/Vertex AI.
  • Customize image size (0.5K to 4K) and aspect ratio (1:1, 16:9, 21:9, and more)
  • Includes standard Gemini safety settings and generation parameters
URL Context Tool for Google/VertexAI Web search and URL content retrieval now available for Google and Vertex AI models in the Playground.
  • Extract and analyze content from web pages during conversations
  • Matches existing functionality available for OpenAI models
Enhanced Custom Scoring System Refactored evaluation scoring with improved reliability and performance.
  • Automatically recalculates report scores when evaluation criteria are updated
  • Prevents score updates on incomplete evaluations

Improvements

  • Fixed WebSocket connection timing to establish only after authentication token is available
  • Increased message history buffer to 400 messages for improved chat continuity
  • Resolved dynamic resolution stack errors in evaluation workflows
  • Enhanced Playground sidebar layout with better widget spacing and control bar positioning
  • Improved clipboard handling for content copy operations in the editor
  • Fixed cost calculations for nano-banana-2 model
  • Streamlined prompt template retrieval logic for better reliability

February 26, 2026

Deployment 1

New Features

OpenAI Images API Support PromptLayer now supports OpenAI’s image generation models including gpt-image-1, gpt-image-1-mini, gpt-image-1.5, dall-e-3, and dall-e-2.
  • Track and log all image generation requests with full parameter support (quality, size, format, moderation)
  • View generated images directly in the request logs with revised prompt accordion
  • Monitor token-based pricing for new GPT image models
Google Gemini Tool Support Enhancements Extended tool support for Google and VertexAI models with additional capabilities.
  • Added URL context tool support for fetching and processing web content
  • Added code execution tool support for running code within model interactions
  • Preserved thinking blocks for extended reasoning visibility in responses
Improved Markdown Rendering Enhanced markdown display across the platform for better content readability.
  • Richer formatting support in chat messages and outputs
  • Improved code block rendering with syntax highlighting
  • Better handling of complex markdown structures in evaluations and logs

Improvements

  • Added human-readable status descriptions in the UI for better request monitoring
  • Fixed refresh button behavior in sidebar navigation for consistent state management
  • Improved error handling for team member invitations with clearer error messages
  • Enhanced clipboard support for copying content from rich text editors
  • Fixed prompt analytics page to correctly display evaluations without scores
  • Improved evaluation table columns to show more detailed metrics
  • Enhanced streaming performance for playground outputs with better state management