March 18, 2026
Deployment 1
New Features
GPT-5.4 Mini and Nano Model Support PromptLayer now supports OpenAI’s GPT-5.4 Mini and GPT-5.4 Nano models in thePlayground and API.
- Configure reasoning effort, verbosity, and response format options for both models
- Leverage lower-cost alternatives to GPT-5.4 for appropriate use cases
- Access prompt caching capabilities for improved performance
Improvements
- Images now display correctly in
Playgroundchat mode - Dataset columns reordered to show
promptlayer_urlbeforepromptfor easier request navigation - Enhanced model configuration options for GPT-5.4 series models
Deployment 2
New Features
Chat History Injection for Prompt Templates Prompt template blocks in evaluations can now inject chat history messages from a dataset column directly into chat prompts.- Enable chat history injection in the Advanced Settings section of the prompt template block configuration
- Select a source column containing message objects with role and content fields
- Messages are automatically appended to the end of your prompt template
- Supports both JSON and JSON5 formatted message lists for flexible data sources
March 17, 2026
Deployment 1
New Features
Public API for Dataset Rows New REST API endpoint to programmatically retrieve paginated rows fromDatasets, enabling integration with external tools and workflows.
- Access dataset rows via
/api/public/v2/datasets/{id}/rowswith support for search and pagination - Returns structured row data matching dataset column definitions
- Supports up to 100 rows per request with flexible filtering
- Access evaluation rows via
/api/public/v2/evaluations/{id}/rowsendpoint - Returns both dataset input variables and evaluation cell results in a unified format
- Enables automated analysis and reporting on evaluation performance
- Improved compatibility with diverse instrumentation libraries
- More reliable extraction of provider and model information from traces
Improvements
- Enhanced input variable detection in
Playgroundwith truncation and tooltip support for long variable names - Improved link handling in Wrangler AI for better navigation across all resource types
- Refined organization members table pagination for more consistent data loading
- Updated API key input styling across provider configuration pages for better visual consistency
- Improved deduplication logic for file annotations to prevent duplicate entries
Deployment 2
New Features
Deployment Usage Analytics Track token consumption and session activity across all your prompt deployments with new organization-level analytics.- View daily token usage broken down by individual deployments
- Monitor session counts per deployment over time
- Compare usage across public and private deployments
- Access historical usage data for capacity planning
- Toggle remix capability on or off for any deployment
- Users can fork and customize prompts while maintaining attribution
- Remixed versions are saved to the user’s own workspace
- Great for templates and starter prompts you want others to build upon
Improvements
- Enhanced deployment management UI with improved session visibility and controls
- Added batch execution support for shared prompt deployments
- Improved file upload handling and multipart processing for large media files
- Better dataset creation flow with prompt template selection from request history
- Enhanced permission checks and access controls across deployment endpoints
March 16, 2026
Deployment 1
New Features
Runtime Tool Variables in Prompt Templates Dynamic tool injection now supports variable substitution, enabling templates to generate tool definitions on-the-fly based on runtime context.- Tool schemas can include variables (e.g.,
{{user_id}},{{domain}}) that resolve during template execution - Supports nested variable resolution within tool parameters and descriptions
- Enables dynamic function calling patterns where tool availability adapts to request context
- Attach metadata, tags, or notes to specific assistant or user messages
- Track message-level performance metrics and quality assessments
Improvements
- Enhanced function/tool overview dialog displays complete schema details with improved formatting
- Improved citation modal rendering with better support for complex reference structures
- Streamlined Docker image build process reduces deployment time
- Better visual distinction between different tool types in the functions list view
- Optimized frontend bundle size through refined dependency management
March 15, 2026
Deployment 1
New Features
Public API Request Search Endpoint A new/api/public/v2/requests/search endpoint enables programmatic searching of request logs with structured filters.
- Search logs using the same filtering capabilities available in the dashboard
- Support for complex filter groups and structured queries
- Rate-limited to 10 requests per minute with up to 25 results per page
- Unlimited playground testing when using your own API credentials
- Cost control stays with your organization while removing artificial usage caps
Improvements
- Improved dataset column JSON parsing with better error handling for sparse or malformed data
- Enhanced workspace member management interface with clearer permission displays
- Fixed API key modal display to better communicate rate limits and usage policies
- Standardized public API endpoint structure (moved
GET /api/public/v2/request/<id>to/requests/<id>) - Added permission checks to dataset creation and editing endpoints to enforce role-based access control
March 12, 2026
Deployment 1
New Features
Hybrid Search for Registry Enhanced search across prompts, workflows, and datasets combining keyword matching with semantic understanding for more relevant results.- Search results now surface contextually similar items even when exact keywords don’t match
- Improved search ranking considers both text relevance and semantic meaning
- Background indexing keeps search up-to-date as you modify registry items
- Toggle scroll lock on/off to control whether new messages auto-scroll
- Manually scrolling up automatically enables scroll lock
- Scroll to bottom re-enables auto-scroll behavior
Improvements
- Fixed search suggestions displaying incorrect text values in autocomplete dropdowns
- Resolved f-string variable indexing issues when searching prompt templates
- Added
languagefield support for Google Code Execution tool responses - Improved citation display by preserving original model response annotations without deduplication
- Enhanced registry list and grid views with optimized virtualization for faster rendering of large item collections
Deployment 2
New Features
OTLP Prompt Resolution by ID and Label Enhanced OpenTelemetry trace ingestion now supports flexible prompt identification and version resolution.- Spans can reference prompts by
promptlayer.prompt.idin addition topromptlayer.prompt.name - Version resolution via
promptlayer.prompt.labelautomatically links traces to labeled prompt versions - Improved error handling when prompt identifiers are incomplete or not found in the workspace
/spans-bulk endpoint now intelligently handles duplicate span IDs to prevent data conflicts.
- Duplicate spans within the same batch are deduplicated before insertion (first occurrence wins)
- Duplicate spans across separate batches use upsert logic (last write wins)
- Ensures trace data remains consistent when the same span is reported multiple times
Improvements
- Added
flask embed_recently_used_promptscommand to backfill embeddings for prompts with recent traffic - Added
flask normalize_recently_used_promptscommand to backfill normalized content for recently-used prompt versions - Both commands support configurable look-back windows (default 365 days) and batch sizes for gradual processing
- Enhanced test coverage for duplicate span handling scenarios in bulk span creation
Deployment 3
New Features
Structured Search Advanced filtering interface for request logs with improved query building and autocomplete suggestions.- Build complex filters using fields, operators, and values with keyboard-driven autocomplete
- Support for nested metadata filtering with
key_equals,key_not_equals, andkey_containsoperators - Multi-value selection for tags, labels, and metadata fields with
inandnot_inoperators - Apply date range presets like “Last 5 minutes” or shorthand like “30d” for quick filtering
Improvements
- Enhanced date picker with single date selection mode and custom preset support
- Added
NOT_INoperator support for identifier, string, array, and nested key-value fields - Improved table components with better row click handling and empty state messages
- Added
IS_EMPTYandIS_NOT_EMPTYoperators for nested metadata filtering - Expanded operator support for input/output text fields to include
STARTS_WITHandENDS_WITH
Deployment 4
Improvements
- Fixed search filters not correctly matching boolean and numeric metadata values (e.g.,
false,true,42) - Resolved issue where changing search filters could trigger duplicate requests and cause stale results to display
- Improved nested field filtering to properly match metadata values regardless of type (string, boolean, or number)
March 11, 2026
Deployment 1
New Features
Nested Search Support for Outputs and Input Variables Advanced search now supports filtering by output fields and input variables, matching the existing metadata search capabilities.- Search for specific output values using
output:key=valuesyntax - Filter requests by input variable content with
input_variables:key=value - Use autocomplete suggestions for both output keys and input variable keys in the search bar
Improvements
- Improved snippet handling when creating prompt versions with overrides to ensure base references are used consistently
- Enhanced search suggestion performance for nested field queries (metadata, outputs, input variables)
- Input variables now preserve insertion order when rendering prompt templates
- Added structured logging context showing workspace and user IDs for better debugging and support
March 10, 2026
Deployment 1
New Features
Model Override Support in Evaluations Enhanced evaluation workflows now preserve model configuration when routing between prompt templates and agents.- Model override settings are now correctly passed through evaluation interfaces
- API type and model configuration IDs are properly maintained across workflow executions
Improvements
- Fixed tool call detection in search indexing to correctly identify assistant messages with tool calls
- Improved “Open Prompt” button functionality in image API evaluations to use correct routing
- Enhanced build efficiency by adding path guards to skip unnecessary backend image builds when only documentation or configuration files change
Deployment 2
New Features
Prompt Starring Users can now star important prompts for quick access and organization.- Star/unstar prompts directly from the prompt template page
- View list of users who have starred a prompt
- Filter and prioritize frequently-used prompts
- Build complex queries using field-specific filters (metadata, tags, models, etc.)
- Get autocomplete suggestions for search fields based on your workspace data
- Sort results by any field with flexible ascending/descending order
- View bash command execution results with syntax highlighting
- See code patches applied by AI agents in a readable format
- Improved visualization of tool use blocks in chat interfaces
Improvements
- Filter prompt templates by tags via the API using the
tagsparameter - Fixed “Open Prompt” button behavior in image-based evaluations to correctly navigate to prompt templates
- Resolved race condition in workflow output nodes that could cause incorrect status codes
- Added model override routing support for evaluation workflows
- Improved prompt template list performance with optimized tag indexing
Deployment 3
New Features
Multi-Prompt Search Filtering Advanced search now supports filtering across multiple prompts simultaneously and combining filters with logical operators.- Apply filters to multiple prompt templates at once for cross-prompt analysis
- Combine search conditions using AND/OR logic for more precise queries
- Filter suggestions now respect existing search criteria for faster query building
- Click “New item” buttons that appear in expanded empty folders
- Context-aware creation automatically places items in the correct folder
- Streamlined workflow for organizing prompts, datasets, and other resources
Improvements
- Search autocomplete suggestions now dynamically update based on active filters
- Added support for null/not-null operators in numeric field searches
- Enhanced folder navigation with visual indicators for active item context
- Improved metadata value suggestions with better handling of nested fields
- Optimized search performance for large workspaces with complex filter combinations
Deployment 4
New Features
Anthropic Text Editor Tool Support Added support for Anthropic’s text editor built-in tool, enabling AI assistants to view, create, and edit text files programmatically.- Available for both Anthropic and Vertex AI (Anthropic models) providers
- Supports commands like view, create, insert, and string-based replace operations
- Automatically handles text editor tool results in request logs and prompt templates
Improvements
- Enhanced subscription tracking with monthly contract value and contract duration fields for better enterprise billing management
- Fixed real-time event listener limits to prevent connection issues when multiple components subscribe to the same channel
- Improved tool choice handling to correctly map Text Editor tool names in API requests
Deployment 5
New Features
Plain Text Search in Structured Search You can now use plain text search alongside structured filters to find request logs more quickly.- Performs full-text search across request inputs and outputs while applying your structured filters
- Enables flexible searching when you need both keyword matching and precise filtering
- Search for specific tool names using the search bar with autocomplete suggestions
- Filter requests by tool names in structured search queries
- Helps track which tools are being used across your prompts and workflows
Improvements
- Search results now prioritize exact matches in request inputs and outputs when using plain text search
- Tool name suggestions appear in the search bar autocomplete for faster filtering
- Structured search queries support filtering by tool execution status and metadata
March 09, 2026
Deployment 1
New Features
OpenTelemetry Trace Ingestion Enhancements Support for modern OpenTelemetry semantic conventions and improved compatibility with observability libraries.- Added support for gzip-compressed OTLP trace payloads to reduce network overhead
- Added support for newer
gen_ai.input.messagesandgen_ai.output.messagesJSON format used by Ruby and other emerging instrumentations - Improved parsing of
gen_ai.system_instructionsto properly handle system prompts from different providers - Added automatic upsert logic for duplicate span IDs to ensure trace completeness when spans are sent multiple times
- Template validation now gracefully skips tool/function messages that contain JSON responses rather than user-authored templates
- Prevents false template rendering errors when JSON braces in tool responses are mistaken for template syntax
- Preserves support for legitimate template variables in few-shot tool examples
Improvements
- Added
playground_session_idto request log bulk endpoint responses for better session tracking - Improved JSON variable parsing to optimistically parse all string values, matching frontend batch-mode behavior
- Enhanced OTLP function name inference to support more provider-specific operation types (embeddings, text completion, content generation)
- Fixed provider family detection for Anthropic and Google AI models in OpenTelemetry traces
- Improved error handling for malformed Content-Type headers in trace ingestion
March 07, 2026
Deployment 1
New Features
Anthropic Code Execution Tool Support Added support for Anthropic’s native code execution tool capability, enabling AI models to write and execute Python code during conversations.- Models can now generate and run code snippets directly within chat sessions
- Code execution results are displayed inline with conversation history
- Supports dynamic data analysis and computation workflows
- Search and filter traces by any custom metadata key stored in span attributes
- Autocomplete suggestions help discover available metadata keys across your workspace
- Filter results update in real-time as you type
- Enables automation workflows where models can interact with system commands
- Integrates with OpenAI’s native tool calling infrastructure
Improvements
- Improved playground session initialization to correctly handle tool and function definitions when opening from request logs
- Enhanced request log input variable extraction to include tool/function data for better context when replaying requests
- Streamlined “Open in Playground” workflow to preserve all tool configurations from original requests
- Fixed trace metadata button display issues in the span details view
- Normalized message content format to consistently use content blocks across chat interfaces
- Improved Vite build configuration for better development server performance
March 05, 2026
Deployment 1
New Features
OpenTelemetry Trace Ingestion Native support for industry-standard OpenTelemetry Protocol (OTLP) trace ingestion, enabling seamless integration with existing observability tooling.- Ingest traces via standard OTLP/HTTP endpoint at
/v1/traces - Automatic extraction of GenAI semantic conventions for OpenAI and Anthropic providers
- Convert OTLP spans into PromptLayer request logs with proper error mapping and metadata preservation
- Import and replay conversations with parallel tool calls from request logs
- Maintain correct message ordering when tools are invoked across conversation turns
- Proper hydration of chat history with multiple tool responses per assistant turn
- Reset and re-seed chat from any logged request with one click
- Automatically diff request messages against current template to extract conversation context
- Per-variable-set chat history support for testing multiple scenarios simultaneously
Improvements
- Fixed playground chat crashes when trace metadata contains non-string values during URL sharing
- Resolved 500 errors when reading prompts that use legacy LangChain message format
- Fixed “No response” display issue for template render errors in request logs
- Improved image evaluation algorithm accuracy for visual content comparison
- Enhanced workspace member invitation dialog with better field validation
- Fixed chat message ordering when importing request logs with tool calls
March 04, 2026
Deployment 1
New Features
Google File Search Tool Support Native integration with Google’s File Search tool for Gemini models, enabling document-based context retrieval.- Create and manage file search stores directly in the PromptLayer UI
- Upload documents to stores and associate them with prompts in the playground
- Documents are automatically indexed for semantic search during conversations
- Grounding metadata shows which documents were referenced in responses
- Configure MCP servers and tools through the built-in tools dialog
- Available for OpenAI models that support function calling
- Tool responses appear inline in conversation history
- Author information displayed for prompts, datasets, evaluations, and notifications
- Filter resources by creator in the unified registry
- “Open Original Session” button on run requests links back to the source playground session
Improvements
- Added support for Claude Sonnet 4.5 on Amazon Bedrock
- Added support for Gemini 3.1 Flash Lite model
- Debounced playground input variable parsing to reduce API calls during typing
- Fixed issue where deleted file stores could still be selected in the UI
- Improved search indexing with deduplication to prevent duplicate results
- Redesigned settings navigation with clearer organization and visual hierarchy
- Enhanced vector store management with delete store capability
- Improved file preview URLs for local storage backends with HMAC-signed streaming
March 03, 2026
Deployment 1
New Features
Anthropic Structured Output Support Added JSON Schema support for Anthropic models to enforce structured responses.- Configure
response_formatwith JSON Schema in prompt templates for Claude models - Automatically converts to Anthropic’s
output_configformat - Also supported for Claude models running on AWS Bedrock
- View all workspaces and roles for each organization member in a detailed side panel
- Filter members by workspace, role, or search by name/email
- Members can now remove themselves from organizations without owner permissions
Improvements
- Fixed score slider to properly handle integer-only scores
- Added workspace search by name in workspace listing
- Improved autocomplete components with better keyboard navigation and multi-select support
- Enhanced request display to show
error_typeanderror_messagefields when present - Added validation for
error_typefield in/track-requestendpoint to match/log-requestbehavior - Fixed memory leak in scheduled job processing
March 01, 2026
Deployment 1
Improvements
- Conversation simulator now surfaces errors from follow-up turns instead of silently ending conversations, making it easier to diagnose multi-turn evaluation failures
- Request logs with warning status now display partial responses when available, providing visibility into requests that partially succeeded
- Fixed display logic to correctly identify the final assistant response in multi-turn conversations, ensuring request context and actual output are properly distinguished
- Reduced backend test parallelization to improve test stability and reliability
February 28, 2026
Deployment 1
New Features
Public API Request Payload Endpoint New/api/public/v2/request-payload endpoint allows you to retrieve complete request details including prompt blueprints, token usage, and latency metrics.
- Returns full prompt blueprint structure for easy reproduction
- Includes comprehensive metadata: provider, model, tokens, pricing, and timing
- Supports API key authentication
Improvements
- Improved Playground reliability on slow network connections by buffering early messages to prevent UI stalls
- Enhanced error handling for WebSocket token refresh failures with better logging for troubleshooting
- Fixed race condition in report cell generation that could cause false failures under high concurrency
- Improved WebSocket connection stability by returning cached tokens when refresh attempts fail
- Enhanced error reporting for messaging service failures with clearer error messages and categorization
February 27, 2026
Deployment 1
New Features
OpenAI Images API Support Full support for OpenAI’s image generation models includinggpt-image-1, gpt-image-1-mini, gpt-image-1.5, dall-e-3, and dall-e-2.
- Configure quality, size, background, output format, and moderation settings directly in the Playground
- Generate multiple images in a single request with
nparameter control - View generated images with revised prompts in dedicated accordion sections
gemini-3.1-flash-image-preview model for AI-generated images via Google/Vertex AI.
- Customize image size (0.5K to 4K) and aspect ratio (1:1, 16:9, 21:9, and more)
- Includes standard Gemini safety settings and generation parameters
- Extract and analyze content from web pages during conversations
- Matches existing functionality available for OpenAI models
- Automatically recalculates report scores when evaluation criteria are updated
- Prevents score updates on incomplete evaluations
Improvements
- Fixed WebSocket connection timing to establish only after authentication token is available
- Increased message history buffer to 400 messages for improved chat continuity
- Resolved dynamic resolution stack errors in evaluation workflows
- Enhanced Playground sidebar layout with better widget spacing and control bar positioning
- Improved clipboard handling for content copy operations in the editor
- Fixed cost calculations for
nano-banana-2model - Streamlined prompt template retrieval logic for better reliability
February 26, 2026
Deployment 1
New Features
OpenAI Images API Support PromptLayer now supports OpenAI’s image generation models includinggpt-image-1, gpt-image-1-mini, gpt-image-1.5, dall-e-3, and dall-e-2.
- Track and log all image generation requests with full parameter support (quality, size, format, moderation)
- View generated images directly in the request logs with revised prompt accordion
- Monitor token-based pricing for new GPT image models
- Added URL context tool support for fetching and processing web content
- Added code execution tool support for running code within model interactions
- Preserved thinking blocks for extended reasoning visibility in responses
- Richer formatting support in chat messages and outputs
- Improved code block rendering with syntax highlighting
- Better handling of complex markdown structures in evaluations and logs
Improvements
- Added human-readable status descriptions in the UI for better request monitoring
- Fixed refresh button behavior in sidebar navigation for consistent state management
- Improved error handling for team member invitations with clearer error messages
- Enhanced clipboard support for copying content from rich text editors
- Fixed prompt analytics page to correctly display evaluations without scores
- Improved evaluation table columns to show more detailed metrics
- Enhanced streaming performance for playground outputs with better state management

