← prompt library
testing

Multi-Platform AI Agent Deployment Validator

Tests AI agents across messaging platforms like iMessage, WhatsApp, and Telegram for consistent behaviour and performance.

agent-testing cross-platform messaging deployment
prompt
# Multi-Platform AI Agent Deployment Validator

You are an expert in testing AI agents across messaging platforms. Your task is to create comprehensive test suites for agents that deploy to multiple chat platforms simultaneously.

## Agent Details
**Agent Name:** [agent name]
**Target Platforms:** [list platforms: iMessage, WhatsApp, Telegram, Discord, etc.]
**Core Functionality:** [describe what the agent does]

## Agent Configuration
```
[paste agent configuration/manifest here]
```

## Current Test Coverage
[describe existing tests or paste current test files]

## Analysis Required

### 1. Platform-Specific Constraints Analysis
- Message length limits per platform
- Media handling differences
- Rate limiting variations
- Authentication flows
- Rich message format support

### 2. Cross-Platform Consistency Tests
Create test cases that verify:
- Identical responses to same inputs across platforms
- Graceful degradation when platform features differ
- Consistent conversation state management
- Error handling uniformity

### 3. Platform Integration Edge Cases
Generate tests for:
- Network connectivity issues per platform
- Platform-specific API failures
- Message delivery confirmation handling
- User blocking/unblocking scenarios
- Group chat vs direct message behaviour

### 4. Performance Benchmarks
Define performance tests for:
- Response latency per platform
- Memory usage during multi-platform operation
- Concurrent user handling across platforms
- Message throughput under load

### 5. Security Validation
Test for:
- Data leakage between platform conversations
- Proper user context isolation
- Secure credential handling per platform
- Message encryption compatibility

## Deliverables
1. **Test Suite Structure** - Organised test files with clear naming
2. **Mock Platform Responses** - Simulated platform API responses
3. **Automated Test Scripts** - Runnable test commands
4. **Performance Baselines** - Expected metrics per platform
5. **Failure Scenarios** - What to expect when things break

Provide specific, runnable test code with clear assertions and expected outcomes.

Use this when deploying AI agents across multiple messaging platforms to ensure consistent behaviour and catch platform-specific issues early. Works with Claude, GPT-4, and Gemini to generate comprehensive test suites that cover the unique constraints of each messaging platform.