Filecatalyst Profiles May 2026
source: type: "postgresql" connection: "pg://backup-user@primary/db" dump_before_transfer: true
notifications: on_success: "slack:#backups - Success: size transferred" on_failure: - "pagerduty:Database Backup Failed" - "email:db-admin@company.com" filecatalyst profiles
destination: type: "s3" bucket: "backup-bucket" path: "database/ date /" lifecycle: "delete after 90 days" filecatalyst profiles
transfer: adaptive_bandwidth: true min_bandwidth: "20Mbps" max_bandwidth: "200Mbps" compression: "zstd" encryption: "AES-256-GCM" parallel_chunks: 8 verify_checksum: "SHA-256" filecatalyst profiles
# Python SDK from filecatalyst import Profile, Orchestrator
Profile: "Video_Transcode_Sync" ├── Transfers (24h): 1,247 ├── Total Data: 3.2 TB ├── Avg Speed: 245 Mbps ├── Success Rate: 99.87% ├── Bottleneck: Disk I/O on source (42% of delay) └── Recommendations: • Enable local caching on source • Increase thread count from 4 to 8 Advanced rule engine for file selection:

