# Python SDK from filecatalyst import Profile, Orchestrator
destination: type: "s3" bucket: "backup-bucket" path: "database/ date /" lifecycle: "delete after 90 days" filecatalyst profiles
filecatalyst profile test --profile new_config.yaml --dry-run # Python SDK from filecatalyst import Profile, Orchestrator
# CLI examples filecatalyst profile create --from-template backup \ --source s3://my-bucket/ \ --dest /backup/ \ --schedule "0 3 * * *" filecatalyst profile apply --profile marketing_sync --override bandwidth=200Mbps filecatalyst profiles
transfer: adaptive_bandwidth: true min_bandwidth: "20Mbps" max_bandwidth: "200Mbps" compression: "zstd" encryption: "AES-256-GCM" parallel_chunks: 8 verify_checksum: "SHA-256"
source: type: "postgresql" connection: "pg://backup-user@primary/db" dump_before_transfer: true
Profile: "Video_Transcode_Sync" ├── Transfers (24h): 1,247 ├── Total Data: 3.2 TB ├── Avg Speed: 245 Mbps ├── Success Rate: 99.87% ├── Bottleneck: Disk I/O on source (42% of delay) └── Recommendations: • Enable local caching on source • Increase thread count from 4 to 8 Advanced rule engine for file selection: