



Have any questions? We’re here to help You
API tokens must be stored securely and should never be exposed on the client side or in public repositories. Store tokens in secure environment variables or dedicated secrets management systems like AWS Secrets Manager, HashiCorp Vault, or Azure Key Vault. Never hardcode tokens in application code or commit them to version control. Implement proper access controls so only authorized services can access stored tokens. For production environments, use separate tokens from development/testing environments. Rotate tokens periodically and immediately revoke tokens if you suspect they've been compromised. Makini tokens provide access to customer data, so treat them with the same security standards you'd apply to database credentials.
Yes, Makini provides extensive customization options for field mappings. Through the connection settings interface, you can view how each system's fields map to Makini's unified model. You can remap fields, add custom field mappings, or create entirely new custom fields that will appear in API responses. These customizations are connection-specific, allowing different mapping configurations for different customers. Mapping changes take effect immediately without requiring code changes. For standardized workflows, default mappings typically provide sufficient coverage. Custom mappings are most useful when integrating with heavily customized systems or when you need fields beyond the standard unified model.
For bulk operations, we recommend batch processing with appropriate rate limiting and error handling. Makini Flows provides built-in batch processing capabilities with configurable batch sizes, delays between batches, and error handling. For API-based bulk operations, implement pagination when retrieving large datasets—our API returns results in pages with continuation tokens for fetching subsequent pages. When writing large volumes of data, break operations into smaller batches (typically 50-100 records per batch) with delays between batches to avoid overwhelming the target system. Implement comprehensive error logging to identify which specific records fail in a batch. For very large operations (thousands of records), consider asynchronous processing patterns where you queue operations and process them in the background.
Our standard SLA targets 99.9% uptime for cloud deployments, which translates to less than 9 hours of downtime per year. For enterprise customers with critical integration requirements, we offer enhanced SLAs up to 99.99% through multi-region redundancy and dedicated infrastructure. SLAs cover the Makini platform itself—availability of connected third-party systems is outside our control, though we monitor their health and alert you to issues. For self-hosted deployments, uptime depends on your infrastructure configuration, and we provide architecture guidance to help you achieve your availability targets. We maintain a public status page showing real-time system health and incident history.
