




Have any questions? We’re here to help You
Industrial systems are often heavily customized, and Makini is built to handle this. For reading data, Makini can access virtually any field or custom table in connected systems. Through the connection settings interface, you can specify custom fields, tables, or entities to include in API responses. These show up alongside standard fields in the unified model. For custom objects not in our default model, you can request them through the interface and they'll be available immediately. For writing data, customization support varies by system but covers most common scenarios. During implementation, we work with you to identify required customizations and ensure they're properly configured before going live.
Makini provides webhook testing tools in the dashboard where you can trigger test webhook deliveries to verify your endpoint configuration. Test webhooks use sample payloads matching actual event structures. Verify your endpoint receives the webhook, validates the signature correctly, and responds with a 200 status code within 10 seconds. Test webhook retries by having your endpoint return error codes or timeout, then verify Makini retries as expected. Test duplicate handling by processing the same webhook multiple times. For local development, use tools like ngrok to expose your local endpoint for webhook testing. The webhook logs in the Makini dashboard show delivery attempts, response codes, and timing, helping debug delivery issues.
For bulk operations, we recommend batch processing with appropriate rate limiting and error handling. Makini Flows provides built-in batch processing capabilities with configurable batch sizes, delays between batches, and error handling. For API-based bulk operations, implement pagination when retrieving large datasets—our API returns results in pages with continuation tokens for fetching subsequent pages. When writing large volumes of data, break operations into smaller batches (typically 50-100 records per batch) with delays between batches to avoid overwhelming the target system. Implement comprehensive error logging to identify which specific records fail in a batch. For very large operations (thousands of records), consider asynchronous processing patterns where you queue operations and process them in the background.
Makini provides several performance monitoring capabilities. API responses include timing information showing request processing time. The dashboard includes performance metrics showing average response times, throughput, and error rates over time. You can set up alerts for performance degradation or error rate increases. Each request generates a unique request ID that enables detailed performance analysis. For workflow-based integrations, execution logs show per-step timing, helping identify bottlenecks. We recommend implementing client-side monitoring to track end-to-end latency including network time. Monitor trends over time rather than individual requests—occasional slow requests are normal, but sustained increases may indicate issues requiring investigation.
