Current Rate Limits
Rate limits protect against abuse and ensure system stability.| Endpoint | Limit | Identifier | Window |
|---|---|---|---|
/api/v1/* (All API endpoints) | 1,000 requests | Per API key | Per hour |
/api/health (Health check) | 60 requests | Per IP address | Per minute |
How It Works
- API Endpoints (
/api/v1/*): Rate limited by API key after authentication - Health Check (
/api/health): Rate limited by IP address before any processing - Algorithm: Sliding window (distributed evenly across the time period)
Rate Limit Headers
Every API response includes rate limit information in the headers:| Header | Description |
|---|---|
X-RateLimit-Limit | Maximum requests allowed in the window |
X-RateLimit-Remaining | Requests remaining in current window |
X-RateLimit-Reset | Unix timestamp when the limit resets |
Rate Limit Exceeded Response
When you exceed the rate limit, you’ll receive a429 status code:
Retry-After header with seconds until you can retry:
Best Practices
1. Monitor Rate Limit Headers
JavaScript
2. Implement Exponential Backoff
Python
3. Respect Rate Limit Reset Time
Python
4. Batch Requests Efficiently
Instead of making 1000 individual requests, use bulk endpoints:5. Distribute Load Over Time
Python
Advanced Rate Limiting Strategies
Request Queue with Rate Limiting
JavaScript
Need Higher Limits?
If your use case requires higher rate limits:1
Assess Your Needs
Document your expected request volume and use case
2
Contact Us
Email info@acrely.ai with your requirements
3
Review Options
We’ll work with you to find the right limits for your needs
4
Upgrade
Higher limits may be available on enterprise plans
Common Scenarios
High-Volume Lead Import
If importing thousands of leads:- Use bulk endpoint (
POST /leads/bulk) - 1000 leads per request - Batch your imports - 10 bulk requests per minute = 10,000 leads/min
- Run during off-hours if possible
- Monitor progress and handle errors gracefully
Real-Time Integrations
For webhook-triggered lead creation:- Queue incoming webhooks instead of immediate API calls
- Process queue at controlled rate (90-95 requests/min)
- Leave headroom for retries and other operations
- Implement circuit breakers for cascading failures
Analytics Dashboards
For dashboards that query the API:- Cache responses for at least 1 minute
- Aggregate data on the backend
- Use webhooks for real-time updates instead of polling
- Implement request debouncing for user interactions
Monitoring Rate Limits
Track your usage to avoid surprises:Python