noob here using Arduino Wifi (Adafruit CC3000) to send data. Or try. I have read and understood about rate limits and a limit on number of data items uploaded per connection. But I searched and cannot find what the numbers are for those limits. One per minute? I have deduced by testing that the number of data items is two, and tried to send my 6 data points in three consecutive calls of two. That crashes and burns with assorted errors, so I suspect I'm hitting the rate limit.
P.S. Two datapoints per minute work, so API key, etc is not an issue.
Can anyone tell me what these limits are?
Thanks very much for your time.
If you have a developer account (free), you have a max of 25 requests per minute. This includes if you have an app on the front end pulling/sampling the data. It includes all the GET PUT and socket connections as well.
Src: http://forums.electricimp.com/discussion/2108/max-frequency-of-updates-to-xively-wo-getting-booted/p1
Related
I recently subscribed on HERE Api for Freemium account, cause as far as I understood, I could make 250k requests per month and only would be charged after that, recently I'm working on Waypoints Sequence, and suddenly during a dev test, I got the error "Daily limit of 10 requests has been reached".
I thought that the limit was 250k requests per month, but then I read that the user can only make 10 requests per day by App ID, but makes no sense when the month limit is significantly higher.
Out of Freemium Account there's only the Pro/Premier Plans, which are actually very expensive for me right now.
Does anyone have a workaround for this? Or have another "Waypoints Sequence" tool to recomend?
Thank you all in advance
Currently, for Freemium plans only, the following limits apply:
Custom Locations: you may upload a maximum of 3 layers with a maximum of 100 polygons or polylines.
Waypoints Sequence: up to 10 requests per day.
Advanced Data Sets: up to 100 requests per hour and 1,000 per day.
Please contact our Sales team (Contact Sales) if you want to raise your limits. Thank you.
I have been searching for a while, but have not found any information on whether there are limits on calls to Amazon Rekognition service.
Does anyone know the numbers, or any source where I can look?
What I want to know is if they limit the number of calls allowed per minute or second. I'm looking for information on the the paid (not free) service tier.
You can find documentation on service limits for Rekognition at https://docs.aws.amazon.com/general/latest/gr/aws_service_limits.html#limits_rekognition.
The limits vary depending on which API you are accessing and from which region. Ex. DetectLabels has a limit of 5 transactions per second in US East (Ohio). You may request a limit increase on these numbers.
Here are the limits they have, at no time says anything about maximum number of requests per second as in other services
For example in amazon polly if it says the limits it has per second:
Any 2 transactions per second (tps) from these operations(DeleteLexicon, PutLexicon, GetLexicon, ListLexicons) combined.
Maximum allowed burst of 4 tps.
That's why I believe that it has no operations limit per minute, otherwise I would put it as they do in other services. I am using it to authenticate in the company where I am, that we are 23 and at the moment it has not given me any problems of operations per second.
References:
http://docs.aws.amazon.com/rekognition/latest/dg/limits.html
http://docs.aws.amazon.com/polly/latest/dg/limits.html
I have routines that synchronize Class/Roster information between an SIS and Google Classroom. Everything has been running smoothly until very recently (11/1/2016). Now we're seeing the following message in all of our log files for routines that handle Classroom syncs.
Insufficient tokens for quota group and limit 'DefaultGroupUSER-100s' of service 'classroom.googleapis.com', using the limit by ID...
We perform batch requests whenever possible and these errors are showing up in individual batch "part" responses. The fact that these errors suddenly started showing up for ALL of our Classroom routines makes me think that something changed on the Google end of things.
I've been playing around with the throttling on our end by changing both the number of requests that we send in each batch (docs say that you can send 1000 per batch) as well as the total number of requests/batches that we're sending per 100 seconds (docs say you can send 50/s/client and also 5/s/user). Interestingly, the quotas indicated in the development console display slightly different but I assume they are to be interpreted in conjunction with one another.
I've throttled things down to the point where we're not even getting close to 5 requests per second and I'm still getting these errors back from the server.
Can someone provide some suggestions or solutions. Has anyone experienced this lately?
Let me know if I any additional information is needed.
I want to monitor some live scores on soccer matches. I have 2 ways to do this:
official api from the website(free)
parse websites source code myself and get data from it( need to do it every second)
What is the difference? Is calling API faster?
This can depend on quite a lot external to this specific scenario, but given the context, yes the API's would much faster. The difference is in what data is being sent/received/parsed.
In either scenario you'd need some timer to tick and parse the results (website or API) so there's no performance difference in the "wait code", but the big difference will be in the data itself that is parsed. When you call the API, chances are more likely that you will send a specific parameter or call a specific function that indicates what you're looking for, pseudo-code example:
SoccerSiteApi.GetValue(SCORE, team1, team2);
Or
SoccerSiteApi.GetCurrentScores(team1, team2);
By calling the API, you are only sending and receiving a few hundred bytes (or more depending on data) and getting back exactly what you want, that is, you don't need to parse the scores out of the values sent back since they are the scores, so no processing time is spent doing anything additional with the data itself.
If, however, you were to parse the entire web site, you would need to make an HTTP GET request (and all that entails) to get the entire page (which could be a couple hundred KB or MB depending on content) and then spend processing time extracting the exact data you were looking for, and then doing this every second.
So the biggest difference is amount of data and time spent processing it.
Hope that can help
I am making an application based on Google maps API. This requires requesting for distance between two cities. Now I want distances between many cities.
So should I use "for loop" and make many requests separately or should I send all the cities names in one link. Which one will work faster? And which one will be better?
For sure you should avoid sending multiple requests, because each roundtrip to a server takes time.
However when you are grouping many requests this can also take a long time (both to send, and to process on the server), and affect the user experience (long waiting time).
In your case I suspect that the "for loop" will not load to a lot of data, and server side processing will also not be too heavy, so sending a grouped single request should be the way to go.
You can use the "DirectionService" sevices ,which is providing by Google i.e "api3".
You can find the distance between the Many cities ,it takes one origin point ,destination point and 8 way points (total 10 places) for one request and it provides a JSON file
in return ,which contains all the information (distance in KM,value meters,city names and lot more) .Please check this link, https://developers.google.com/maps/documentation/javascript/directions . i hope this answer will meet your requirement,otherwise don't mind.