Hot on the heels of our Improved Call Data Visualizations and Reporting, the APImetrics user interface now has a fresh new look for 2017. Our improved interface makes our service even more user-friendly and straightforward to use than before. Over the past year, we’ve been focused on enhancing the core functionality of APImetrics that makes our product the industry-leading API quality monitoring solution, especially with our unique APImetrics Insights intelligent analytics suite and CASC Score so you can see at glance how well your APIs are performing. User interface upgrades We re-examined every aspect of our user interface and user [...]
Welcome to 2017 and with it, the first of a bunch of new reporting features. They're designed to tell you more about what's actually impacting your API and SaaS performance. First out of the gate – we've improved visualizations and comparison data of the steps involved in an API call. We've been breaking down the individual steps of the DNS lookup, handshake, processing and download. But we've now visualized that data along with gauges. It shows how the numbers stack up to your historical averages. SLOW reporting We also added functionality into the 'SLOW' alerts to let you know roughly [...]
The world is increasingly reliant on APIs to drive mobile apps, services and corporate operations. But how do you measure the quality of the APIs you use or depend on? APImetrics have been working on this problem for over 2 years, looking at the different factors that impact API and Cloud quality. And we came to realize that the classically used metrics don't really answer the question of actual quality of service. You need more than the server uptime or server-side response latency you're delivering or receiving from a supplier. To simplify this process, APImetrics have developed the CASC Score [...]
First Utility is a leading energy supplier in the United Kingdom. They were having issues with the quality of the APIs that were powering their mobile apps. The challenge they faced was simple. They had extensive testing and monitoring including API Gateway monitoring, Selenium based web test tools and Splunk based logs. But they were hitting performance issues. And they were taking significant time to resolve and identify. The company implemented APImetrics. And within days, First Utility was able to resolve problems that had been impacting users for weeks. Even better, they were able to stop intra-company finger pointing. “Rather [...]
70% of all API problems have no easy way to identify root cause In a survey of 20 of the leading corporate infrastructure APIs, we found that in over 70% of the API problems, there was no clear root cause in the cluster of poor performance. We used our Machine Learning system to learn the normal performance of each of the APIs. This included services from Docusign, Microsoft and Dropbox. We looked for periods where the performance degraded. We then clustered the events that seem to be linked or related (again using machine learning techniques). Then we looked [...]
It's come to our attention that we had a significant issue affecting data collection from our remote agents over the 4th of July weekend. We've traced the problem to our services bus that connects the remote agents to our data store. This has been rectified and we are taking steps to ensure this type of event can't happen again, unfortunately, it will have resulted in a weekend of lost data from the different collection points. Calls made from our default server will be unaffected. If you have any additional questions please don't hesitate to contact us.
We're in the process of rolling out some new analytics tools and we've been looking at some headline numbers. The first headline number is 83,000,000 - that's how many API calls we've run since we started APImetrics and, because we believe in learning from data, that's how many records we have in our database. The second number, which is what surprised us, of that number about 1,600,000 are out and out failures - that is the API returned, for some reason, a 5XX error. We have a much higher rate of 4XX errors, but they could be related to token [...]
An article including a number of things we've discovered over the last year and very handy for anybody who is starting to try and figure out what a Service Level Agreement (SLA) actually means for an API, a micro-service or, for that matter, the cloud. We're going to be expanding our feature set around SLA monitoring and we'll keep you informed of the status over the next few months. Enjoy this one, it's an excellent read from CIO.com.
API Time Travel We frequently check general performance data to look for 'odd' responses and we found an interesting one today which involved an API call on a test server which took -284,027ms, or just a little under 5 minutes. We assume that the host had to have a clock reset in the middle of making the API call, but it was an interesting results and one which has led to a small change on our side to error out such platform induced time issues in the future. Once again, this makes it clear to us that just [...]
As one of our feature enhancements we've improved the way our agents work, we are now consistently capturing key data on the actual API performance including: DNS Lookup time: i.e. 28355µs (28ms) Time to Connect: i.e 76106µs (76ms) Time for Handshake: i.e 0µs (0ms) Upload time: i.e. 48µs (0ms) Processing time: i.e. 120568µs (121ms) Download time: i.e. 2545µs (3ms) However, this improvement has raised two issues. Firstly, we have realized that we had some minor reporting issues with our old collection agents which means that some of the latencies we are recording were actually better than actually were being experienced. [...]