Yesterday, Ookla released the Net Index, a ranking of average broadband speeds by city, state, and country.
As background, Ookla is the company that powers the bandwidth tests at Speedtest.net and Pingtest.net. They also license their bandwidth testing engine to most of the major ISPs as well as the FCC, which used it along with MLabs to run bandwidth tests on broadband.gov.
This release of data is significant for a number of reasons.
Arguably the biggest reason is the sheer amount of data. While the FCC was able to collect about 100,000 bandwidth tests on their own, Ookla conducts more than a million tests a day, with more than 1.5 billion completed to date. So this data is an order of magnitude greater than anything else that's come out to date.
This data is also significant because it shows actual speeds being realized by users. Too much of our broadband speed data to date has been based on guesses about theoretical performance, with estimates generated through looking at maps and relying on network engineering principles, whereas this data comes from actual users on real-life broadband networks conducting bandwidth tests.
Ookla cites their bandwidth testing methodology as another significant strength of this dataset. "We right-size the payload," says Mike Apgar, Ookla's co-founder and managing partner. "In our engine on the server side, we have nine or ten different file sizes. What we do is a brief test that happens at the beginning that sends a single file that gives some idea of the connection we're talking about and whether we ought to use the 100MB payload or the 10MB."
Ookla also has the ability to use multiple threads where they can use more than one web server, domain name, and IP address to provide more avenues to make sure they're truly filling the pipe. Apgar shares the following analogy to explain the importance of doing this: "If you have a connection to your city's water, if you turn on a single faucet you're not measuring the system's full capacity. By turning on multiple faucets you can."
For these reasons, Ookla's bandwidth testing engine is able to accurately gauge the throughput of high-capacity broadband networks.
But that doesn't mean this data isn't without its limitations.
For example, in the ranking of cities by average broadband speed, Ookla only include those cities with at least 75,000 unique IP addresses. While it makes sense to do this to make sure you don't end up with smaller cities with skewed numbers, an unintended side effect is that it basically rules out cities with full fiber networks from these rankings since most fiber communities don't have 75,000 homes let alone 75,000 customers connected to fiber.
I mentioned this to Mike and he offered to pull data from smaller cities to help rectify this as he's a big supporter of fiber. And that's one of the great things about this Net Index is that the index data is available for free to everyone to use and analyze. They're also making the raw source data available to academics for research purposes. So we now have a resource that we can pull from for all sorts of reasons.
For example, Apgar shared that they have future plans to expand upon this, like creating a Value Index, which will collect and rank data about broadband networks' promised vs. actual speeds relative to the cost of service.
Another potential weakspot of this data is that it doesn't get into answering granular questions about where bandwidth bottlenecks might be occurring. While users can choose to test bandwidth between servers in different geographic locations to help determine where slowdowns might be occurring, this data doesn't reflect that.
A final limitation is one that's potentially unavoidable, namely that the data is skewed because it self selects its participants. Average joes likely aren't going out of their way to test their bandwidth; Ookla's primary users are techies. Techies tend to value bandwidth more and therefore their results will likely skew higher as they're more likely to sign up for higher capacity service.
But that being said, in talking it through with Apgar I came to realize that the full truth is a bit more nuanced than that.
One major trend is that with debates around broadband reaching the mainstream media that we have more of a public emphasis on bandwidth and broadband performance than ever, which is likely driving more people to become aware of bandwidth testing who might not have before.
It also dawned on me that all of those non-techies need techies to help them when their computers break, so that could be another way that non-techie homes are getting their bandwidth tested.
And Mike also shared with me that he's hearing from many people that installation techs for major broadband providers are using SpeedTest.net as the final step of setting a home up with broadband to have a way to easily show customers the kinds of speeds they're realizing. So in other words, there are a number of ways that these tests are reaching a more mainstream audience.
All in all this is an exciting development for broadband policy debates. More data should help us make better broadband policy decisions. And the Net Index represents one of the largest such datasets to date.
I'm going to be reviewing through their data and will report on what interesting tidbits I find.
But for now, if you're interested in starting to learn more for yourself, go check out the Net Index at www.netindex.com.