Wednesday, 28 May 2014

Do we need the Delphic Oracle to make sensible telecoms investments?

The Delphic Oracle was the leading seer of ancient Greece. This was reliably established by King Croesus (as in 'rich as'), who had his messengers ask a sample of seers at a pre-agreed time what he was doing at that very moment. The Delphic Oracle correctly said he was making lamb and tortoise stew.

However, the Oracle's statements were rarely this unambiguous -  they tended to be a bit more, well, Delphic. For instance, asked about a prospective military expedition, she replied: "You will go you will return never in war will you perish" - place punctuation at your own risk.

It is sometimes argued that, absent reliable seers, we have to invest in superfast broadband because of the unknown unknowns - the applications that are surely coming, but which we just lack the foresight to predict today.

There are many problems with this argument, but one of them is that its proponents tend to underestimate our foresight. Here for instance the view of my old friends the FTTH Council on how little we knew in 2000 about the drivers of demand for today's broadband:


Just how accurate is it that these things were unforeseen in 2000?

Videoconferencing with Skype
Skype wasn't founded until 2003, but video calls over the internet have been long discussed - at least since 1995, when Stewart Loken of the Berkeley Lab said "internet videoconferencing is about to become commonplace". Obviously to know what bandwidth you might need, you don't need to know the name of the company that's going to be most successful, you just need to know what the application is, so the fact that Skype didn't exist in 2000 is neither here nor there.

HD-TVs with 42" and more in 3D
The first formal HDTV research programme began in 1970. Consumer sets went on sale in the US in 1998, and some of them had 55" displays. 3D TV was trialled as early as 1994.

Facebook
We certainly didn't know about Facebook in 2000, it wasn't founded until 2004. (Though you can discuss with the Winklevoss twins exactly when it was conceived). However social media had been around for a long time - GeoCities, an early example, was founded in 1994. And again, we don't need to know the name of the provider to know the necessary bandwidth. Facebook uses text, pictures and a bit of video - all well understood as internet media in 2000.

Online shops
Amazon was founded in 1994. 'nuff said.

Google
By 2000 Google was already available in 10 languages (and had hired its first chef a year prior)

Digital Photography
The first consumer digital cameras were released in 1990 (the same year as Photoshop). Webshots, founded in 1999, was one of the first web-based photo sharing sites, but consumers had been uploading photos to BBSs (not always savoury ones) for some years before that.

iPad and Smartphones
The Palm VII, one of the first PDAs with wireless capability for internet access, shipped in 1999. I'll give the FTTH Council the iPad, which wasn't widely anticipated. Of course, it doesn't need particularly high bandwidth (though it has driven more traffic by extending hours of internet use in the home).


So, of the Council's seven things "we did not know in 2000" it turns out we did know 6½ of them. Their view of our ignorance is a bit ... ignorant.

The vast majority of things we do with the internet today were in fact anticipated in 2000, at least in broad brush strokes. That's why it's particularly problematic for FTTH fans that there are (in their own words) 'no really compelling applications yet' for FTTH. In 2000 we knew (roughly) what we would do with broadband speeds. In 2014 we have no real idea what we might do with superfast.

Thursday, 15 May 2014

The killer app for FTTH? Piracy

Fibre enthusiasts like to point out that those with very high speed connections have higher usage of the internet. This is certainly true, though whether this is because heavier users trade up to superfast broadband, or because superfast actually changes behaviour is a vexed question.

However, a more intriguing issue is what makes up that extra usage. Markets like Hong Kong and Korea have very fast broadband, and high per-line traffic. When you ask insiders in those markets what all that traffic is, they often look a bit shifty and then whisper 'piracy'. Until today I hadn't seen any hard data to back up this assertion.

However, I've just come across a report from COMBO, an EU funded project looking at fixed-mobile convergence. (It has blue-chip participants - France Telecom, Deutsche Telekom, Telefonica, Alcatel Lucent, Ericsson and  several others). On page 104 of the report is the following data for traffic of FTTH and ADSL customers of France Telecom in October 2013 :

Upstream and downstream traffic combined.

As you can see, FTTH users have roughly double the usage of ADSL customers. However, 58% of the additional traffic is down to P2P. Peer-to-peer is a protocol with legitimate uses, but it's very largely used for piracy. It's what enables Bittorrent for example (and is beloved of the copyright-sceptics of The Pirate Bay).

FTTH customers have 4.2x the P2P traffic of ADSL customers, mostly because they upload 7.5x more than ADSL customers. Indeed, for every pirate-byte they download, FTTH users are uploading 2.6. In the words of the report, "this allows us to conclude that some FTTH customers are becoming P2P servers".

Superfast advocates claim significant externalities (societal and economic impacts) for faster broadband. But externalities can be negative as well as positive. See the rigging on the rather fine pirate ship above? Fibre-optic cable, every strand.

Friday, 2 May 2014

Why websites are like turnips

We are sometimes told to eat locally - to consume food that is well suited to our domestic climate and is grown nearby. (For those of us in the UK, this implies a diet heavy in turnips, one of our relatively few native vegetables). By eating locally, we can reduce food-miles and our carbon foot print.

While perhaps less threatening to the the planet, bit-miles can be damaging too, harming your surfing experience.

In my previous post I noted that increasing the speed of your home broadband connection doesn't necessarily lead to an improved experience. One scenario where this can happen is when you are accessing a remote service, such as a website hosted in a remote location. While the transfer of data between your computer and the server may be constrained by your access bandwidth, latency (the time taken for a packet to travel between the two) can make an enormous difference. Moreover, the more hops the packets are taking between your computer and the server, the more opportunities for them to hit congestion somewhere along the way.

It's easy to test the impact of this. Speedtest.net allows you to run a test of transfer rates between your computer and one of their servers. By default, they automatically select a server near you, reducing the likelihood that transfer rates are constrained by factors other than your own access link. (There's a lot of subtlety in speed tests though - see this for a detailed discussion of how they work and their limits).

Here's the result of such a test from my computer:



This test used Speedtest's defaults, and it has selected a server in Maidenhead, about 50 miles away. As a result, my ping (round trip latency) is 31 milliseconds. In terms of bandwidth, I'm getting 18 Mbps, which sounds about right - I'm on DSL, relatively close to the exchange.

Now here's a test where I've overridden the defaults, and used a server in LA:


Ping has quintupled to 162 ms (though before we get too outraged, this still means that the packets were averaging 1/3 the speed of light, given a 10,800 mile round trip). The speed of connection has dropped to 11.4 Mbps. This certainly suggests that the bandwidth of the last mile - the access link to my house - is not the key constraint, since we know that's capable of at least 18 Mbps.

Of course, there are places more remote from my UK home than LA. Here's what happens when we run the test on a Sydney server:


Ping has doubled again, and the effective bandwidth has dropped to 8 Mbps. By visiting a remote server, I've 'lost' 10 Mbps of the capability of my line. Put another way, I'd likely get just as good a performance from the Sydney server if I had an 8 Mbps line as opposed to my actual 18 Mbps line.

While one could in theory abandon tasty vegetables and only eat turnips, it is a bit harder to only source local websites. If I'm desperate for the results of the Brisbane cockroach races, then substituting a visit to a local website about snail racing in Norfolk just won't do. (Though you've got to love a sport where a 6 year old can be the world champion). There's always going to be a portion of our internet use where improving our last mile bandwidth won't make any difference at all.