In January 2015, I was invited to speak at a bank’s
Corporate day. The audience comprised
analysts from the Finance industry, drawn from various firms. Their normal job
is to look at company financials, speak to other analysts and industry pundits
and come up with pronouncements on the future performance of companies’
stocks.
Since the audience was not from the IT industry but have
been subjected to the blizzard of IT terms we in the industry carelessly toss
about, I thought I’d give them a layman’s perspective on two
frequently-encountered buzzwords: Big
Data and the IoT (Internet of things).
We talk a lot about Big Data as if we grew up with the
stuff, but in fact, the truth is far more interesting than this blasé attitude
suggests. There’s lots of data in the
world, and 90% of it was created in the last 2 years. You’ve probably heard
this before, but what you may not know is that that statement will be true in
2, 3, 5, years time from now and was true last year, and the year before that
and the year before.
What it means is that the rate of data generation is not
rising at a steady, arithmetic pace, but at something far more frenetic. It’s
coming at us from our cellphones, from sensors embedded in the ground, in
machines, satellites, the Internet, it’s being generated by social media,
machines programmed to churn out spam (most of the world’s email is spam, by
the way), the mass media, etc.
The rate of data accumulation is truly quite staggering – by
2020, IDC expects that there will be 40 zettabytes of data generated. That
number makes no sense if I tell you it’s a 40 followed by 21 zeros, but a sense
of the scale can be obtained if I say it is about 57 times the number of grains
of sand on all the beaches on Earth, combined.
And by the way, that’s 40 Zb of data generated PER YEAR!
I could chuck out a couple more staggering statistics like
that, but it does beg the question: how did we get here?
The world's first solid-state transistor,
invented at Bell Labs in 1947
The answer is somewhat prosaically described by Moore’s Law,
which is actually an astute observation made by Gordon Moore back in 1965 that
the density of transistors that could be crammed into an integrated circuit
doubles every two years. Consider that
the transistor was invented in 1947, and that we currently make 400 million
transistors for every one of the 7.3 billion persons on this planet – per year –
and it rather gives you a headache trying to get a sense of the scale, but that’s
how we continue to generate prodigious amounts of data, and how it keeps
increasing, year after year, at an exponential rate.
This doubling of computing capacity, by the way, has the
corollary that processing power has grown by almost unimaginable leaps and
bounds. The ignition chip in your car,
for example, contains more computing power than NASA had at its disposal back
in 1969 when they sent man to the moon.
And at a tiny, tiny, tiny, fraction of the price.
The second part of my talk focused on the IoT and how it,
and Big Data, are intimately related and the applications of all this, but I
think for this story, that simply getting a scale of things is a fairly good
start.
No comments:
Post a Comment