For any company to be competitive in local or global economies today, it needs to look at big data aspect for its future operations. In other words, what is today's sales data not last month's, what kind of feedback are the sales people bring back to the organization with big data aspects.
Dannie Stelyn, Intel Corporation General Manager for East and Sub-Saharan Africa says big data is today's biggest single opportunity and a threat too, for any company. "Why is it an opportunity, it increases your top line today, because you will understand what is happening, and make more informed decisions, turns you to be much more agile".
Stelyn notes that definitely big data will give a company an advantage over the competitor for not using it. "However if you ignore it, it will be a big challenge for you since other companies will adopt it".
Meet the new boss
The Wall Street Journal in September this year had an article, "Meet the new boss: Big Data", and it is recognized as a big deal for companies. They will no longer have to rely on executives intel about how successful or not a company will be.
However big data will help you decide which side to go on to business and to realize that consumers are not talking about your company's particular product; of which your future is build on or there is another huge window of an opportunity from your production line.
Big data is the new boss, and it is completely changing how business is being done.
What are companies doing about it?
"We were used to this, being unstructured; we then moved the whole situation where IT has become the business, and extremely dependant on what kind of product companies deliver. Therefore you can no longer be for example in manufacturing industry without your IT environment being in place", says Stelyn. This is the case because the competitiveness of a company's IT environment has a direct effect to its income.
An example of what is happening at BMW, the company's IT used to have 100,000 employees worldwide and so their IT environment had to support these 100,000 people. However this year they have 1 million vehicles that they are now adding as customers to their IT environment.
This because all these vehicles are connected to various IT platforms, by 2015 there will be 10 million connected vehicles which among other data mining features they should be able to upload and pull numerous data to and from the cloud, and by 2018 the company foresees that it will create 1 terabyte of big data in day. So BMW is looking at how it can make use of this data.
What to watch
Therefore companies need to watch out for four things in this IT transformation era, self service, customers, on demand services and resilience.
In the 90s, the IT environment was a mess, it was not a nicely integrated, come 2000 and 2010 period IT shifted into the virtualization era, and in 6 years it was all about talks, companies' implemented virtualization only at the end of this period.
There was however a substantial improvement in terms of IT utilization then. So far IT become a service environment, the network, storage and processing facilities become a complete unit as opposed to separate systems that does computing, storage etc, even the network switching is no longer traditional, it is more like a server, and so companies needs to be convergent to new capabilities, competiveness and work capabilities in big data.
These three things are imperatives for big data and for any company to be competitive why is this? Because they can no longer divorce their IT environment from doing business, a company's IT cost has an implication on its product, a number of people are moving away from propriety systems into open standard solutions, and this is what is affecting efficiency.
Another aspect is to look at a company competitiveness verses the cloud environment, for example how competitive is a company investing on its own data centre than doing the same in the cloud.
"Cloud services in Kenya are a reality especially now that the country is cabled. Having your own data center rather than the cloud is no longer a CAPEX question but an efficiency question. So the current existing benchmark is a baseline of 10 cents per Vmh, this has come down a lot in the past year and Intel foresees that it will come down further", says Stelyn.The cloud or in house virtual hosting
Now the big question is if companies have their own data centre can they measure the cost per VM? And this is one thing CIOs need to look at, how do they quantify cost for running their VM and at what point should they decide what to keep in the cloud and what applications should they host in house.
Regulations that prevent data from being online and offline will change, they will no longer be a problem in future, for a CIO, he/she will decide what the cost effective solution is, and will they build their own solution in the data center to make sure that they are competitive?
With supercomputing, companies will be able to decide when to switch their applications to the cloud and having them in their own data centre depending on the time and cost. This is the competitive side of it.
The convergence side, traditionally we used to have storage, servers and networks, these are the three big areas that CIOs looked at when they looked at data centre environments, however the lines in-between these areas have completely blurred, in the sense that you very often can no longer tell the difference of your data centre's object store, distribution analysis, databases and actual servers. These are becoming blurred because of open standards, where companies have been able to do things most cost effectively.
The other reason why this is the case is that companies have started to get devices that do more than one of this things, switches and routers are becoming smart, so how companies are going to handle data generated from these devices is the bigger question. So there are a few ways in which storage is being addressed to allow big data get addressed efficiently, one of them is intelligent tiering where important data is crunched rather than storing it in the backend and then bring it forward into a new environment every time.
Thin provisioning is also another important concept, where a little data storage is allocated and scaled depending on the application on demand. Realtime compression is another aspect when dealing with big data where one needs to decompress lots of data in real time off storage, and because of the capacity, the solutions have become more powerful, for example de-duplication.