Welcome!

Open Web Authors: Lori MacVittie, Liz McMillan, Gilad Parann-Nissany, Carmen Gonzalez, Mark R. Hinkle

Related Topics: Cloud Expo, SOA & WOA, Virtualization, Web 2.0, Open Web, Security

Cloud Expo: Article

Lessons Learned from Real-World Big Data Implementations

The value of Big Data is in the insights that the data can provide

In the past few weeks I visited several Cloud and Big Data conferences that provided me with a lot of insight. Some people only consider the technology side of Big Data technologies like Hadoop or Cassandra. The real driver however is a different one. Business analysts have discovered Big Data technologies as a way to leverage tons of existing data and ask questions about customer behavior and all sorts relationships to drive business strategy. By doing that they are pushing their IT departments to run ever bigger Hadoop environments and ever faster real-time systems.

What's interesting from a technical side is that ad-hoc analytics on existing data is allowed to take some time. However ad-hoc implies people waiting for an answer, meaning we are talking about minutes and not hours. Another interesting insight is that Hadoop environments are never static or standalone. Most companies take in new data on a continuous basis via technologies like flume. This means Hadoop MapReduce jobs need to be able to keep up with the data flow, either by adding more hardware or by optimizing them.

There are multiple drivers to Big Data (actually there are a lot) but the two most important ones are these: Analytics and Technical Need for Speed. Let's look at some of those and the resulting takeaways.

The Value Is in the Insight Not the Volume
The value of Big Data is in the insights that the data can provide, not the sheer volume of it. The reason that more and more companies are keeping all of their log and transaction data is that they want to gain those insights. The sheer size of the data is rather an obstacle to this goal and has been for a long time. With Big Data technologies this value can be harnessed.

Don't Forget That Data Analysts Are People Too
Ad-hoc analytics doesn't have to be instant, but must not take hours either. It was interesting to see that time to result on ad-hoc analytics is considered important. This is because people are doing those queries, and people don't like to wait for hours. But even more important is that business analytics is often an iterative process. Ask a question, check the answer, refine or change the question. Hours long MapReduce jobs are prohibitive to this process.

New Data Is Coming in All the Time
Big Data environments are constantly fed new data. This is not really big news, but I was still surprised by the constant reiteration of this fact. The constant data growth means that ad-hoc queries get either slower over time or need to work on samples. To remedy this, companies are writing, scrubbing and categorizing MapReduce jobs. These jobs basically strip out all the unimportant stuff and put cleansed, streamline easy-to-access data into new files. Instead of executing analytics against raw files, the analyst works on a cleansed data set. The implications are that scrubbing jobs need to be maintained all the time (as data input is changing over time) and they need to be able to keep up with the velocity of the input. MapReduce is not allowed to run for hours, but needs to be quick and iterative.

Big Data Is Not Cheap
While it sounds obvious, it's something that's not talked about by the vendors unless specifically asked. Hadoop requires a lot of hardware and a lot of expertise. Especially the expertise is hard to come by as of yet. While hardware might be cheap (you don't need expensive boxes for Hadoop) the bigger the environment the higher the operational costs. That operational cost is the reason some Hadoop vendors exist on services alone and also why customers are demanding better monitoring and management solutions.

Data Must Be Accessible at Low Latencies to Provide Value
One very interesting fact is that most early adopters that use Hadoop for analytics use it for ad-hoc analytics and not as a traditional warehouse. They use MapReduce to do the heavy lifting that is usually reserved for ETL jobs and put the resulting dimensions in existing data warehouses or into a NoSQL solution like HBase, Cassandra or MongoDB. These solutions provide low latency access semantics and are then integrated in the transactional application world, e.g. to provide recommendations to the end users.

This does not absolve them from optimizing their Hadoop environment where they can, but it gives them the much needed real time access that Hadoop so far does not provide. This also makes for additional complexity that needs to be maintained and monitored.

NoSQL Solutions Need Management and Monitoring as Well
NoSQL solutions are most often used to provide low latency databases with failover and horizontal scaling characteristics. As expected, practitioners quickly run into new issues like distribution and wrong access patterns. Most NoSQL solutions lack sophisticated monitoring or performance analysis tools and require experts instead. Fortunately several companies are working on providing those tools and some APM vendors work hard to support NoSQL databases similar to normal databases. This is emphasized by another interesting finding: With a fast and scalable data storage, the application itself quickly becomes the response time and scaling bottleneck.

Applications Using NoSQL Technologies Are More Complex
Most NoSQL solutions surrender more complex logic like joins in order to achieve horizontally scalable data distribution. That logic is moved to the application - arguably this is where it should be anyway. NoSQL solutions require data to be stored in a query access optimized way - de-normalization is the key. The flip side of storing data multiple times and the need to keep it in sync on updates, is that the storage logic again becomes more complex. More application logic usually means less performance.

My conclusion as a performance engineer is relatively clear: Big Data requires Performance Management and Monitoring Tools to fulfill its promise in a cost effective and timely manner. Here are some suggestions on what you should think about when you start a Big Data project.

  1. Large Hadoop environments are hard to manage and operate. Without automation in terms of deployment, operations, monitoring and root cause analysis they quickly become unmanageable. Make sure to have a monitoring solution in place that informs you pro-actively of any infrastructure or software issues that would affect your operation. It needs to give you an easy way to pinpoint the root cause.
  2. The easiest way to identify new performance issues is to detect and analyze change. Adopt a life cycle and 24/7 production APM approach. It will enable you to notice changes in data and compute distribution over time. In addition a life cycle approach will allow you to immediately pin point any negative changes introduced by a new software release.
  3. Don't just throw more and more hardware at the problem. While you can use cheaper hardware for Hadoop, it's still cost. But more than that you have to consider the operational drag. Every node you add will make traditional log based analysis more complicated. Instead ensure that you have an APM solution in place that lets you understand and optimize MapReduce jobs at their core and reduce both the time and resources it takes to run them.
  4. Your Hadoop cluster is no island, but will always be connected in some form or the other to a real time or at least transactional system. Make sure that you have a monitoring solution in place that can support both.

NoSQL applications tend to have more complex logic. The very performance and scalability of the store depends on correct data access and data distribution. An good monitoring solution allows you to monitor and optimize that additional complexity with ease; it also enables you to understand how your application access the data and how that access is distributed across your NoSQL cluster in your production system. The best way to ensure a scalable and fast NoSQL store is to ensure optimal distribution and access patterns.

Conclusion
Big Data is still very much an emerging technology and its promises are huge. But in order to deliver on those promises it must be cost and time effective to those that harness its value - The Business and not just technology experts.

More Stories By Michael Kopp

Michael Kopp has over 12 years of experience as an architect and developer in the Enterprise Java space. Before coming to CompuwareAPM dynaTrace he was the Chief Architect at GoldenSource, a major player in the EDM space. In 2009 he joined dynaTrace as a technology strategist in the center of excellence. He specializes application performance management in large scale production environments with special focus on virtualized and cloud environments. His current focus is how to effectively leverage BigData Solutions and how these technologies impact and change the application landscape.

Comments (0)

Share your thoughts on this story.

Add your comment
You must be signed in to add a comment. Sign-in | Register

In accordance with our Comment Policy, we encourage comments that are on topic, relevant and to-the-point. We will remove comments that include profanity, personal attacks, racial slurs, threats of violence, or other inappropriate material that violates our Terms and Conditions, and will block users who make repeated violations. We ask all readers to expect diversity of opinion and to treat one another with dignity and respect.


@ThingsExpo Stories
“In the past year we've seen a lot of stabilization of WebRTC. You can now use it in production with a far greater degree of certainty. A lot of the real developments in the past year have been in things like the data channel, which will enable a whole new type of application," explained Peter Dunkley, Technical Director at Acision, in this SYS-CON.tv interview at @ThingsExpo, held Nov 4–6, 2014, at the Santa Clara Convention Center in Santa Clara, CA.
SYS-CON Events announced today that IDenticard will exhibit at SYS-CON's 16th International Cloud Expo®, which will take place on June 9-11, 2015, at the Javits Center in New York City, NY. IDenticard™ is the security division of Brady Corp (NYSE: BRC), a $1.5 billion manufacturer of identification products. We have small-company values with the strength and stability of a major corporation. IDenticard offers local sales, support and service to our customers across the United States and Canada. Our partner network encompasses some 300 of the world's leading systems integrators and security s...
The BPM world is going through some evolution or changes where traditional business process management solutions really have nowhere to go in terms of development of the road map. In this demo at 15th Cloud Expo, Kyle Hansen, Director of Professional Services at AgilePoint, shows AgilePoint’s unique approach to dealing with this market circumstance by developing a rapid application composition or development framework.
The major cloud platforms defy a simple, side-by-side analysis. Each of the major IaaS public-cloud platforms offers their own unique strengths and functionality. Options for on-site private cloud are diverse as well, and must be designed and deployed while taking existing legacy architecture and infrastructure into account. Then the reality is that most enterprises are embarking on a hybrid cloud strategy and programs. In this Power Panel at 15th Cloud Expo (http://www.CloudComputingExpo.com), moderated by Ashar Baig, Research Director, Cloud, at Gigaom Research, Nate Gordon, Director of T...
"BSQUARE is in the business of selling software solutions for smart connected devices. It's obvious that IoT has moved from being a technology to being a fundamental part of business, and in the last 18 months people have said let's figure out how to do it and let's put some focus on it, " explained Dave Wagstaff, VP & Chief Architect, at BSQUARE Corporation, in this SYS-CON.tv interview at @ThingsExpo, held Nov 4-6, 2014, at the Santa Clara Convention Center in Santa Clara, CA.
SYS-CON Events announced today that Windstream, a leading provider of advanced network and cloud communications, has been named “Silver Sponsor” of SYS-CON's 16th International Cloud Expo®, which will take place on June 9–11, 2015, at the Javits Center in New York, NY. Windstream (Nasdaq: WIN), a FORTUNE 500 and S&P 500 company, is a leading provider of advanced network communications, including cloud computing and managed services, to businesses nationwide. The company also offers broadband, phone and digital TV services to consumers primarily in rural areas.
The Internet of Things is not new. Historically, smart businesses have used its basic concept of leveraging data to drive better decision making and have capitalized on those insights to realize additional revenue opportunities. So, what has changed to make the Internet of Things one of the hottest topics in tech? In his session at @ThingsExpo, Chris Gray, Director, Embedded and Internet of Things, discussed the underlying factors that are driving the economics of intelligent systems. Discover how hardware commoditization, the ubiquitous nature of connectivity, and the emergence of Big Data a...

ARMONK, N.Y., Nov. 20, 2014 /PRNewswire/ --  IBM (NYSE: IBM) today announced that it is bringing a greater level of control, security and flexibility to cloud-based application development and delivery with a single-tenant version of Bluemix, IBM's platform-as-a-service. The new platform enables developers to build ap...

DevOps Summit 2015 New York, co-located with the 16th International Cloud Expo - to be held June 9-11, 2015, at the Javits Center in New York City, NY - announces that it is now accepting Keynote Proposals. The widespread success of cloud computing is driving the DevOps revolution in enterprise IT. Now as never before, development teams must communicate and collaborate in a dynamic, 24/7/365 environment. There is no time to wait for long development cycles that produce software that is obsolete at launch. DevOps may be disruptive, but it is essential.
"People are a lot more knowledgeable about APIs now. There are two types of people who work with APIs - IT people who want to use APIs for something internal and the product managers who want to do something outside APIs for people to connect to them," explained Roberto Medrano, Executive Vice President at SOA Software, in this SYS-CON.tv interview at Cloud Expo, held Nov 4–6, 2014, at the Santa Clara Convention Center in Santa Clara, CA.
Nigeria has the largest economy in Africa, at more than US$500 billion, and ranks 23rd in the world. A recent re-evaluation of Nigeria's true economic size doubled the previous estimate, and brought it well ahead of South Africa, which is a member (unlike Nigeria) of the G20 club for political as well as economic reasons. Nigeria's economy can be said to be quite diverse from one point of view, but heavily dependent on oil and gas at the same time. Oil and natural gas account for about 15% of Nigera's overall economy, but traditionally represent more than 90% of the country's exports and as...
The Internet of Things is a misnomer. That implies that everything is on the Internet, and that simply should not be - especially for things that are blurring the line between medical devices that stimulate like a pacemaker and quantified self-sensors like a pedometer or pulse tracker. The mesh of things that we manage must be segmented into zones of trust for sensing data, transmitting data, receiving command and control administrative changes, and peer-to-peer mesh messaging. In his session at @ThingsExpo, Ryan Bagnulo, Solution Architect / Software Engineer at SOA Software, focused on desi...
"At our booth we are showing how to provide trust in the Internet of Things. Trust is where everything starts to become secure and trustworthy. Now with the scaling of the Internet of Things it becomes an interesting question – I've heard numbers from 200 billion devices next year up to a trillion in the next 10 to 15 years," explained Johannes Lintzen, Vice President of Sales at Utimaco, in this SYS-CON.tv interview at @ThingsExpo, held Nov 4–6, 2014, at the Santa Clara Convention Center in Santa Clara, CA.
"For over 25 years we have been working with a lot of enterprise customers and we have seen how companies create applications. And now that we have moved to cloud computing, mobile, social and the Internet of Things, we see that the market needs a new way of creating applications," stated Jesse Shiah, CEO, President and Co-Founder of AgilePoint Inc., in this SYS-CON.tv interview at 15th Cloud Expo, held Nov 4–6, 2014, at the Santa Clara Convention Center in Santa Clara, CA.
SYS-CON Events announced today that Gridstore™, the leader in hyper-converged infrastructure purpose-built to optimize Microsoft workloads, will exhibit at SYS-CON's 16th International Cloud Expo®, which will take place on June 9-11, 2015, at the Javits Center in New York City, NY. Gridstore™ is the leader in hyper-converged infrastructure purpose-built for Microsoft workloads and designed to accelerate applications in virtualized environments. Gridstore’s hyper-converged infrastructure is the industry’s first all flash version of HyperConverged Appliances that include both compute and storag...
Today’s enterprise is being driven by disruptive competitive and human capital requirements to provide enterprise application access through not only desktops, but also mobile devices. To retrofit existing programs across all these devices using traditional programming methods is very costly and time consuming – often prohibitively so. In his session at @ThingsExpo, Jesse Shiah, CEO, President, and Co-Founder of AgilePoint Inc., discussed how you can create applications that run on all mobile devices as well as laptops and desktops using a visual drag-and-drop application – and eForms-buildi...
We certainly live in interesting technological times. And no more interesting than the current competing IoT standards for connectivity. Various standards bodies, approaches, and ecosystems are vying for mindshare and positioning for a competitive edge. It is clear that when the dust settles, we will have new protocols, evolved protocols, that will change the way we interact with devices and infrastructure. We will also have evolved web protocols, like HTTP/2, that will be changing the very core of our infrastructures. At the same time, we have old approaches made new again like micro-services...
Code Halos - aka "digital fingerprints" - are the key organizing principle to understand a) how dumb things become smart and b) how to monetize this dynamic. In his session at @ThingsExpo, Robert Brown, AVP, Center for the Future of Work at Cognizant Technology Solutions, outlined research, analysis and recommendations from his recently published book on this phenomena on the way leading edge organizations like GE and Disney are unlocking the Internet of Things opportunity and what steps your organization should be taking to position itself for the next platform of digital competition.
The 3rd International Internet of @ThingsExpo, co-located with the 16th International Cloud Expo - to be held June 9-11, 2015, at the Javits Center in New York City, NY - announces that its Call for Papers is now open. The Internet of Things (IoT) is the biggest idea since the creation of the Worldwide Web more than 20 years ago.
As the Internet of Things unfolds, mobile and wearable devices are blurring the line between physical and digital, integrating ever more closely with our interests, our routines, our daily lives. Contextual computing and smart, sensor-equipped spaces bring the potential to walk through a world that recognizes us and responds accordingly. We become continuous transmitters and receivers of data. In his session at @ThingsExpo, Andrew Bolwell, Director of Innovation for HP's Printing and Personal Systems Group, discussed how key attributes of mobile technology – touch input, sensors, social, and ...