Nine Best Practices for Working With Hadoop in the Enterprise

 
 
By Chris Preimesberger  |  Posted 2015-05-12
 
 
 
 
 
 
 
 
 
  • Previous
    1 - Nine Best Practices for Working With Hadoop in the Enterprise
    Next

    Nine Best Practices for Working With Hadoop in the Enterprise

    by Chris Preimesberger
  • Previous
    2 - Promote Collaboration Around Performance
    Next

    Promote Collaboration Around Performance

    When there's a performance problem on Hadoop, there can be several culprits: your code, your data, your hardware or the way you're sharing resources (your cluster configuration). At a startup, a single person—data scientist, developer and operator all rolled into one—might be responsible for all that. But at a large enterprise, multiple teams have to cooperate to figure out what went wrong and how to fix it. If you're managing a big data operation at a large, distributed organization, nurture collaboration by giving your team tools that let developers, operators and managers work together to address performance issues.
  • Previous
    3 - Make It Easy to Share Application Context Around Errors
    Next

    Make It Easy to Share Application Context Around Errors

    When execution errors do arise, teams can spend hours tracking down what went wrong if all they've got are Hadoop's log files and Job Tracker. Invest in tools that help your team quickly connect errors to application context—where in your code they're happening—and share that information easily.
  • Previous
    4 - Monitor the Fleet, Not the Vehicle
    Next

    Monitor the Fleet, Not the Vehicle

    To an operator running hundreds or thousands of applications on a Hadoop cluster, all of them look the same—until there's a problem. So you need tools that let you look at performance over groups of applications. Ideally, you should be able to segment performance tracking by application types, departments, teams and data-sensitivity levels.
  • Previous
    5 - Define and Enforce Service-Level Agreements
    Next

    Define and Enforce Service-Level Agreements

    Monitoring a fleet still means knowing when an individual vehicle performs poorly. Similarly, operators need to set SLA bounds on performance and define alerts and escalation paths when they're violated. SLA bounds should incorporate both raw metadata, such as job status, as well as business-level events, such as sensitive data access. Successful practitioners of operational readiness also set up metrics that help predict future SLA violations, so they can proactively address and avoid them.
  • Previous
    6 - Understand Inter-App Dependencies
    Next

    Understand Inter-App Dependencies

    Large, traditional enterprises tend to run their Hadoop clusters as a shared service across many lines of business. As a result, each application has at least a few "roommates" in the cluster, some of which can be detrimental to its own performance. To understand the errant behavior of one Hadoop application, operators must understand what others were doing on the cluster when it ran. Therefore, provide your operations team with as much cluster-related context as possible.
  • Previous
    7 - Ration Your Cluster
    Next

    Ration Your Cluster

    To optimize cluster use and ROI, operators must ration resources on the cluster and enforce the limits. An operator can budget mappers for the execution of a particular application, and if the application doesn't perform appropriately, rationing rules should prevent the application from being deployed. Establishing and enforcing the rules for rationing cluster resources is vital for achieving meaningful operational readiness and meeting SLA commitments.
  • Previous
    8 - Trace Data Access at the Operational Level
    Next

    Trace Data Access at the Operational Level

    Good Hadoop management isn't only about rationing compute resources; it also means regulating access to sensitive data, especially in industries with heightened privacy concerns like health care, insurance and financial services. Solving for data lineage and governance in an unstructured environment like Hadoop is difficult. Traditional techniques to manually maintain a metadata dictionary quickly lead to stale and old repositories, and they offer no way to prove that a production dataset is dependent on some fields and not on others. As a result, visibility and enforcement on the use of data fields are required at the operational level. If you can reliably track if and when a data field is accessed by an app, your compliance teams will be happy.
  • Previous
    9 - Record Data Misfires
    Next

    Record Data Misfires

    Compliance professionals at large enterprises also want proof that a Hadoop application processed every record in a dataset, and they look for documentation when it fails to do so. Failures can result from format changes in upstream data sets or plain old data corruption. Keeping track of all records that the application failed to process is particularly vital in regulated industries.
  • Previous
    10 - Tune Your Engine Before You Replace It
    Next

    Tune Your Engine Before You Replace It

    With new compute fabrics emerging all the time, teams are sometimes too quick to junk their old ones in pursuit of better performance. However, it's often the case that you can achieve equal or greater performance gains just by optimizing code and data flows on your existing fabrics. That way, you can avoid expensive infrastructure upgrades unless they're truly necessary.
 

Getting started with Hadoop is like buying a car—about 100 years ago. Back then, only hobbyists and status seekers drove their own vehicles, which were temperamental, to put it mildly. To operate a car, you pretty much had to be a mechanic because the only thing you could count on was a breakdown. The dashboard showed little, if any, actionable information. Only after automobiles became user-friendly did they really take off. There are parallels with Apache Hadoop. Enterprise IT teams were once led to believe that if they built on top of Hadoop, they could replicate the business successes of giants like Twitter, LinkedIn and Netflix. However, working on Hadoop often has proved problematic, and enterprises are struggling to deploy Hadoop applications with the standards of quality, reliability and manageability that they expect. Out of the box, Hadoop is not operationally ready. This slide show, developed using eWEEK reporting and insight from Supreet Oberoi, vice president of field engineering at big data specialist Concurrent, discusses best practices for how IT organizations can achieve operational readiness on Hadoop—or how to transform every big data project from a Model T into a Camry.

 
 
 
 
 
 
 
 
 
 
 

Submit a Comment

Loading Comments...
 
Manage your Newsletters: Login   Register My Newsletters























 
 
 
 
 
 
 
 
 
Rocket Fuel