DataOps

DataOps is an automated, process-oriented methodology, used by analytic and data teams, to improve the quality and reduce the cycle time of data analytics. While DataOps began as a set of best practices, it has now matured to become a new and independent approach to data analytics.[1] DataOps applies to the entire data lifecycle[2] from data preparation to reporting, and recognizes the interconnected nature of the data analytics team and information technology operations.[3] From a process and methodology perspective, DataOps applies Agile software development, DevOps[3] and the statistical process control used in lean manufacturing, to data analytics.[4]

In DataOps, development of new analytics is streamlined using Agile software development, an iterative project management methodology that replaces the traditional Waterfall sequential methodology. Studies show that software development projects complete significantly faster and with far fewer defects when Agile Development is used. The Agile methodology is particularly effective in environments where requirements are quickly evolving — a situation well known to data analytics professionals.[5]

DevOps focuses on continuous delivery by leveraging on-demand IT resources and by automating test and deployment of analytics. This merging of software development and IT operations has improved velocity, quality, predictability and scale of software engineering and deployment. Borrowing methods from DevOps, DataOps seeks to bring these same improvements to data analytics.[3]

Like lean manufacturing, DataOps utilizes statistical process control (SPC) to monitor and control the data analytics pipeline. With SPC in place, the data flowing through an operational system is constantly monitored and verified to be working. If an anomaly occurs, the data analytics team can be notified through an automated alert.[6]

DataOps is not tied to a particular technology, architecture, tool, language or framework. Tools that support DataOps promote collaboration, orchestration[4], agility[5], quality, security, access and ease of use.[7]

History

DataOps was first introduced by Lenny Liebmann, Contributing Editor, InformationWeek, in a blog post on the IBM Big Data & Analytics Hub titled "3 reasons why DataOps is essential for big data success" on June 19, 2014.[8] The term DataOps was later popularized by Andy Palmer at Tamr.[3] DataOps is a moniker for "Data Operations."[2] 2017 was a significant year for DataOps with significant ecosystem development, analyst coverage, increased keyword searches, surveys, publications, and open source projects.[9]

DataOps heritage from DevOps, Agile, and manufacturing

Goals and Philosophy

The volume of data is forecast to grow at a rate of 32% CAGR to 180 Zettabytes by the year 2025 (Source: IDC).[7] DataOps seeks to provide the tools, processes, and organizational structures to cope with this significant increase in data.[7] Automation streamlines the daily demands of managing large integrated databases, freeing the data team to develop new analytics in a more efficient and effective way.[10]

DataOps embraces the need to manage many sources of data, numerous data pipelines and a wide variety of transformations.[3] DataOps seeks to increase velocity, reliability, and quality of data analytics.[11] It emphasizes communication, collaboration, integration, automation, measurement and cooperation between data scientists, analysts, data/ETL(extract, transform, load) engineers, information technology (IT), and quality assurance/governance.[12] It aims to help organizations rapidly produce insight, turn that insight into operational tools, and continuously improve analytic operations and performance.[12]

Implementation

Toph Whitmore at Blue Hill Research, offers these DataOps leadership principles for the information technology department:[1]

  • “Establish progress and performance measurements at every stage of the data flow. Where possible, benchmark data-flow cycle times.
  • Define rules for an abstracted semantic layer. Ensure everyone is “speaking the same language” and agrees upon what the data (and metadata) is and is not.
  • Validate with the “eyeball test”: Include continuous-improvement -oriented human feedback loops. Consumers must be able to trust the data, and that can only come with incremental validation.
  • Automate as many stages of the data flow as possible including BI, data science, and analytics.
  • Using benchmarked performance information, identify bottlenecks and then optimize for them. This may require investment in commodity hardware, or automation of a formerly-human-delivered data-science step in the process.
  • Establish governance discipline, with a particular focus on two-way data control, data ownership, transparency, and comprehensive data-lineage tracking through the entire workflow.
  • Design process for growth and extensibility. The data flow model must be designed to accommodate volume and variety of data. Ensure enabling technologies are priced affordably to scale with that enterprise data growth.”

DataOps Manifesto

Individuals and organizations supporting DataOps have produced a DataOps manifesto, consisting of 18 DataOps principles, which summarize the mission, values, philosophies, goals and best practices of DataOps practitioners.

The Role of DataOps Engineer

The DataOps Engineer orchestrates and automates the data analytics pipeline, promotes features to production and automates quality.[13] In many organizations, the DataOps engineer is a separate role. In others, it is a shared function.[14]

More DataOps Information

References

  1. 1 2 "DataOps – It's a Secret". www.datasciencecentral.com. Retrieved 2017-04-05.
  2. 1 2 "What is DataOps (data operations)? - Definition from WhatIs.com". SearchDataManagement. Retrieved 2017-04-05.
  3. 1 2 3 4 5 "From DevOps to DataOps, By Andy Palmer - Tamr Inc". Tamr Inc. 2015-05-07. Retrieved 2017-03-21.
  4. 1 2 "The DataOps Ecosystem Emerges - Tamr Inc". Tamr Inc. 2017-05-04. Retrieved 2017-08-24.
  5. 1 2 DataKitchen (2017-02-21). "How Software Teams Accelerated Average Release Frequency from 12 Months to Three Weeks". Medium. Retrieved 2017-08-24.
  6. DataKitchen (2017-03-07). "Lean Manufacturing Secrets that You Can Apply to Data Analytics". Medium. Retrieved 2017-08-24.
  7. 1 2 3 "What is DataOps? | Nexla: Scalable Data Operations Platform for the Machine Learning Age". www.nexla.com. Retrieved 2017-09-07.
  8. "3 reasons why DataOps is essential for big data success". IBM Big Data & Analytics Hub. Retrieved 2018-08-10.
  9. DataKitchen (2017-12-19). "2017: The Year of DataOps". data-ops. Retrieved 2018-01-24.
  10. "5 trends driving Big Data in 2017". CIO Dive. Retrieved 2017-09-07.
  11. "Unravel Data Advances Application Performance Management for Big Data". Database Trends and Applications. 2017-03-10. Retrieved 2017-09-07.
  12. 1 2 DataKitchen (2017-03-15). "How to Become a Rising Star with Data Analytics". data-ops. Retrieved 2017-09-07.
  13. DataKitchen (2017-07-19). "Building a DataOps Team". data-ops. Retrieved 2017-08-24.
  14. DataKitchen (2017-05-16). "DataOps Engineer Will Be the Sexiest Job in Analytics". Medium. Retrieved 2017-08-24.
This article is issued from Wikipedia. The text is licensed under Creative Commons - Attribution - Sharealike. Additional terms may apply for the media files.