big data, big thinking: unifying your it architecture

8
SAP Big Data, Big Thinking webinar series Big Data, Big Thinking: Unifying your IT architecture Research shows that most enterprises are using more than one platform in their Big Data environments. To be effective, a hybrid approach must be simple, efficient and offer real time insights. In this penultimate “Big Data, Big Thinking” webinar, Hortonworks’ Ajay Singh, and SAP’s Sid Sipes explored how a unified, future-proof IT architecture can help you simplify data analysis across various sources and supercharge the processing of structured and unstructured data residing in Hadoop.

Upload: sap-technology

Post on 27-Jun-2015

1.168 views

Category:

Technology


0 download

DESCRIPTION

Research shows that most enterprises are using more than one platform in their Big Data environments. To be effective, a hybrid approach must be simple, efficient and offer real time insights. In this penultimate “Big Data, Big Thinking” webinar, Hortonworks’ Ajay Singh, and SAP’s Sid Sipes explored how a unified, future-proof IT architecture can help you simplify data analysis across various sources and supercharge the processing of structured and unstructured data residing in Hadoop.

TRANSCRIPT

Page 1: Big Data, Big Thinking: Unifying Your IT Architecture

SAP Big Data, Big Thinking webinar series

Big Data, Big Thinking: Unifying your IT architectureResearch shows that most enterprises are using more than one platform in their Big Data environments. To be effective, a hybrid approach must be simple, efficient and offer real time insights. In this penultimate “Big Data, Big Thinking” webinar, Hortonworks’ Ajay Singh, and SAP’s Sid Sipes explored how a unified, future-proof IT architecture can help you simplify data analysis across various sources and supercharge the processing of structured and unstructured data residing in Hadoop.

Page 2: Big Data, Big Thinking: Unifying Your IT Architecture

SAP Big Data, Big Thinking webinar series

The story of Hadoop

Once upon a time, a team at Yahoo was tasked to develop a large-scale data storage and processing technology that would accelerate search performance. Their challenge was twofold:

To capture and archive the contents of the internet

Toprocessthedatasouserscouldsearchmoreefficiently

Traditional approaches were technically and commercially impractical due to data volume and cost considerations.

In 2005, Hadoop as we know it was born. The team came up with the Hadoop Distributed File System, providing high aggregate bandwidth across a cluster of commodity machines, and Hadoop Mapreduce, a programming model for large-scale data processing.

2

Page 3: Big Data, Big Thinking: Unifying Your IT Architecture

SAP Big Data, Big Thinking webinar series

The story of HadoopYahoo committed to an open source approach that would enable Hadoop to work within the Apache Software Foundation, so it could be developed publicly and collaboratively, accelerating progress for all.

Soon, all the leading consumer web companies began to use and advance it, until by 2011, Hadoop underpinned every click at Yahoo, and the infrastructure had grown to 35,000 nodes.

Hadoop was named after a toy elephant belonging to co-creator Doug Cutting’s son.

3

Page 4: Big Data, Big Thinking: Unifying Your IT Architecture

SAP Big Data, Big Thinking webinar series

The story of HadoopBefore long, mainstream IT started to look closely at Hadoop as a way to address the architectural challenges faced by the explosion of data generated by mobile, social and machine sensors.

At this point, to facilitate broader market adoption, Yahoo gave its blessing to the core Hadoop team to fly the nest and form Hortonworks.

The new company had a singular vision: to mature its offspring into Enterprise Hadoop – a complete open source data platform enabling a modern architecture that would allow organizations to address their Big Data challenges.

4

Page 5: Big Data, Big Thinking: Unifying Your IT Architecture

SAP Big Data, Big Thinking webinar series

The Hortonworks Data PlatformHadoop has come a long way. The HDP is an enterprise-class data management platform that delivers the key requirements of governance, security and operations.

YARN (Yet Another Resource Negotiator) forms the architectural epicentre of HDP and allows existing apps to plug directly into Hadoop. It enables multiple data access engines across batch, interactive and real-time workloads on a single set of data, and also extends Hadoop to integrate with the tools and systems already found in the enterprise data centre.

A wide range of deployment options are supported, including Linux and Windows, on-premise and cloud.

5

Page 6: Big Data, Big Thinking: Unifying Your IT Architecture

SAP Big Data, Big Thinking webinar series

SAP and HDP: the best of both worldsHadoop is great for economically storing machine-generated data, analysis and model-building, but it lacks the transactional capability of enterprise-class databases that provide sub-second multi-user response.

Enterprises currently tend to move transactional data through a series of systems – operational data stores, data warehouses, data marts, aggregates and cubes – to allow a niche community of power users to analyse or report on data. This process can take days or weeks.

Big Data demands an integrated platform.

Such organisations are now revisiting their data centres and modernising their architecture and data platforms with a combination of HPD and SAP HANA, the class-leading enterprise database, to get the best of both worlds.

SAP Big Data, Big Thinking webinar series6

Page 7: Big Data, Big Thinking: Unifying Your IT Architecture

SAP Big Data, Big Thinking webinar series

The combined platform offers the infinite storage and deep processing capabilities of Hadoop, with the seamless user experience of Big Data analysis with SAP BusinessObjects Business Intelligence tools, and rapid text data processing, data movement and data quality management with SAP Data Services.

Together, SAP HANA and Hortonworks rewrite the book on Big Data to enable real-time analytics within applications and core operational processes. Hadoop expands the scope of capabilities within the SAP real-time data platform to create a seamless experience for a broad community of decision makers, middle managers and frontline employees – not just data scientists.

SAP and HDP: the best of both worlds

SAP Big Data, Big Thinking webinar series7

Page 8: Big Data, Big Thinking: Unifying Your IT Architecture

SAP Big Data, Big Thinking webinar seriesCopyright statement – © 2014 SAP SE or an SAP affiliate company. All rights reserved.

Join usJoin the last session in our 2014 Big Data, Big Thinking webinar series, as the experts from SAP explore the Internet of Things – making it real. For content from previous Big Data, Big Thinking webinars and to register for the season finale,

Visit bit.ly/BigDataBigThinking