Day One Kickoff | Spark Summit 2016
01. John Walls And George Gilber Kickoff Day On At Spark Summit 2016. (00:20) 02. What Do You Think Of Evolving Spark Into Continuous Applications. (00:56) 03. There's Going To Be A Launch Of Spark 2.0 With A Lot Of Improvements. (02:18) 04. How Do You Keep Up With The Massive Amounts Of Data. (03:50) 05. Is Hadoop And Spark Coming Together Like A Marriage. (05:30) 06. Lets Talk About The Future. (08:30) 07. What About The Education Piece And The Skills Gap. (10:33) 08. Great Line Up Of Guests. (12:05) Track List created with http://www.vinjavideo.com. --- --- How is Spark creating seamless cutting-edge data processing? Analysts speak at #SparkSummit by Tim Hawkins | Jun 7, 2016 The way in which incoming information is handled by the programs dealing with it has changed drastically since its beginnings 50 years ago. The Internet of Things (IoT) and the speed of how businesses are operating and interfacing today has made it critical that these applications keep up. Giving an overview of how Spark is combining structured streaming and traditional data input, as well as where the platform is going in the future, were George Gilbert (@ggilbert41) and John Walls, cohosts of theCUBE, from the SiliconANGLE Media team, live from HPE Discover 2016. Bringing together the past and present One of the biggest challenges facing developers today is learning to work with structured streaming data and the sheer volume of data coming in, as well as more traditional data, at the same time. Spark is designed with this in mind and has the capability to scale itself to the data being received. “One of the first things Spark did that was appealing to developers was they made it very easy to work with streams and traditional tables. Now, they’re making it so they’re exactly the same,” explained Gilbert. Smarter software means less work for developers In addition to helping programmers work seamlessly with a wide range and speed of data input, Spark is also designed with machine learning in mind. The program will actually train itself to deal with certain functions and scenarios as they arise. This takes a great deal of pressure off of developers who would normally have to continuously train to learn new nuances of the software. “We are going to train people to be better with data science, but also be building better tools,” said Gilbert. #SparkSummit #theCUBE