In the world of distributed computing, Spark has simplified development and open the doors for many to start writing distributed programs. Folks with little to none distributed coding experience can now start writing just a couple lines of code that will get 100s or 1000s of machines, immediately, working on creating business value. However, even through Spark code is easy to write and read, that doesn’t mean that users don’t run into issues of long running, slow performing jobs or out of memory errors. Thankfully most of the issues with using Spark have nothing to do with Spark but the approach we take when using it. This session will go over the top 5 things that we’ve seen in the field that prevent people from getting the most out of their Spark clusters. When some of these issues are addressed, it is not uncommon to see the same job running 10x or 100x faster with the same clusters, the same data, just a different approach.
Mark is a software engineer working on Apache Spark at Cloudera. He is a co-author of Hadoop Application Architectures book and also wrote a section in Programming Hive book. Mark is also a committer on Apache Bigtop and a committer and PMC member on Apache Sentry. He has contributed to a number of open source projects including Apache Hadoop, Apache Hive, Apache Sqoop and Apache Flume projects. Mark is sought after speaker on topics related to Big Data at various national and international conferences. He occasionally blogs on topics related to technology on his blog.
Ted has seen the world of data from helping out hundreds of different companies while serving as a Printable Solutions Architect at Cloudera to multiple years at the leading game company Blizzard building out data pipelines, and managing data engineering efforts. Now Ted, servers as a Directory of Enterprise Architecture at Capital One, solving data problems at every level of the company.