The Third Phoenix Data Conference – the largest big data event in the Phoenix Valley had a tremendous response this year. Having established itself as a must attend big data event in the Phoenix area, the event this year attracted thought leaders and key companies discussing latest trends and technologies in the big data space. Over 350 technologists, business leaders, data analysts and engineers attended the 2016 conference The Phoenix data conference 2016, that concluded last Saturday focussed on practical big data use cases from leading analytics companies like Cloudera, SAP, Clairvoyant, MapR, StreamSets, Microsoft, Confluent, SnapLogic, DataTorrent, Tresata, Amazon, Choice Hotels, MemSQL, Wells Fargo and others. Technology leaders in the big data space shared innovative implementations and advances in the hadoop space. Specific challenges around security, talent availability, technical deployments, managed services etc were discussed by the speakers. We would like to thank our speakers, sponsors, all the attendees and the volunteers for making the Phoenix Data Conference – 2016 a huge success….
Author: Avinash Ramineni
How to upgrade Spark on CDH5.5
CDH is normally on a six month release cycle and Spark is on 4 month release cycle. This often leads to an extra latency before a spark version gets integrated and supported in CDH. At times Cloudera might even choose to skip a couple of Spark versions. Reason for this latency seems to be the effort needed with integration testing and bugs to fix in Spark and other projects to get it all work together in time. But if you are like our teams, and can’t wait to get your hands on latest and greatest of both spark and CDH, you can always run the latest version of Spark on CDH. We encountered a similar scenario recently, one of our clients wanted to use the Spark Thrift Server on CDH5.5, but Spark Thrift JDBC/ODBC server is not included on CDH5.5. We figured this might be a common use case for many of you who…