Apache Spark for Scala - Professional

Learn More

Apache Spark for Scala – Professional 

scala_learningWhen it comes to data analytics, Apache Spark is an open-source powerhouse. Developed at UC-Berkeley’s AMPLab, and later given to the Apache Software Foundation, Spark is a powerful programming interface that can be leveraged to create a wide variety of analytical, graph processing, and other frameworks. When combined with the flexibility of the Scala programming language, Spark developers can build unique software solutions applicable in nearly every area of business today.

This two-day Apache Scala workshop is designed to teach developers how to implement data analytics using Apache Spark for Reactive applications. In this workshop, developers will use hands-on exercises to learn the principles of Spark programming and idioms for specific problems, such as event stream processing, SQL-based analysis on structured data in files, integration with Reactive frameworks like Akka, Hadoop and related tools, and advanced analytics such as machine learning and graph algorithms.

Fees

$1200.00 USD

When you take the step toward learning Spark programming with Scala, you gain a professional skill that you can leverage throughout your career as a programmer. At YoppWorks, we strive to price our online courses fairly so that your investment in your career and skill set is one you can afford to make.

Participant

Our professional-level Apache Spark training can be valuable to a wide variety of developers working in a broad range of industry types. From those simply looking to expand their programming knowledge base to developers specifically seeking to create data analysis solutions for their company, all students with internet access can find something valuable in this convenient two-day, 14-hour online Scala course.

The following are a few examples of types of professionals who might be interested in taking the Apache Spark for Scala course:

  • Developers with basic knowledge of Scala, as covered in “Fast Track to Scala”
  • Developers with an interest in data science looking to put theory into high-scale practice
  • Managers who want to understand how to field applications powered by fast data analytics

Benefit

The benefits of learning to use Apache Spark with Scala are numerous, from the basic satisfaction of personal improvement to the ability to solve very specific problems with regard to business data.

Here are just a few possible advantages of taking our professional Apache Spark for Scala programming course:

  • Developers gain knowledge and skills to build Spark-based offline and event-streaming applications.
  • Certified Lightbend Trainers share best practices for developing Scala Spark applications.
  • Knowledge is useful for production data (integrating Spark with Akka, Hadoop).

Outline

This Apache Spark certification course covers six major areas of knowledge:

  1. Introduction to Spark
  2. Writing Spark RDD Applications I
  3. Writing Spark RDD Applications II
  4. Spark SQL and DataFrames
  5. MLib, GraphX, and Others
  6. Spark Clusters (optional)

Logistics

When you sign up for a YoppWorks Scala training course, you should make sure that the class in which you enroll appropriately suits your current skill level and professional goals. Before you go ahead and book a course with us, take a brief look at the following logistics to make sure you meet all necessary requirements and are able to commit to the class.

  • Level: Intermediate – knowledge of and practical experience with Scala is assumed
  • Length: Two days – 14 hours of classroom time, plus breaks
  • Approach: Extensive hands-on coding – students develop a workshop case study and produce a fully functional application that is event-driven and resilient
  • Requirements: Students must bring their own laptops with Java 6 or higher installed

Related Trainings

At YoppWorks, we’re passionate about helping professional developers expand their knowledge base and create new and better software solutions throughout their careers. As such, we offer many other courses beyond this Apache Spark for Scala class. Whether you have an interest in learning something entirely new or taking your existing Scala developer skills to an advanced level, we’ve got a Scala training class to suit your interests.

If you’re interested in pursuing further Scala learning with us, consider these other great online classes:

Our team of dedicated technology and programming experts is excited about helping you learn new and valuable software development skills. If you’re interested in learning how you can use Apache Spark with Scala to create reactive, user-friendly software solutions for data analytics and more, book a course online or contact us at (888) 322-6002 for more information.

Book Your Training Session Now

Apache Spark for Scala - Professional (Formerly Apache Spark Monthly Workshop)

Monday, Nov 13, 2017
- Tuesday, Nov 14, 2017

Time: 9:00am - 5:00pm

Thumbnail Image



Summary


This two-day course, created by Dean Wampler, Ph.D., is designed to teach developers how to implement data processing pipelines and analytics using Apache Spark . Developers will use hands-on exercises to learn the Spark Core, SQL/DataFrame, Streaming, and MLlib (machine learning) APIs. Developers will also learn about Spark internals and tips for improving application performance. Additional coverage includes integration with Mesos, Hadoop, and Reactive frameworks like Akka.

Duration


2 Days

Objectives


After having participated in this course you should:

  • Understand how to use the Spark Scala APIs to implement various data analytics algorithms for offline (batch-mode) and event-streaming applications

  • Understand Spark internals

  • Understand Spark performance considerations

  • Understand how to test and deploy Spark applications

  • Understand the basics of integrating Spark with Mesos, Hadoop, and Akka



Audience


Developers wishing to learn how to write data-centric applications using Spark.


Prerequisite



  • Experience with Scala, such as completion of Fast Track to Scala course

  • Experience with SQL, machine learning, and other Big Data tools will be helpful, but not required.



Setup Requirements




Program



  • Introduction – Why Spark

    • How Spark improves on Hadoop MapReduce

    • The core abstractions in Spark

    • What happens during a Spark job?

    • The Spark ecosystem

    • Deployment options

    • References for more information



  • Spark’s Core API

    • Resilient Distributed Datasets (RDD) and how they implement your job

    • Using the Spark Shell (interpreter) vs submitting Spark batch jobs

    • Using the Spark web console.

    • Reading and writing data files

    • Working with structured and unstructured data

    • Building data transformation pipelines

    • Spark under the hood: caching, checkpointing, partitioning, shuffling, etc.

    • Mastering the RDD API

    • Broadcast variables, accumulators



  • Spark SQL and DataFrames

    • Working with the DataFrame API for structured data

    • Working with SQL

    • Performance optimizations

    • Support for JSON and Parquet formats

    • Integration with Hadoop Hive



  • Processing events with Spark Streaming:

    • Working with time slices, “mini-batches”, of events

    • Working with moving windows of mini-batches

    • Reuse of code in batch-mode and streaming: the Lambda Architecture

    • Working with different streaming sources: sockets, file systems, Kafka, etc.

    • Resiliency and fault tolerance considerations

    • Stateful transformations (e.g., running statistics)



  • Other Spark-based Libraries:

    • MLlib for machine learning

    • Discussion of GraphX for graph algorithms, Tachyon for distributed caching, and BlinkDB for approximate queries



  • Deploying to clusters:

    • Spark’s clustering abstractions: cluster vs. client deployments, coarse-grained and fine-grained process management

    • Standalone mode

    • Mesos

    • Hadoop YARN

    • EC2

    • Cassandra rings



  • Using Spark with the Lightbend Reactive Platform:

    • Akka Streams and Spark Streaming



  • Conclusions


To Secure your seat for Fast Track to Scala Register Now! Enter Contact Details

Have questions about YoppWorks training, speak to one of the YoppWorks Training Team!


Event Location

Virtual – Live On Line

,

Event Fees:

$ 1200.00

REGISTER

Registration is closed for this event.

For more information or questions, please email:

sophie.alder@yoppworks.com

 

 

 

Show Buttons
Hide Buttons
Contact Us