Data Analytics with Spark Using Python (Addison-Wesley Data & Analytics Series)

Jeffrey Aven

  • 出版商: Addison-Wesley Professional
  • 出版日期: 2018-06-16
  • 售價: $1,575
  • 貴賓價: 9.5$1,496
  • 語言: 英文
  • 頁數: 320
  • 裝訂: Paperback
  • ISBN: 013484601X
  • ISBN-13: 9780134846019
  • 相關分類: Python 程式語言資料科學Spark

立即出貨 (庫存 < 4)

相關活動主題

Pearson s
20180917 python%e7%b0%a1%e4%b8%ad%e8%8b%b1%e6%96%87 small %e8%a4%87%e6%9c%ac1

商品描述

Solve Data Analytics Problems with Spark, PySpark, and Related Open Source Tools

Spark is at the heart of today’s Big Data revolution, helping data professionals supercharge efficiency and performance in a wide range of data processing and analytics tasks. In this guide, Big Data expert Jeffrey Aven covers all you need to know to leverage Spark, together with its extensions, subprojects, and wider ecosystem.

Aven combines a language-agnostic introduction to foundational Spark concepts with extensive programming examples utilizing the popular and intuitive PySpark development environment. This guide’s focus on Python makes it widely accessible to large audiences of data professionals, analysts, and developers—even those with little Hadoop or Spark experience.

Aven’s broad coverage ranges from basic to advanced Spark programming, and Spark SQL to machine learning. You’ll learn how to efficiently manage all forms of data with Spark: streaming, structured, semi-structured, and unstructured. Throughout, concise topic overviews quickly get you up to speed, and extensive hands-on exercises prepare you to solve real problems.

Coverage includes:
• Understand Spark’s evolving role in the Big Data and Hadoop ecosystems
• Create Spark clusters using various deployment modes
• Control and optimize the operation of Spark clusters and applications
• Master Spark Core RDD API programming techniques
• Extend, accelerate, and optimize Spark routines with advanced API platform constructs, including shared variables, RDD storage, and partitioning
• Efficiently integrate Spark with both SQL and nonrelational data stores
• Perform stream processing and messaging with Spark Streaming and Apache Kafka
• Implement predictive modeling with SparkR and Spark MLlib