The Azure Data Lakehouse Toolkit

Ron L'Esteve

ISBN 10: 1484282329 ISBN 13: 9781484282328
Published by APress, US, 2022
New Paperback

From Rarewaves USA, OSWEGO, IL, U.S.A. Seller rating 5 out of 5 stars 5-star rating, Learn more about seller ratings

AbeBooks Seller since June 10, 2025

This specific item is no longer available.

About this Item

Description:

Design and implement a modern data lakehouse on the Azure Data Platform using Delta Lake, Apache Spark, Azure Databricks, Azure Synapse Analytics, and Snowflake. This book teaches you the intricate details of the Data Lakehouse Paradigm and how to efficiently design a cloud-based data lakehouse using highly performant and cutting-edge Apache Spark capabilities using Azure Databricks, Azure Synapse Analytics, and Snowflake. You will learn to write efficient PySpark code for batch and streaming ELT jobs on Azure. And you will follow along with practical, scenario-based examples showing how to apply the capabilities of Delta Lake and Apache Spark to optimize performance, and secure, share, and manage a high volume, high velocity, and high variety of data in your lakehouse with ease.The patterns of success that you acquire from reading this book will help you hone your skills to build high-performing and scalable ACID-compliant lakehouses using flexible and cost-efficient decoupled storage and compute capabilities. Extensive coverage of Delta Lake ensures that you are aware of and can benefit from all that this new, open source storage layer can offer. In addition to the deep examples on Databricks in the book, there is coverage of alternative platforms such as Synapse Analytics and Snowflake so that you can make the right platform choice for your needs.After reading this book, you will be able to implement Delta Lake capabilities, including Schema Evolution, Change Feed, Live Tables, Sharing, and Clones to enable better business intelligence and advanced analytics on your data within the Azure Data Platform.What You Will LearnImplement the Data Lakehouse Paradigm on Microsoft's Azure cloud platformBenefit from the new Delta Lake open-source storage layer for data lakehouses Take advantage of schema evolution, change feeds, live tables, and moreWritefunctional PySpark code for data lakehouse ELT jobsOptimize Apache Spark performance through partitioning, indexing, and other tuning optionsChoose between alternatives such as Databricks, Synapse Analytics, and SnowflakeWho This Book Is ForData, analytics, and AI professionals at all levels, including data architect and data engineer practitioners. Also for data professionals seeking patterns of success by which to remain relevant as they learn to build scalable data lakehouses for their organizations and customers who are migrating into the modern Azure Data Platform. Seller Inventory # LU-9781484282328

Report this item

Synopsis:

Design and implement a modern data lakehouse on the Azure Data Platform using Delta Lake, Apache Spark, Azure Databricks, Azure Synapse Analytics, and Snowflake. This book teaches you the intricate details of the Data Lakehouse Paradigm and how to efficiently design a cloud-based data lakehouse using highly performant and cutting-edge Apache Spark capabilities using Azure Databricks, Azure Synapse Analytics, and Snowflake. You will learn to write efficient PySpark code for batch and streaming ELT jobs on Azure. And you will follow along with practical, scenario-based examples showing how to apply the capabilities of Delta Lake and Apache Spark to optimize performance, and secure, share, and manage a high volume, high velocity, and high variety of data in your lakehouse with ease.
The patterns of success that you acquire from reading this book will help you hone your skills to build high-performing and scalable ACID-compliant lakehouses using flexible and cost-efficient decoupled storage and compute capabilities. Extensive coverage of Delta Lake ensures that you are aware of and can benefit from all that this new, open source storage layer can offer. In addition to the deep examples on Databricks in the book, there is coverage of alternative platforms such as Synapse Analytics and Snowflake so that you can make the right platform choice for your needs.
After reading this book, you will be able to implement Delta Lake capabilities, including Schema Evolution, Change Feed, Live Tables, Sharing, and Clones to enable better business intelligence and advanced analytics on your data within the Azure Data Platform.
 
What You Will Learn

  • Implement the Data Lakehouse Paradigm on Microsoft's Azure cloud platform
  • Benefit from the new Delta Lake open-source storage layer for data lakehouses 
  • Take advantage of schema evolution, change feeds, live tables, and more
  • Write functional PySpark code for data lakehouse ELT jobs
  • Optimize Apache Spark performance through partitioning, indexing, and other tuning options
  • Choose between alternatives such as Databricks, Synapse Analytics, and Snowflake
 
Who This Book Is For
 
Data, analytics, and AI professionals at all levels, including data architect and data engineer practitioners. Also for data professionals seeking patterns of success by which to remain relevant as they learn to build scalable data lakehouses for their organizations and customers who are migrating into the modern Azure Data Platform. 
 
Source Code Repository
 
The complete source code for this book can be found within the following GitHub respository:

github.com/Apress/the-azure-data-lakehouse-toolkit

About the Author: Ron C. L’Esteve is a professional author, trusted technology leader, and digital innovation strategist residing in Chicago, IL, USA. He is well-known for his impactful books and award-winning article publications about Azure Data & AI Architecture and Engineering. He possesses deep technical skills and experience in designing, implementing, and delivering modern Azure Data & AI projects for numerous clients around the world.

Having several Azure Data, AI, and Lakehouse certifications under his belt, Ron has been a go-to technical advisor for some of the largest and most impactful Azure implementation projects on the planet. He has been responsible for scaling key data architectures, defining the road map and strategy for the future of data and business intelligence needs, and challenging customers to grow by thoroughly understanding the fluid business opportunities and enabling change by translating them into high-quality and sustainable technical solutionsthat solve the most complex challenges and promote digital innovation and transformation.

Ron is a gifted presenter and trainer, known for his innate ability to clearly articulate and explain complex topics to audiences of all skill levels. He applies a practical and business-oriented approach by taking transformational ideas from concept to scale. He is a true enabler of positive and impactful change by championing a growth mindset.

 


"About this title" may belong to another edition of this title.

Bibliographic Details

Title: The Azure Data Lakehouse Toolkit
Publisher: APress, US
Publication Date: 2022
Binding: Paperback
Condition: New
Edition: 1st ed.

Top Search Results from the AbeBooks Marketplace

Stock Image

Ron L'Esteve
Published by APress, Berkley, 2022
ISBN 10: 1484282329 ISBN 13: 9781484282328
New Paperback First Edition

Seller: Grand Eagle Retail, Bensenville, IL, U.S.A.

Seller rating 5 out of 5 stars 5-star rating, Learn more about seller ratings

Paperback. Condition: new. Paperback. Design and implement a modern data lakehouse on the Azure Data Platform using Delta Lake, Apache Spark, Azure Databricks, Azure Synapse Analytics, and Snowflake. This book teaches you the intricate details of the Data Lakehouse Paradigm and how to efficiently design a cloud-based data lakehouse using highly performant and cutting-edge Apache Spark capabilities using Azure Databricks, Azure Synapse Analytics, and Snowflake. You will learn to write efficient PySpark code for batch and streaming ELT jobs on Azure. And you will follow along with practical, scenario-based examples showing how to apply the capabilities of Delta Lake and Apache Spark to optimize performance, and secure, share, and manage a high volume, high velocity, and high variety of data in your lakehouse with ease.The patterns of success that you acquire from reading this book will help you hone your skills to build high-performing and scalable ACID-compliant lakehouses using flexible and cost-efficient decoupled storage and compute capabilities. Extensive coverage of Delta Lake ensures that you are aware of and can benefit from all that this new, open source storage layer can offer. In addition to the deep examples on Databricks in the book, there is coverage of alternative platforms such as Synapse Analytics and Snowflake so that you can make the right platform choice for your needs.After reading this book, you will be able to implement Delta Lake capabilities, including Schema Evolution, Change Feed, Live Tables, Sharing, and Clones to enable better business intelligence and advanced analytics on your data within the Azure Data Platform.What You Will LearnImplement the Data Lakehouse Paradigm on Microsofts Azure cloud platformBenefit from the new Delta Lake open-source storage layer for data lakehouses Take advantage of schema evolution, change feeds, live tables, and moreWritefunctional PySpark code for data lakehouse ELT jobsOptimize Apache Spark performance through partitioning, indexing, and other tuning optionsChoose between alternatives such as Databricks, Synapse Analytics, and SnowflakeWho This Book Is ForData, analytics, and AI professionals at all levels, including data architect and data engineer practitioners. Also for data professionals seeking patterns of success by which to remain relevant as they learn to build scalable data lakehouses for their organizations and customers who are migrating into the modern Azure Data Platform. Intermediate user level Shipping may be from multiple locations in the US or from the UK, depending on stock availability. Seller Inventory # 9781484282328

Contact seller

Buy New

US$ 57.49
Convert currency
Shipping: FREE
Within U.S.A.
Destination, rates & speeds

Quantity: 1 available

Add to basket

Stock Image

L?Esteve, Ron
Published by Apress Publishers, 2022
ISBN 10: 1484282329 ISBN 13: 9781484282328
New Softcover First Edition

Seller: Kennys Bookshop and Art Galleries Ltd., Galway, GY, Ireland

Seller rating 5 out of 5 stars 5-star rating, Learn more about seller ratings

Condition: New. 2022. 1st ed. paperback. . . . . . Seller Inventory # V9781484282328

Contact seller

Buy New

US$ 70.99
Convert currency
Shipping: US$ 12.19
From Ireland to U.S.A.
Destination, rates & speeds

Quantity: 15 available

Add to basket

Stock Image

Ron L'Esteve
Published by APress, Berkley, 2022
ISBN 10: 1484282329 ISBN 13: 9781484282328
New Paperback First Edition

Seller: CitiRetail, Stevenage, United Kingdom

Seller rating 5 out of 5 stars 5-star rating, Learn more about seller ratings

Paperback. Condition: new. Paperback. Design and implement a modern data lakehouse on the Azure Data Platform using Delta Lake, Apache Spark, Azure Databricks, Azure Synapse Analytics, and Snowflake. This book teaches you the intricate details of the Data Lakehouse Paradigm and how to efficiently design a cloud-based data lakehouse using highly performant and cutting-edge Apache Spark capabilities using Azure Databricks, Azure Synapse Analytics, and Snowflake. You will learn to write efficient PySpark code for batch and streaming ELT jobs on Azure. And you will follow along with practical, scenario-based examples showing how to apply the capabilities of Delta Lake and Apache Spark to optimize performance, and secure, share, and manage a high volume, high velocity, and high variety of data in your lakehouse with ease.The patterns of success that you acquire from reading this book will help you hone your skills to build high-performing and scalable ACID-compliant lakehouses using flexible and cost-efficient decoupled storage and compute capabilities. Extensive coverage of Delta Lake ensures that you are aware of and can benefit from all that this new, open source storage layer can offer. In addition to the deep examples on Databricks in the book, there is coverage of alternative platforms such as Synapse Analytics and Snowflake so that you can make the right platform choice for your needs.After reading this book, you will be able to implement Delta Lake capabilities, including Schema Evolution, Change Feed, Live Tables, Sharing, and Clones to enable better business intelligence and advanced analytics on your data within the Azure Data Platform.What You Will LearnImplement the Data Lakehouse Paradigm on Microsofts Azure cloud platformBenefit from the new Delta Lake open-source storage layer for data lakehouses Take advantage of schema evolution, change feeds, live tables, and moreWritefunctional PySpark code for data lakehouse ELT jobsOptimize Apache Spark performance through partitioning, indexing, and other tuning optionsChoose between alternatives such as Databricks, Synapse Analytics, and SnowflakeWho This Book Is ForData, analytics, and AI professionals at all levels, including data architect and data engineer practitioners. Also for data professionals seeking patterns of success by which to remain relevant as they learn to build scalable data lakehouses for their organizations and customers who are migrating into the modern Azure Data Platform. Intermediate user level Shipping may be from our UK warehouse or from our Australian or US warehouses, depending on stock availability. Seller Inventory # 9781484282328

Contact seller

Buy New

US$ 74.24
Convert currency
Shipping: US$ 49.40
From United Kingdom to U.S.A.
Destination, rates & speeds

Quantity: 1 available

Add to basket

Stock Image

Ron L'Esteve
Published by APress, Berkley, 2022
ISBN 10: 1484282329 ISBN 13: 9781484282328
New Paperback First Edition

Seller: AussieBookSeller, Truganina, VIC, Australia

Seller rating 5 out of 5 stars 5-star rating, Learn more about seller ratings

Paperback. Condition: new. Paperback. Design and implement a modern data lakehouse on the Azure Data Platform using Delta Lake, Apache Spark, Azure Databricks, Azure Synapse Analytics, and Snowflake. This book teaches you the intricate details of the Data Lakehouse Paradigm and how to efficiently design a cloud-based data lakehouse using highly performant and cutting-edge Apache Spark capabilities using Azure Databricks, Azure Synapse Analytics, and Snowflake. You will learn to write efficient PySpark code for batch and streaming ELT jobs on Azure. And you will follow along with practical, scenario-based examples showing how to apply the capabilities of Delta Lake and Apache Spark to optimize performance, and secure, share, and manage a high volume, high velocity, and high variety of data in your lakehouse with ease.The patterns of success that you acquire from reading this book will help you hone your skills to build high-performing and scalable ACID-compliant lakehouses using flexible and cost-efficient decoupled storage and compute capabilities. Extensive coverage of Delta Lake ensures that you are aware of and can benefit from all that this new, open source storage layer can offer. In addition to the deep examples on Databricks in the book, there is coverage of alternative platforms such as Synapse Analytics and Snowflake so that you can make the right platform choice for your needs.After reading this book, you will be able to implement Delta Lake capabilities, including Schema Evolution, Change Feed, Live Tables, Sharing, and Clones to enable better business intelligence and advanced analytics on your data within the Azure Data Platform.What You Will LearnImplement the Data Lakehouse Paradigm on Microsofts Azure cloud platformBenefit from the new Delta Lake open-source storage layer for data lakehouses Take advantage of schema evolution, change feeds, live tables, and moreWritefunctional PySpark code for data lakehouse ELT jobsOptimize Apache Spark performance through partitioning, indexing, and other tuning optionsChoose between alternatives such as Databricks, Synapse Analytics, and SnowflakeWho This Book Is ForData, analytics, and AI professionals at all levels, including data architect and data engineer practitioners. Also for data professionals seeking patterns of success by which to remain relevant as they learn to build scalable data lakehouses for their organizations and customers who are migrating into the modern Azure Data Platform. Intermediate user level Shipping may be from our Sydney, NSW warehouse or from our UK or US warehouse, depending on stock availability. Seller Inventory # 9781484282328

Contact seller

Buy New

US$ 96.65
Convert currency
Shipping: US$ 37.00
From Australia to U.S.A.
Destination, rates & speeds

Quantity: 1 available

Add to basket