Items related to Foundations of Deep Learning (Machine Learning: Foundations,...

Foundations of Deep Learning (Machine Learning: Foundations, Methodologies, and Applications) - Hardcover

 
9789811682322: Foundations of Deep Learning (Machine Learning: Foundations, Methodologies, and Applications)

Synopsis

Deep learning has significantly reshaped a variety of technologies, such as image processing, natural language processing, and audio processing. The excellent generalizability of deep learning is like a “cloud” to conventional complexity-based learning theory: the over-parameterization of deep learning makes almost all existing tools vacuous. This irreconciliation considerably undermines the confidence of deploying deep learning to security-critical areas, including autonomous vehicles and medical diagnosis, where small algorithmic mistakes can lead to fatal disasters. This book seeks to explaining the excellent generalizability, including generalization analysis via the size-independent complexity measures, the role of optimization in understanding the generalizability, and the relationship between generalizability and ethical/security issues. 

The efforts to understand the excellent generalizability are following two major paths: (1) developing size-independent complexity measures, which can evaluate the “effective” hypothesis complexity that can be learned, instead of the whole hypothesis space; and (2) modelling the learned hypothesis through stochastic gradient methods, the dominant optimizers in deep learning, via stochastic differential functions and the geometry of the associated loss functions. Related works discover that over-parameterization surprisingly bring many good properties to the loss functions. Rising concerns of deep learning are seen on the ethical and security issues, including privacy preservation and adversarial robustness. Related works also reveal an interplay between them and generalizability: a good generalizability usually means a good privacy-preserving ability; and more robust algorithms might have a worse generalizability.

 We expect readers can have a big picture of the current knowledge in deep learning theory, understand how the deep learning theory can guide new algorithm designing, and identify future research directions. Readers need knowledge of calculus, linear algebra, probability, statistics, and statistical learning theory.

"synopsis" may belong to another edition of this title.

About the Author


From the Back Cover

Deep learning has significantly reshaped a variety of technologies, such as image processing, natural language processing, and audio processing. The excellent generalizability of deep learning is like a “cloud” to conventional complexity-based learning theory: the over-parameterization of deep learning makes almost all existing tools vacuous. This irreconciliation considerably undermines the confidence of deploying deep learning to security-critical areas, including autonomous vehicles and medical diagnosis, where small algorithmic mistakes can lead to fatal disasters. This book seeks to explaining the excellent generalizability, including generalization analysis via the size-independent complexity measures, the role of optimization in understanding the generalizability, and the relationship between generalizability and ethical/security issues. 

The efforts to understand the excellent generalizability are following two major paths: (1) developing size-independent complexity measures, which can evaluate the “effective” hypothesis complexity that can be learned, instead of the whole hypothesis space; and (2) modelling the learned hypothesis through stochastic gradient methods, the dominant optimizers in deep learning, via stochastic differential functions and the geometry of the associated loss functions. Related works discover that over-parameterization surprisingly bring many good properties to the loss functions. Rising concerns of deep learning are seen on the ethical and security issues, including privacy preservation and adversarial robustness. Related works also reveal an interplay between them and generalizability: a good generalizability usually means a good privacy-preserving ability; and more robust algorithms might have a worse generalizability.

 

We expect readers can have a big picture of the current knowledge in deep learning theory, understand how the deep learning theory can guide new algorithm designing, and identify future research directions. Readers need knowledge of calculus, linear algebra, probability, statistics, and statistical learning theory.

"About this title" may belong to another edition of this title.

Search results for Foundations of Deep Learning (Machine Learning: Foundations,...

Stock Image

He, Fengxiang; Tao, Dacheng
Published by Springer, 2025
ISBN 10: 9811682321 ISBN 13: 9789811682322
New Hardcover

Seller: California Books, Miami, FL, U.S.A.

Seller rating 5 out of 5 stars 5-star rating, Learn more about seller ratings

Condition: New. Seller Inventory # I-9789811682322

Contact seller

Buy New

US$ 190.00
Convert currency
Shipping: FREE
Within U.S.A.
Destination, rates & speeds

Quantity: Over 20 available

Add to basket

Stock Image

He, Fengxiang; Tao, Dacheng
Published by Springer, 2025
ISBN 10: 9811682321 ISBN 13: 9789811682322
New Hardcover

Seller: Ria Christie Collections, Uxbridge, United Kingdom

Seller rating 5 out of 5 stars 5-star rating, Learn more about seller ratings

Condition: New. In. Seller Inventory # ria9789811682322_new

Contact seller

Buy New

US$ 177.90
Convert currency
Shipping: US$ 16.06
From United Kingdom to U.S.A.
Destination, rates & speeds

Quantity: Over 20 available

Add to basket

Seller Image

Fengxiang He|Dacheng Tao
Published by Springer Nature Singapore, 2024
ISBN 10: 9811682321 ISBN 13: 9789811682322
New Hardcover

Seller: moluna, Greven, Germany

Seller rating 5 out of 5 stars 5-star rating, Learn more about seller ratings

Condition: New. The first comprehensive overview book on the foundations of deep learningWritten by leading experts in the fieldExplicates excellent generalizability of deep learning, including generalization analysisDeep learning has si. Seller Inventory # 521847426

Contact seller

Buy New

US$ 147.93
Convert currency
Shipping: US$ 56.96
From Germany to U.S.A.
Destination, rates & speeds

Quantity: 1 available

Add to basket

Stock Image

Fengxiang He
ISBN 10: 9811682321 ISBN 13: 9789811682322
New Hardcover

Seller: Grand Eagle Retail, Mason, OH, U.S.A.

Seller rating 5 out of 5 stars 5-star rating, Learn more about seller ratings

Hardcover. Condition: new. Hardcover. Deep learning has significantly reshaped a variety of technologies, such as image processing, natural language processing, and audio processing. The excellent generalizability of deep learning is like a cloud to conventional complexity-based learning theory: the over-parameterization of deep learning makes almost all existing tools vacuous. This irreconciliation considerably undermines the confidence of deploying deep learning to security-critical areas, including autonomous vehicles and medical diagnosis, where small algorithmic mistakes can lead to fatal disasters. This book seeks to explaining the excellent generalizability, including generalization analysis via the size-independent complexity measures, the role of optimization in understanding the generalizability, and the relationship between generalizability and ethical/security issues. The efforts to understand the excellent generalizability are following two major paths: (1) developing size-independent complexity measures, which can evaluate the effective hypothesis complexity that can be learned, instead of the whole hypothesis space; and (2) modelling the learned hypothesis through stochastic gradient methods, the dominant optimizers in deep learning, via stochastic differential functions and the geometry of the associated loss functions. Related works discover that over-parameterization surprisingly bring many good properties to the loss functions. Rising concerns of deep learning are seen on the ethical and security issues, including privacy preservation and adversarial robustness. Related works also reveal an interplay between them and generalizability: a good generalizability usually means a good privacy-preserving ability; and more robust algorithms might have a worse generalizability. We expect readers can have a big picture of the current knowledge in deep learning theory, understand how the deep learning theory can guide new algorithm designing, and identify future research directions. Readers need knowledge of calculus, linear algebra, probability, statistics, and statistical learning theory. The excellent generalizability of deep learning is like a cloud to conventional complexity-based learning theory: the over-parameterization of deep learning makes almost all existing tools vacuous. Shipping may be from multiple locations in the US or from the UK, depending on stock availability. Seller Inventory # 9789811682322

Contact seller

Buy New

US$ 205.17
Convert currency
Shipping: FREE
Within U.S.A.
Destination, rates & speeds

Quantity: 1 available

Add to basket

Seller Image

Fengxiang He
ISBN 10: 9811682321 ISBN 13: 9789811682322
New Hardcover
Print on Demand

Seller: BuchWeltWeit Ludwig Meier e.K., Bergisch Gladbach, Germany

Seller rating 5 out of 5 stars 5-star rating, Learn more about seller ratings

Buch. Condition: Neu. This item is printed on demand - it takes 3-4 days longer - Neuware -Deep learning has significantly reshaped a variety of technologies, such as image processing, natural language processing, and audio processing. The excellent generalizability of deep learning is like a 'cloud' to conventional complexity-based learning theory: the over-parameterization of deep learning makes almost all existing tools vacuous. This irreconciliation considerably undermines the confidence of deploying deep learning to security-critical areas, including autonomous vehicles and medical diagnosis, where small algorithmic mistakes can lead to fatal disasters. This book seeks to explaining the excellent generalizability, including generalization analysis via the size-independent complexity measures, the role of optimization in understanding the generalizability, and the relationship between generalizability and ethical/security issues.The efforts to understand the excellent generalizability are following two major paths: (1) developing size-independent complexity measures, which can evaluate the 'effective' hypothesis complexity that can be learned, instead of the whole hypothesis space; and (2) modelling the learned hypothesis through stochastic gradient methods, the dominant optimizers in deep learning, via stochastic differential functions and the geometry of the associated loss functions. Related works discover that over-parameterization surprisingly bring many good properties to the loss functions. Rising concerns of deep learning are seen on the ethical and security issues, including privacy preservation and adversarial robustness. Related works also reveal an interplay between them and generalizability: a good generalizability usually means a good privacy-preserving ability; and more robust algorithms might have a worse generalizability.We expect readers can have a big picture of the current knowledge in deep learning theory, understand how the deep learning theory can guide new algorithm designing, and identify future research directions. Readers need knowledge of calculus, linear algebra, probability, statistics, and statistical learning theory. 284 pp. Englisch. Seller Inventory # 9789811682322

Contact seller

Buy New

US$ 179.37
Convert currency
Shipping: US$ 26.74
From Germany to U.S.A.
Destination, rates & speeds

Quantity: 2 available

Add to basket

Seller Image

Dacheng Tao
ISBN 10: 9811682321 ISBN 13: 9789811682322
New Hardcover

Seller: AHA-BUCH GmbH, Einbeck, Germany

Seller rating 5 out of 5 stars 5-star rating, Learn more about seller ratings

Buch. Condition: Neu. Druck auf Anfrage Neuware - Printed after ordering - Deep learning has significantly reshaped a variety of technologies, such as image processing, natural language processing, and audio processing. The excellent generalizability of deep learning is like a 'cloud' to conventional complexity-based learning theory: the over-parameterization of deep learning makes almost all existing tools vacuous. This irreconciliation considerably undermines the confidence of deploying deep learning to security-critical areas, including autonomous vehicles and medical diagnosis, where small algorithmic mistakes can lead to fatal disasters. This book seeks to explaining the excellent generalizability, including generalization analysis via the size-independent complexity measures, the role of optimization in understanding the generalizability, and the relationship between generalizability and ethical/security issues.The efforts to understand the excellent generalizability are following two major paths: (1) developing size-independent complexity measures, which can evaluate the 'effective' hypothesis complexity that can be learned, instead of the whole hypothesis space; and (2) modelling the learned hypothesis through stochastic gradient methods, the dominant optimizers in deep learning, via stochastic differential functions and the geometry of the associated loss functions. Related works discover that over-parameterization surprisingly bring many good properties to the loss functions. Rising concerns of deep learning are seen on the ethical and security issues, including privacy preservation and adversarial robustness. Related works also reveal an interplay between them and generalizability: a good generalizability usually means a good privacy-preserving ability; and more robust algorithms might have a worse generalizability.We expect readers can have a big picture of the current knowledge in deep learning theory, understand how the deep learning theory can guide new algorithm designing, and identify future research directions. Readers need knowledge of calculus, linear algebra, probability, statistics, and statistical learning theory. Seller Inventory # 9789811682322

Contact seller

Buy New

US$ 180.90
Convert currency
Shipping: US$ 36.19
From Germany to U.S.A.
Destination, rates & speeds

Quantity: 1 available

Add to basket

Stock Image

He, Fengxiang; Tao, Dacheng
Published by Springer, 2025
ISBN 10: 9811682321 ISBN 13: 9789811682322
New Hardcover

Seller: Books Puddle, New York, NY, U.S.A.

Seller rating 4 out of 5 stars 4-star rating, Learn more about seller ratings

Condition: New. 1st ed. 2022 edition NO-PA16APR2015-KAP. Seller Inventory # 26396346908

Contact seller

Buy New

US$ 230.88
Convert currency
Shipping: US$ 3.99
Within U.S.A.
Destination, rates & speeds

Quantity: 4 available

Add to basket

Stock Image

Fengxiang He
ISBN 10: 9811682321 ISBN 13: 9789811682322
New Hardcover

Seller: CitiRetail, Stevenage, United Kingdom

Seller rating 5 out of 5 stars 5-star rating, Learn more about seller ratings

Hardcover. Condition: new. Hardcover. Deep learning has significantly reshaped a variety of technologies, such as image processing, natural language processing, and audio processing. The excellent generalizability of deep learning is like a cloud to conventional complexity-based learning theory: the over-parameterization of deep learning makes almost all existing tools vacuous. This irreconciliation considerably undermines the confidence of deploying deep learning to security-critical areas, including autonomous vehicles and medical diagnosis, where small algorithmic mistakes can lead to fatal disasters. This book seeks to explaining the excellent generalizability, including generalization analysis via the size-independent complexity measures, the role of optimization in understanding the generalizability, and the relationship between generalizability and ethical/security issues. The efforts to understand the excellent generalizability are following two major paths: (1) developing size-independent complexity measures, which can evaluate the effective hypothesis complexity that can be learned, instead of the whole hypothesis space; and (2) modelling the learned hypothesis through stochastic gradient methods, the dominant optimizers in deep learning, via stochastic differential functions and the geometry of the associated loss functions. Related works discover that over-parameterization surprisingly bring many good properties to the loss functions. Rising concerns of deep learning are seen on the ethical and security issues, including privacy preservation and adversarial robustness. Related works also reveal an interplay between them and generalizability: a good generalizability usually means a good privacy-preserving ability; and more robust algorithms might have a worse generalizability. We expect readers can have a big picture of the current knowledge in deep learning theory, understand how the deep learning theory can guide new algorithm designing, and identify future research directions. Readers need knowledge of calculus, linear algebra, probability, statistics, and statistical learning theory. The excellent generalizability of deep learning is like a cloud to conventional complexity-based learning theory: the over-parameterization of deep learning makes almost all existing tools vacuous. Shipping may be from our UK warehouse or from our Australian or US warehouses, depending on stock availability. Seller Inventory # 9789811682322

Contact seller

Buy New

US$ 187.11
Convert currency
Shipping: US$ 49.61
From United Kingdom to U.S.A.
Destination, rates & speeds

Quantity: 1 available

Add to basket

Seller Image

Dacheng Tao
ISBN 10: 9811682321 ISBN 13: 9789811682322
New Hardcover

Seller: buchversandmimpf2000, Emtmannsberg, BAYE, Germany

Seller rating 5 out of 5 stars 5-star rating, Learn more about seller ratings

Buch. Condition: Neu. Neuware -Deep learning has significantly reshaped a variety of technologies, such as image processing, natural language processing, and audio processing. The excellent generalizability of deep learning is like a 'cloud' to conventional complexity-based learning theory: the over-parameterization of deep learning makes almost all existing tools vacuous. This irreconciliation considerably undermines the confidence of deploying deep learning to security-critical areas, including autonomous vehicles and medical diagnosis, where small algorithmic mistakes can lead to fatal disasters. This book seeks to explaining the excellent generalizability, including generalization analysis via the size-independent complexity measures, the role of optimization in understanding the generalizability, and the relationship between generalizability and ethical/security issues.Springer Verlag GmbH, Tiergartenstr. 17, 69121 Heidelberg 308 pp. Englisch. Seller Inventory # 9789811682322

Contact seller

Buy New

US$ 179.37
Convert currency
Shipping: US$ 63.94
From Germany to U.S.A.
Destination, rates & speeds

Quantity: 2 available

Add to basket

Stock Image

Fengxiang He
ISBN 10: 9811682321 ISBN 13: 9789811682322
New Hardcover

Seller: AussieBookSeller, Truganina, VIC, Australia

Seller rating 5 out of 5 stars 5-star rating, Learn more about seller ratings

Hardcover. Condition: new. Hardcover. Deep learning has significantly reshaped a variety of technologies, such as image processing, natural language processing, and audio processing. The excellent generalizability of deep learning is like a cloud to conventional complexity-based learning theory: the over-parameterization of deep learning makes almost all existing tools vacuous. This irreconciliation considerably undermines the confidence of deploying deep learning to security-critical areas, including autonomous vehicles and medical diagnosis, where small algorithmic mistakes can lead to fatal disasters. This book seeks to explaining the excellent generalizability, including generalization analysis via the size-independent complexity measures, the role of optimization in understanding the generalizability, and the relationship between generalizability and ethical/security issues. The efforts to understand the excellent generalizability are following two major paths: (1) developing size-independent complexity measures, which can evaluate the effective hypothesis complexity that can be learned, instead of the whole hypothesis space; and (2) modelling the learned hypothesis through stochastic gradient methods, the dominant optimizers in deep learning, via stochastic differential functions and the geometry of the associated loss functions. Related works discover that over-parameterization surprisingly bring many good properties to the loss functions. Rising concerns of deep learning are seen on the ethical and security issues, including privacy preservation and adversarial robustness. Related works also reveal an interplay between them and generalizability: a good generalizability usually means a good privacy-preserving ability; and more robust algorithms might have a worse generalizability. We expect readers can have a big picture of the current knowledge in deep learning theory, understand how the deep learning theory can guide new algorithm designing, and identify future research directions. Readers need knowledge of calculus, linear algebra, probability, statistics, and statistical learning theory. The excellent generalizability of deep learning is like a cloud to conventional complexity-based learning theory: the over-parameterization of deep learning makes almost all existing tools vacuous. Shipping may be from our Sydney, NSW warehouse or from our UK or US warehouse, depending on stock availability. Seller Inventory # 9789811682322

Contact seller

Buy New

US$ 212.23
Convert currency
Shipping: US$ 37.00
From Australia to U.S.A.
Destination, rates & speeds

Quantity: 1 available

Add to basket

There are 3 more copies of this book

View all search results for this book