Items related to Distributed Machine Learning and Gradient Optimization...

Distributed Machine Learning and Gradient Optimization (Big Data Management) - Softcover

 
9789811634222: Distributed Machine Learning and Gradient Optimization (Big Data Management)

Synopsis

This book presents the state of the art in distributed machine learning algorithms that are based on gradient optimization methods. In the big data era, large-scale datasets pose enormous challenges for the existing machine learning systems. As such, implementing machine learning algorithms in a distributed environment has become a key technology, and recent research has shown gradient-based iterative optimization to be an effective solution. Focusing on methods that can speed up large-scale gradient optimization through both algorithm optimizations and careful system implementations, the book introduces three essential techniques in designing a gradient optimization algorithm to train a distributed machine learning model: parallel strategy, data compression and synchronization protocol.

Written in a tutorial style, it covers a range of topics, from fundamental knowledge to a number of carefully designed algorithms and systems of distributed machine learning. It will appealto a broad audience in the field of machine learning, artificial intelligence, big data and database management.


"synopsis" may belong to another edition of this title.

About the Author

Jiawei Jiang obtained his PhD from Peking University 2018, advised by Prof. Bin Cui. His research interests include distributed machine learning, gradient optimization and automatic machine learning. He has served as a program committee member or reviewer for various international events, including SIGMOD, VLDB, ICDE, KDD, AAAI and TKDE. He was awarded the CCF Outstanding Doctoral Dissertation Award (2019) and ACM China Doctoral Dissertation Award (2018).

Bin Cui is a Professor at the School of EECS and Director of the Institute of Network Computing and Information Systems, at Peking University. His research interests include database system architectures, query and index techniques, and big data management and mining. He has published over 200 refereed papers at international conferences and in journals. Dr. Cui has served on the technical program committee of various international conferences, including SIGMOD, VLDB, ICDE and KDD, and as Vice PC Chair of ICDE 2011, Demo Co-Chair of ICDE 2014, Area Chair of VLDB 2014, PC Co-Chair of APWeb 2015 and WAIM 2016. He is currently a member of the trustee board of VLDB Endowment, is on the editorial board of the VLDB Journal, Distributed and Parallel Databases Journal, and Information Systems, and was formerly an associate editor of IEEE Transactions on Knowledge and Data Engineering (TKDE, 2009-2013). He was selected for a Microsoft Young Professorship award (MSRA 2008), CCF Young Scientist award (2009), Second Prize of Natural Science Award of MOE China (2014), and appointed a Cheung Kong distinguished Professor by the MOE in 2016.


From the Back Cover

This book presents the state of the art in distributed machine learning algorithms that are based on gradient optimization methods. In the big data era, large-scale datasets pose enormous challenges for the existing machine learning systems. As such, implementing machine learning algorithms in a distributed environment has become a key technology, and recent research has shown gradient-based iterative optimization to be an effective solution. Focusing on methods that can speed up large-scale gradient optimization through both algorithm optimizations and careful system implementations, the book introduces three essential techniques in designing a gradient optimization algorithm to train a distributed machine learning model: parallel strategy, data compression and synchronization protocol.

Written in a tutorial style, it covers a range of topics, from fundamental knowledge to a number of carefully designed algorithms and systems of distributed machine learning. It will appeal toa broad audience in the field of machine learning, artificial intelligence, big data and database management.

"About this title" may belong to another edition of this title.

  • PublisherSpringer
  • Publication date2023
  • ISBN 10 981163422X
  • ISBN 13 9789811634222
  • BindingPaperback
  • LanguageEnglish
  • Edition number1
  • Number of pages184

Buy New

View this item

US$ 16.09 shipping from United Kingdom to U.S.A.

Destination, rates & speeds

Other Popular Editions of the Same Title

9789811634192: Distributed Machine Learning and Gradient Optimization (Big Data Management)

Featured Edition

ISBN 10:  981163419X ISBN 13:  9789811634192
Publisher: Springer, 2022
Hardcover

Search results for Distributed Machine Learning and Gradient Optimization...

Stock Image

Jiang, Jiawei; Cui, Bin; Zhang, Ce
Published by Springer, 2023
ISBN 10: 981163422X ISBN 13: 9789811634222
New Softcover

Seller: Ria Christie Collections, Uxbridge, United Kingdom

Seller rating 5 out of 5 stars 5-star rating, Learn more about seller ratings

Condition: New. In. Seller Inventory # ria9789811634222_new

Contact seller

Buy New

US$ 178.18
Convert currency
Shipping: US$ 16.09
From United Kingdom to U.S.A.
Destination, rates & speeds

Quantity: Over 20 available

Add to basket

Seller Image

Jiawei Jiang
ISBN 10: 981163422X ISBN 13: 9789811634222
New Taschenbuch
Print on Demand

Seller: BuchWeltWeit Ludwig Meier e.K., Bergisch Gladbach, Germany

Seller rating 5 out of 5 stars 5-star rating, Learn more about seller ratings

Taschenbuch. Condition: Neu. This item is printed on demand - it takes 3-4 days longer - Neuware -This book presents the state of the art in distributed machine learning algorithms that are based on gradient optimization methods. In the big data era, large-scale datasets pose enormous challenges for the existing machine learning systems. As such, implementing machine learning algorithms in a distributed environment has become a key technology, and recent research has shown gradient-based iterative optimization to be an effective solution. Focusing on methods that can speed up large-scale gradient optimization through both algorithm optimizations and careful system implementations, the book introduces three essential techniques in designing a gradient optimization algorithm to train a distributed machine learning model: parallel strategy, data compression and synchronization protocol.Written in a tutorial style, it covers a range of topics, from fundamental knowledge to a number of carefully designed algorithms and systems of distributed machine learning. It will appealto a broad audience in the field of machine learning, artificial intelligence, big data and database management. 184 pp. Englisch. Seller Inventory # 9789811634222

Contact seller

Buy New

US$ 185.64
Convert currency
Shipping: US$ 25.83
From Germany to U.S.A.
Destination, rates & speeds

Quantity: 2 available

Add to basket

Seller Image

Jiang, Jiawei|Cui, Bin|Zhang, Ce
ISBN 10: 981163422X ISBN 13: 9789811634222
New Softcover
Print on Demand

Seller: moluna, Greven, Germany

Seller rating 5 out of 5 stars 5-star rating, Learn more about seller ratings

Condition: New. Dieser Artikel ist ein Print on Demand Artikel und wird nach Ihrer Bestellung fuer Sie gedruckt. This book presents the state of the art in distributed machine learning algorithms that are based on gradient optimization methods. In the big data era, large-scale datasets pose enormous challenges for the existing machine learning systems. As such, imp. Seller Inventory # 807960357

Contact seller

Buy New

US$ 158.78
Convert currency
Shipping: US$ 55.02
From Germany to U.S.A.
Destination, rates & speeds

Quantity: Over 20 available

Add to basket

Seller Image

Jiawei Jiang
ISBN 10: 981163422X ISBN 13: 9789811634222
New Taschenbuch

Seller: AHA-BUCH GmbH, Einbeck, Germany

Seller rating 5 out of 5 stars 5-star rating, Learn more about seller ratings

Taschenbuch. Condition: Neu. Druck auf Anfrage Neuware - Printed after ordering - This book presents the state of the art in distributed machine learning algorithms that are based on gradient optimization methods. In the big data era, large-scale datasets pose enormous challenges for the existing machine learning systems. As such, implementing machine learning algorithms in a distributed environment has become a key technology, and recent research has shown gradient-based iterative optimization to be an effective solution. Focusing on methods that can speed up large-scale gradient optimization through both algorithm optimizations and careful system implementations, the book introduces three essential techniques in designing a gradient optimization algorithm to train a distributed machine learning model: parallel strategy, data compression and synchronization protocol.Written in a tutorial style, it covers a range of topics, from fundamental knowledge to a number of carefully designed algorithms and systems of distributed machine learning. It will appealto a broad audience in the field of machine learning, artificial intelligence, big data and database management. Seller Inventory # 9789811634222

Contact seller

Buy New

US$ 188.45
Convert currency
Shipping: US$ 33.06
From Germany to U.S.A.
Destination, rates & speeds

Quantity: 1 available

Add to basket

Seller Image

Jiawei Jiang
ISBN 10: 981163422X ISBN 13: 9789811634222
New Taschenbuch

Seller: buchversandmimpf2000, Emtmannsberg, BAYE, Germany

Seller rating 5 out of 5 stars 5-star rating, Learn more about seller ratings

Taschenbuch. Condition: Neu. Neuware -This book presents the state of the art in distributed machine learning algorithms that are based on gradient optimization methods. In the big data era, large-scale datasets pose enormous challenges for the existing machine learning systems. As such, implementing machine learning algorithms in a distributed environment has become a key technology, and recent research has shown gradient-based iterative optimization to be an effective solution. Focusing on methods that can speed up large-scale gradient optimization through both algorithm optimizations and careful system implementations, the book introduces three essential techniques in designing a gradient optimization algorithm to train a distributed machine learning model: parallel strategy, data compression and synchronization protocol.Written in a tutorial style, it covers a range of topics, from fundamental knowledge to a number of carefully designed algorithms and systems of distributed machine learning. It will appealto a broad audience in the field of machine learning, artificial intelligence, big data and database management.Springer Verlag GmbH, Tiergartenstr. 17, 69121 Heidelberg 184 pp. Englisch. Seller Inventory # 9789811634222

Contact seller

Buy New

US$ 173.27
Convert currency
Shipping: US$ 61.77
From Germany to U.S.A.
Destination, rates & speeds

Quantity: 2 available

Add to basket

Stock Image

Jiang, Jiawei/ Cui, Bin/ Zhang, Ce
Published by Springer, 2023
ISBN 10: 981163422X ISBN 13: 9789811634222
New Paperback

Seller: Revaluation Books, Exeter, United Kingdom

Seller rating 5 out of 5 stars 5-star rating, Learn more about seller ratings

Paperback. Condition: Brand New. 180 pages. 9.25x6.10x0.39 inches. In Stock. Seller Inventory # x-981163422X

Contact seller

Buy New

US$ 266.60
Convert currency
Shipping: US$ 13.43
From United Kingdom to U.S.A.
Destination, rates & speeds

Quantity: 2 available

Add to basket