10% off all books and free delivery over £50
Buy from our bookstore and 25% of the cover price will be given to a school of your choice to buy more books. *15% of eBooks.

Advancements in Knowledge Distillation

View All Editions (1)

The selected edition of this book is not available to buy right now.
Add To Wishlist
Write A Review

About

Advancements in Knowledge Distillation Synopsis

The book provides a timely coverage of the paradigm of knowledge distillation-an efficient way of model compression. Knowledge distillation is positioned in a general setting of transfer learning, which effectively learns a lightweight student model from a large teacher model. The book covers a variety of training schemes, teacher-student architectures, and distillation algorithms. The book covers a wealth of topics including recent developments in vision and language learning, relational architectures, multi-task learning, and representative applications to image processing, computer vision, edge intelligence, and autonomous systems. The book is of relevance to a broad audience including researchers and practitioners active in the area of machine learning and pursuing fundamental and applied research in the area of advanced learning paradigms.

About This Edition

ISBN: 9783031320972
Publication date:
Author: Witold Pedrycz, ShyiMing Chen
Publisher: Springer an imprint of Springer International Publishing
Format: Paperback
Pagination: 232 pages
Series: Studies in Computational Intelligence
Genres: Artificial intelligence