The book provides a timely coverage of the paradigm of knowledge distillation—an efficient way of model compression. Knowledge distillation is positioned in a general setting of transfer learning, which effectively learns a lightweight student model from a large teacher model. The book covers a variety of training schemes, teacher–student architectures, and distillation algorithms. The book covers a wealth of topics including recent developments in vision and language learning, relational architectures, multi-task learning, and representative applications to image processing, computer vision, edge intelligence, and autonomous systems. The book is of relevance to a broad audience including researchers and practitioners active in the area of machine learning and pursuing fundamental and applied research in the area of advanced learning paradigms.
Serier
Computere og teknologi
Bedøm denne e-bog
Fortæl os, hvad du mener.
Oplysninger om læsning
Smartphones og tablets
Installer appen Google Play Bøger til Android og iPad/iPhone. Den synkroniserer automatisk med din konto og giver dig mulighed for at læse online eller offline, uanset hvor du er.
Bærbare og stationære computere
Du kan høre lydbøger, du har købt i Google Play via browseren på din computer.
e-læsere og andre enheder
Hvis du vil læse på e-ink-enheder som f.eks. Kobo-e-læsere, skal du downloade en fil og overføre den til din enhed. Følg den detaljerede vejledning i Hjælp for at overføre filerne til understøttede e-læsere.