Generative Deep Learning: Teaching Machines to Paint, Write, Compose, and Play

This document was uploaded by one of our users. The uploader already confirmed that they had the permission to publish it. If you are author/publisher or own the copyright of this documents, please report to us by using this DMCA report form.

Simply click on the Download Book button.

Yes, Book downloads on Ebookily are 100% Free.

Sometimes the book is free on Amazon As well, so go ahead and hit "Search on Amazon"

Generative modeling is one of the hottest topics in AI. It’s now possible to teach a machine to excel at human endeavors such as painting, writing, and composing music. With this practical book, machine-learning engineers and data scientists will discover how to re-create some of the most impressive examples of generative deep learning models, such as variational autoencoders,generative adversarial networks (GANs), encoder-decoder models and world models. Author David Foster demonstrates the inner workings of each technique, starting with the basics of deep learning before advancing to some of the most cutting-edge algorithms in the field. Through tips and tricks, you’ll understand how to make your models learn more efficiently and become more creative. • Discover how variational autoencoders can change facial expressions in photos • Build practical GAN examples from scratch, including CycleGAN for style transfer and MuseGAN for music generation • Create recurrent generative models for text generation and learn how to improve the models using attention • Understand how generative models can help agents to accomplish tasks within a reinforcement learning setting • Explore the architecture of the Transformer (BERT, GPT-2) and image generation models such as ProGAN and StyleGAN

Author(s): David Foster
Edition: 1
Publisher: O’Reilly Media
Year: 2019

Language: English
Commentary: EPUB
Pages: 330
City: Sebastopol, CA
Tags: Machine Learning; Neural Networks; Deep Learning; Natural Language Processing; Image Processing; Python; Convolutional Neural Networks; Recurrent Neural Networks; Generative Adversarial Networks; Keras; TensorFlow; Naive Bayes; Variational Autoencoders; Batch Learning; Music