# coding: utf-8 #from google.colab import drive import tensorflow as tf import tensorflow_datasets as tfds import numpy as np import matplotlib.pyplot as plt import time import itertools # *Python Machine Learning 3rd Edition* by [Sebastian Raschka](https://sebastianraschka.com) & [Vahid Mirjalili](http://vahidmirjalili.com), Packt Publishing Ltd. 2019 # # Code Repository: https://github.com/rasbt/python-machine-learning-book-3rd-edition # # Code License: [MIT License](https://github.com/rasbt/python-machine-learning-book-3rd-edition/blob/master/LICENSE.txt) # # Chapter 17 - Generative Adversarial Networks for Synthesizing New Data (Part 2/2) # Note that the optional watermark extension is a small IPython notebook plugin that I developed to make the code reproducible. You can just skip the following line(s). # # Improving the quality of synthesized images using a convolutional and Wasserstein GAN # ## Transposed convolution # ## Batch normalization # ## Implementing the generator and discriminator # * **Setting up the Google Colab** #! pip install -q tensorflow-gpu==2.0.0-beta1 #drive.mount('/content/drive/') print(tf.__version__) print("GPU Available:", tf.test.is_gpu_available()) if tf.test.is_gpu_available(): device_name = tf.test.gpu_device_name() else: device_name = 'CPU:0' print(device_name) def make_dcgan_generator( z_size=20, output_size=(28, 28, 1), n_filters=128, n_blocks=2): size_factor = 2**n_blocks hidden_size = ( output_size[0]//size_factor, output_size[1]//size_factor ) model = tf.keras.Sequential([ tf.keras.layers.Input(shape=(z_size,)), tf.keras.layers.Dense( units=n_filters*np.prod(hidden_size), use_bias=False), tf.keras.layers.BatchNormalization(), tf.keras.layers.LeakyReLU(), tf.keras.layers.Reshape( (hidden_size[0], hidden_size[1], n_filters)), tf.keras.layers.Conv2DTranspose( filters=n_filters, kernel_size=(5, 5), strides=(1, 1), padding='same', use_bias=False), tf.keras.layers.BatchNormalization(), tf.keras.layers.LeakyReLU() ]) nf = n_filters for i in range(n_blocks): nf = nf // 2 model.add( tf.keras.layers.Conv2DTranspose( filters=nf, kernel_size=(5, 5), strides=(2, 2), padding='same', use_bias=False)) model.add(tf.keras.layers.BatchNormalization()) model.add(tf.keras.layers.LeakyReLU()) model.add( tf.keras.layers.Conv2DTranspose( filters=output_size[2], kernel_size=(5, 5), strides=(1, 1), padding='same', use_bias=False, activation='tanh')) return model def make_dcgan_discriminator( input_size=(28, 28, 1), n_filters=64, n_blocks=2): model = tf.keras.Sequential([ tf.keras.layers.Input(shape=input_size), tf.keras.layers.Conv2D( filters=n_filters, kernel_size=5, strides=(1, 1), padding='same'), tf.keras.layers.BatchNormalization(), tf.keras.layers.LeakyReLU() ]) nf = n_filters for i in range(n_blocks): nf = nf*2 model.add( tf.keras.layers.Conv2D( filters=nf, kernel_size=(5, 5), strides=(2, 2),padding='same')) model.add(tf.keras.layers.BatchNormalization()) model.add(tf.keras.layers.LeakyReLU()) model.add(tf.keras.layers.Dropout(0.3)) model.add(tf.keras.layers.Conv2D( filters=1, kernel_size=(7, 7), padding='valid')) model.add(tf.keras.layers.Reshape((1,))) return model gen_model = make_dcgan_generator() gen_model.summary() disc_model = make_dcgan_discriminator() disc_model.summary() # ## Dissimilarity measures between two distributions # ## Using EM distance in practice for GANs # ## Gradient penalty # ## Implementing WGAN-GP to train the DCGAN model mnist_bldr = tfds.builder('mnist') mnist_bldr.download_and_prepare() mnist = mnist_bldr.as_dataset(shuffle_files=False) def preprocess(ex, mode='uniform'): image = ex['image'] image = tf.image.convert_image_dtype(image, tf.float32) image = image*2 - 1.0 if mode == 'uniform': input_z = tf.random.uniform( shape=(z_size,), minval=-1.0, maxval=1.0) elif mode == 'normal': input_z = tf.random.normal(shape=(z_size,)) return input_z, image num_epochs = 100 batch_size = 128 image_size = (28, 28) z_size = 20 mode_z = 'uniform' lambda_gp = 10.0 tf.random.set_seed(1) np.random.seed(1) ## Set-up the dataset mnist_trainset = mnist['train'] mnist_trainset = mnist_trainset.map(preprocess) mnist_trainset = mnist_trainset.shuffle(10000) mnist_trainset = mnist_trainset.batch( batch_size, drop_remainder=True) ## Set-up the model with tf.device(device_name): gen_model = make_dcgan_generator() gen_model.build(input_shape=(None, z_size)) gen_model.summary() disc_model = make_dcgan_discriminator() disc_model.build(input_shape=(None, np.prod(image_size))) disc_model.summary() ## optimizers: g_optimizer = tf.keras.optimizers.Adam(0.0002) d_optimizer = tf.keras.optimizers.Adam(0.0002) if mode_z == 'uniform': fixed_z = tf.random.uniform( shape=(batch_size, z_size), minval=-1, maxval=1) elif mode_z == 'normal': fixed_z = tf.random.normal( shape=(batch_size, z_size)) def create_samples(g_model, input_z): g_output = g_model(input_z, training=False) images = tf.reshape(g_output, (batch_size, *image_size)) return (images+1)/2.0 all_losses = [] epoch_samples = [] start_time = time.time() for epoch in range(1, num_epochs+1): epoch_losses = [] for i,(input_z,input_real) in enumerate(mnist_trainset): ## Compute discriminator's loss and gradients: with tf.GradientTape() as d_tape, tf.GradientTape() as g_tape: g_output = gen_model(input_z, training=True) d_critics_real = disc_model(input_real, training=True) d_critics_fake = disc_model(g_output, training=True) ## Compute generator's loss: g_loss = -tf.math.reduce_mean(d_critics_fake) ## Compute discriminator's losses d_loss_real = -tf.math.reduce_mean(d_critics_real) d_loss_fake = tf.math.reduce_mean(d_critics_fake) d_loss = d_loss_real + d_loss_fake ## Gradient penalty: with tf.GradientTape() as gp_tape: alpha = tf.random.uniform( shape=[d_critics_real.shape[0], 1, 1, 1], minval=0.0, maxval=1.0) interpolated = ( alpha*input_real + (1-alpha)*g_output) gp_tape.watch(interpolated) d_critics_intp = disc_model(interpolated) grads_intp = gp_tape.gradient( d_critics_intp, [interpolated,])[0] grads_intp_l2 = tf.sqrt( tf.reduce_sum(tf.square(grads_intp), axis=[1, 2, 3])) grad_penalty = tf.reduce_mean(tf.square(grads_intp_l2 - 1.0)) d_loss = d_loss + lambda_gp*grad_penalty ## Optimization: Compute the gradients apply them d_grads = d_tape.gradient(d_loss, disc_model.trainable_variables) d_optimizer.apply_gradients( grads_and_vars=zip(d_grads, disc_model.trainable_variables)) g_grads = g_tape.gradient(g_loss, gen_model.trainable_variables) g_optimizer.apply_gradients( grads_and_vars=zip(g_grads, gen_model.trainable_variables)) epoch_losses.append( (g_loss.numpy(), d_loss.numpy(), d_loss_real.numpy(), d_loss_fake.numpy())) all_losses.append(epoch_losses) print('Epoch {:-3d} | ET {:.2f} min | Avg Losses >>' ' G/D {:6.2f}/{:6.2f} [D-Real: {:6.2f} D-Fake: {:6.2f}]' .format(epoch, (time.time() - start_time)/60, *list(np.mean(all_losses[-1], axis=0))) ) epoch_samples.append( create_samples(gen_model, fixed_z).numpy() ) #import pickle #pickle.dump({'all_losses':all_losses, # 'samples':epoch_samples}, # open('/content/drive/My Drive/Colab Notebooks/PyML-3rd-edition/ch17-WDCGAN-learning.pkl', 'wb')) #gen_model.save('/content/drive/My Drive/Colab Notebooks/PyML-3rd-edition/ch17-WDCGAN-gan_gen.h5') #disc_model.save('/content/drive/My Drive/Colab Notebooks/PyML-3rd-edition/ch17-WDCGAN-gan_disc.h5') fig = plt.figure(figsize=(8, 6)) ## Plotting the losses ax = fig.add_subplot(1, 1, 1) g_losses = [item[0] for item in itertools.chain(*all_losses)] d_losses = [item[1] for item in itertools.chain(*all_losses)] plt.plot(g_losses, label='Generator loss', alpha=0.95) plt.plot(d_losses, label='Discriminator loss', alpha=0.95) plt.legend(fontsize=20) ax.set_xlabel('Iteration', size=15) ax.set_ylabel('Loss', size=15) epochs = np.arange(1, 101) epoch2iter = lambda e: e*len(all_losses[-1]) epoch_ticks = [1, 20, 40, 60, 80, 100] newpos = [epoch2iter(e) for e in epoch_ticks] ax2 = ax.twiny() ax2.set_xticks(newpos) ax2.set_xticklabels(epoch_ticks) ax2.xaxis.set_ticks_position('bottom') ax2.xaxis.set_label_position('bottom') ax2.spines['bottom'].set_position(('outward', 60)) ax2.set_xlabel('Epoch', size=15) ax2.set_xlim(ax.get_xlim()) ax.tick_params(axis='both', which='major', labelsize=15) ax2.tick_params(axis='both', which='major', labelsize=15) #plt.savefig('images/ch17-wdcgan-learning-curve.pdf') plt.show() selected_epochs = [1, 2, 4, 10, 50, 100] fig = plt.figure(figsize=(10, 14)) for i,e in enumerate(selected_epochs): for j in range(5): ax = fig.add_subplot(6, 5, i*5+j+1) ax.set_xticks([]) ax.set_yticks([]) if j == 0: ax.text( -0.06, 0.5, 'Epoch {}'.format(e), rotation=90, size=18, color='red', horizontalalignment='right', verticalalignment='center', transform=ax.transAxes) image = epoch_samples[e-1][j] ax.imshow(image, cmap='gray_r') #plt.savefig('images/ch17-wdcgan-samples.pdf') plt.show() # ## Mode collapse # # ---- # # # Readers may ignore the next cell. # #