Final Lab Manual
Final Lab Manual
List of Practicals
410251 : Deep Learning
Group 1
1) Linear regression by using Deep Neural network: Implement Boston housing price prediction problem by
Linear regression using Deep Neural network. Use Boston House price prediction dataset.
2) Classification using Deep neural network: Binary classification using Deep Neural Networks Example:
Classify movie reviews into positive" reviews and "negative" reviews, just based on the text content of the
reviews. Use IMDB dataset
3) Convolutional neural network (CNN): Use MNIST Fashion Dataset and create a classifier to classify fashion
clothing into categories.
Group 2
4) Mini Project: Human Face Recognition
Group 1
1
EXPERIMENT 1
Course Objective:
To illustrate the concepts of DNN.
Problem Statement:
Implement Boston housing price prediction problem by Linear regression using Deep Neural
network. Use Boston House price prediction dataset.
Course Outcomes:
Apply the technique of Deep Neural network for implementing Linear regression.
Software-Hardware requirements:
Python, Ubuntu Operating System 14.0 or above, 2 CPU CoreProcessor, 2GB RAM.
THEORY:
Linear Regression:
Linear Regression is usually the first machine learning algorithm that every data scientist comes
across. It is a simple model but everyone needs to master it as it lays the foundation for other machine
learning algorithms.
The objective of a linear regression model is to find a relationship between one or more
features(independent variables) and a continuous target variable(dependent variable). When there is
only feature it is called Uni-variate Linear Regression and if there are multiple features, it is called
Multiple Linear Regression.
2
Hypothesis of Linear Regression
The linear regression model can be represented by the following equation
where
θ is the model’s parameter vector including the bias term θ₀
x is the feature vector with x₀ =1
Data-set
Let’s create some random data-set to train our model.
#imports
import numpy as np
import matplotlib.pyplot as plt
3
# plot
plt.scatter(x,y,s=10)
plt.xlabel('x')
plt.ylabel('y')
plt.show()
The plot for the data set generated using the above code is shown below:
4
To define and measure the error of our model we define the cost function as the sum of the squares
of the residuals. The cost function is denoted by
DATASET: lib.stat.cmu.edu/datasets/boston
IMPLEMENTATION:
Boston Housing Data: This dataset was taken from the StatLib library and is maintained by
Carnegie Mellon University. This dataset concerns the housing prices in the housing city of
Boston. The dataset provided has 506 instances with 13 features.
The Description of the dataset is taken from the below reference as shown in the table
follows:
5
Make the Linear Regression Model, predicting housing prices by Inputting Libraries
and datasets.
Step 1: import the required libraries.
# Importing Libraries
import numpy as np
import pandas as pd
import matplotlib.pyplot as plt
# Importing Data
from sklearn.datasets import load_boston
boston = load_boston()
boston.feature_names
Step 3: Converting data from nd-array to data frame and adding feature names to the
data
data = pd.DataFrame(boston.data)
6
data.columns = boston.feature_names
data.head(10)
data['Price'] = boston.target
data.head()
7
Step 6: Info of Boston Dataset
data.info()
Step 7: Getting input and output data and further splitting data to training and
testing dataset.
# Input Data
x = boston.data
# Output Data
y = boston.target
8
# splitting data to training and testing dataset.
Step 8: Applying Linear Regression Model to the dataset and predicting the prices.
# Fitting Multi Linear regression model to training model
from sklearn.linear_model import LinearRegression
regressor = LinearRegression()
regressor.fit(xtrain, ytrain)
9
Step 10: Results of Linear Regression i.e. Mean Squared Error and Mean Absolute
Error.
from sklearn.metrics import mean_squared_error, mean_absolute_error
mse = mean_squared_error(ytest, y_pred)
mae = mean_absolute_error(ytest,y_pred)
print("Mean Square Error : ", mse)
print("Mean Absolute Error : ", mae)
As per the result, our model is only 66.55% accurate. So, the prepared model is not
very good for predicting housing prices. One can improve the prediction results using
many other possible machine learning algorithms and techniques.
CONCLUSION:
10
EXPERIMENT 2
Objective:
To classify movie reviews as positive or negative
Problem Statement:
Binary classification using Deep Neural Networks Example: Classify movie reviews into
positive" reviews and "negative" reviews, just based on the text content of the reviews. Use
IMDB dataset
Course Outcomes:
Apply the technique of Deep Neural network for implementing binary classification.
Software-Hardware requirements:
Python, Ubuntu Operating System 14.0 or above, 2 CPU CoreProcessor, 2GB RAM.
THEORY:
Binary Classification
Binary Classification refers to classifying samples in one of two categories. In this example, we will
design a neural network to perform two-class classification, or binary classification, of reviews, from
the IMDB movie reviews dataset, to determine whether the reviews are positive or negative. We will
use the Python library, Keras.
11
To prepare our data, we will One-hot Encode our lists and turn them into vectors of 0’s and
1’s. This would blow up all of our sequences into 10,000-dimensional vectors containing 1
at all indices corresponding to integers present in that sequence. This vector will have
element 0 at all index, which is not present in the integer sequence.
Simply put, the 10,000-dimensional vector corresponding to each review will have
Hidden layers, simply put, are layers of mathematical functions each designed to produce an
output specific to an intended result.
Hidden layers allow for the function of a neural network to be broken down into specific
transformations of the data. Each hidden layer function is specialized to produce a defined
output.For example, a hidden layer functions that are used to identify human eyes and ears
may be used in conjunction by subsequent layers to identify faces in images. While the
functions to identify eyes alone are not enough to independently recognize objects, they can
function jointly within a neural network.
In this network, we will leverage hidden layers. We will define our layers as such.
Dense(16, activation='relu')
The argument being passed to each Dense layer, (16) is the number of hidden units of a layer.
The output from a Dense layer with relu activation is generated after a chain of tensor
operations. This chain of operations is implemented as
12
Having 16 hidden units means that the matrix W will be of the shape (input_Dimension, 16
). In this case, where the dimension of the input vector is 10,000, the shape of the Weight
matrix will be (10000, 16). If you were to represent this network as a graph, you would see
16 neurons in this hidden layer.
Each of these balls or hidden units is a dimension in the representation space of the layer.
Representation space is the set of all viable representations for the data. Every hidden layer
composed of its hidden units aims to learn one specific transformation of the data or one
feature/pattern from the data.
Model Architecture
For our model, we will use
Two intermediate layers with 16 hidden units each
Third layer that will output the scalar sentiment prediction
Intermediate layers will use the relu activation function. relu or Rectified linear unit function will
zero out the negative values.
Sigmoid activation for the final layer or output layer. A sigmoid function “squashes” arbitrary values
into the [0,1] range.
13
The Sigmoid Activation Function
There are formal principles that guide our approach in selecting the architectural attributes of a
model.
model.complie(optimizer='rmsprop',
loss = 'binary_crossentropy',
metrics = ['accuracy'])
One could use a customized loss function or optimizer by passing a custom class instance as an
argument to the loss, optimizer or mertics fields.
In this example, we will implement our default choices, but we will do so by passing class instances.
This is precisely how we would do it if we had customized parameters.
Setting up Validation
We will set aside a part of our training data for validation of the accuracy of the model as it trains. A
validation set enables us to monitor the progress of our model on previously unseen data as it goes
through epochs during training.
14
Validation steps help us fine-tune the training parameters of the model.fit function to avoid
Overfitting and underfitting of data.
Calling the fit method returns a History object. This object has an attribute history which is a
dictionary containing four entries: one per monitored metric.
history_dict = history.history
history_dict.keys()
• Training loss
• Training Accuracy
• Validation Loss
• Validation Accuracy
• at the end of each epoch.
Let’s use Matplotlib to plot Training and validation losses and Training and Validation Accuracy
side by side.
We observe that minimum validation loss and maximum validation Accuracy is achieved at around
3–5 epochs. After that, we observe two trends:
To address Overfitting, we will reduce the number of epochs to somewhere between 3 and 5. These
15
results may vary depending on your machine and due to the very nature of the random assignment
of weights that may vary from model to model.
DATASET: https://www.kaggle.com/datasets/lakshmi25npathi/imdb-dataset-of-50k-
movie-reviews?resource=download
IMPLEMENTATION:
# Load the data, keeping only 10,000 of the most frequently occuring
words
(train_data, train_labels), (test_data, test_labels) = imdb.load_data
(num_words = 10000)
16
For kicks, let’s decode the first review.
# Let's quickly decode a review
decoded_review
17
X_train[0]
X_train.shape
Vectorize labels
y_train = np.asarray(train_labels).astype('float32')
y_test = np.asarray(test_labels).astype('float32')
model = models.Sequential()
model.add(layers.Dense(16, activation='relu', input_shape=(10000,)))
model.add(layers.Dense(16, activation='relu'))
model.add(layers.Dense(1, activation='sigmoid'))
model.compile(optimizer=optimizers.RMSprop(lr=0.001),
loss = losses.binary_crossentropy,
metrics = [metrics.binary_accuracy])
18
history_dict = history.history
history_dict.keys()
# Plotting losses
loss_values = history_dict['loss']
val_loss_values = history_dict['val_loss']
plt.show()
19
# Training and Validation Accuracy
acc_values = history_dict['binary_accuracy']
val_acc_values = history_dict['val_binary_accuracy']
plt.show()
20
model.fit(partial_X_train,
partial_y_train,
epochs=3,
batch_size=512,
validation_data=(X_val, y_val))
result
y_pred = np.zeros(len(result))
for i, score in enumerate(result):
y_pred[i] = 1 if score > 0.5 else 0
# Error
mae
CONCLUSION:
Thus, we have successfully classified reviews on IMDB.
21
EXPERIMENT 3
Course Objective:
To classify fashion clothing into categories using CNN model.
Problem Statement:
Use MNIST Fashion Dataset and create a classifier to classify fashion clothing into categories.
Course Outcomes:
Apply the technique of Convolution (CNN) for implementing Deep Learning models.
Software-Hardware requirements:
Python, Ubuntu Operating System 14.0 or above, 2 CPU CoreProcessor, 2GB RAM.
THEORY:
Convolutional neural network (CNN)
CNN is basically a model known to be Convolutional Neural Network and in recent times it has
gained a lot of popularity because of its usefulness. CNN uses multilayer perceptrons to do
computational works. CNN uses relatively little pre-processing compared to other image
classification algorithms. This means the network learns through filters that in traditional algorithms
were hand-engineered. So, for the image processing tasks CNNs are the best-suited option.
These CNN models are being used across different applications and domains, and they’re especially
prevalent in image and video processing projects.
A Convolutional Neural Network is a powerful neural network that uses filters to extract features
from images. It also does so in such a way that position information of pixels is retained.
A convolution is a mathematical operation applied on a matrix. This matrix is usually the image
represented in the form of pixels/numbers. The convolution operation extracts the features from the
image.
The building blocks of CNNs are filters a.k.a. kernels. Kernels are used to extract the relevant features
from the input using the convolution operation.
Let’s try to grasp the importance of filters using images as input data. Convolving an image with
filters results in a feature map:
They can identify faces, individuals, street signs, platypuses, and many other aspects of visual data.
CNNs overlap with text analysis via optical character recognition, but they are also useful when
22
analyzing words as discrete textual units. They’re also good at analyzing sound.
CNN captures the spatial features from an image. Spatial features refer to the arrangement of pixels
and the relationship between them in an image. They help us in identifying the object accurately, the
location of an object, as well as its relation with other objects in an image
In the above image, we can easily identify that its a human’s face by looking at specific features like
eyes, nose, mouth and so on. We can also see how these specific features are arranged in an image.
That’s exactly what CNNs are capable of capturing.
CNN also follows the concept of parameter sharing. A single filter is applied across different parts
of an input to produce a feature map:
The number of filters (kernel) you will use on the input will result in same amount of feature maps.
Think of filter like a membrane that allows only the desired qualities of the input to pass through it.
Figure: CNN
23
Producing a feature map by applying convolution and the activation function ReLU to the
Input Data
24
MNIST dataset:
mnist dataset is a dataset of handwritten images as shown below in the image.
We can get 99.06% accuracy by using CNN(Convolutional Neural Network) with a functional
model. The reason for using a functional model is to maintain easiness while connecting the layers.
25
DATASET: https://www.kaggle.com/datasets/oddrationale/mnist-in-csv
IMPLEMENTATION:
if k.image_data_format() == 'channels_first':
x_train = x_train.reshape(x_train.shape[0], 1, img_rows, img_cols)
x_test = x_test.reshape(x_test.shape[0], 1, img_rows, img_cols)
inpx = (1, img_rows, img_cols)
else:
x_train = x_train.reshape(x_train.shape[0], img_rows, img_cols, 1)
x_test = x_test.reshape(x_test.shape[0], img_rows, img_cols, 1)
inpx = (img_rows, img_cols, 1)
x_train = x_train.astype('float32')
x_test = x_test.astype('float32')
x_train /= 255
x_test /= 255
y_train = keras.utils.to_categorical(y_train)
26
y_test = keras.utils.to_categorical(y_test)
Now, the dataset is ready so let’s move towards the CNN model :
inpx = Input(shape=inpx)
layer1 = Conv2D(32, kernel_size=(3, 3), activation='relu')(inpx)
layer2 = Conv2D(64, (3, 3), activation='relu')(layer1)
layer3 = MaxPooling2D(pool_size=(3, 3))(layer2)
layer4 = Dropout(0.5)(layer3)
layer5 = Flatten()(layer4)
layer6 = Dense(250, activation='sigmoid')(layer5)
layer7 = Dense(10, activation='softmax')(layer6)
Firstly, we made an object of the model as shown in the above-given lines, where [inpx] is the input
in the model and layer7 is the output of the model. We compiled the model using the required
27
optimizer, loss function and printed the accuracy and at the last model.fit was called along with
parameters like x_train(means image vectors), y_train(means the label), number of epochs, and the
batch size. Using fit function x_train, y_train dataset is fed to model in particular batch size.
Output:
CONCLUSION:
Thus, we have learnt how to classify fashion clothing into categories using CNN.
28
Group 2
29
MINI PROJECT
Course Objective:
To illustrate the concepts of Artificial Intelligence/Machine Learning(AI/ML).
Problem Statement:
Face Detection and Recognition is one of the areas of computer vision where the research
actively happens. The applications of Face Recognition include Face Unlock, Security and
Defense, etc.
Course Outcomes:
Identify and apply the suitable algorithms to solve AI/ML problems.
Software-Hardware requirements:
Python, Ubuntu Operating System 14.0 or above, 2 CPU CoreProcessor, 2GB RAM.
THEORY:
Face Detection and Recognition is one of the areas of computer vision where the research actively
happens. The applications of Face Recognition include Face Unlock, Security and Defense, etc.
Doctors and healthcare officials use face recognition to access the medical records and history of
patients and better diagnose diseases.
In this python project, we are going to build a machine learning model that recognizes the persons
from an image. We use the face_recognition API and OpenCV in our project.
Face recognition is the process of identifying or verifying a person’s face from photos and video
frames.
Face detection is defined as the process of locating and extracting faces (location and size) in an
image for use by a face detection algorithm.
Face recognition method is used to locate features in the image that are uniquely specified. The
facial picture has already been removed, cropped, scaled, and converted to grayscale in most cases.
Face recognition involves 3 steps: face detection, feature extraction, face recognition.
OpenCV is an open-source library written in C++. It contains the implementation of various
algorithms and deep neural networks used for computer vision tasks.
30
DATASET:
We can do this face recognition project using our own dataset. For this project, let’s take the
cast of the popular American web series “Friends” as the dataset. The dataset is included
with face recognition project code
IMPLEMENTATION:
31
We loop through each of the face locations and its encoding found in the image. Then we compare
this encoding with the encodings of the faces from the “train” dataset.
Then calculate the facial distance meaning that we calculate the similarity between the encoding of
the test image and that of the train images. Now, we pick the minimum valued distance from it
indicating that this face of the test image is one of the persons from the training dataset.
Now, draw a rectangle with the face location coordinates using the methods from the cv2 module.
for (top, right, bottom, left), face_encoding in zip(face_locations, face_encodings):
matches = fr.compare_faces(known_name_encodings, face_encoding)
name = ""
face_distances = fr.face_distance(known_name_encodings, face_encoding)
best_match = np.argmin(face_distances)
if matches[best_match]:
name = known_names[best_match]
cv2.rectangle(image, (left, top), (right, bottom), (0, 0, 255), 2)
cv2.rectangle(image, (left, bottom - 15), (right, bottom), (0, 0, 255), cv2.FILLED)
font = cv2.FONT_HERSHEY_DUPLEX
cv2.putText(image, name, (left + 6, bottom - 6), font, 1.0, (255, 255, 255), 1)
Display the image using the imshow() method of the cv2 module.
cv2.imshow("Result", image)
Save the image to our current working directory using the imwrite() method.
cv2.imwrite("./output.jpg", image)
OUTPUT:
CONCLUSION:
32
In this machine learning project, we developed a face recognition model in python and opencv using
our own custom dataset..
33