*64*

# Multi-layer Perceptron in TensorFlow

Multi-Layer perceptron defines the most complex architecture of artificial neural networks. It is substantially formed from multiple layers of the perceptron. TensorFlow is a very popular deep learning framework released by, and this notebook will guide to build a neural network with this library. If we want to understand what is a Multi-layer perceptron, we have to develop a multi-layer perceptron from scratch using Numpy.

The pictorial representation of multi-layer perceptron learning is as shown below-

MLP networks are used for supervised learning format. A typical learning algorithm for MLP networks is also called **back propagation’s algorithm**.

A multilayer perceptron (MLP) is a feed forward artificial neural network that generates a set of outputs from a set of inputs. An MLP is characterized by several layers of input nodes connected as a directed graph between the input nodes connected as a directed graph between the input and output layers. MLP uses backpropagation for training the network. MLP is a deep learning method.

Now, we are focusing on the implementation with MLP for an image classification problem.

**The above line of codes generating the following output-**

## Creating an interactive section

We have two basic options when using TensorFlow to run our code:

- Build graphs and run sessions [Do all the set-up and then execute a session to implement a session to evaluate tensors and run operations].
- Create our coding and run on the fly.

For this first part, we will use the interactive session that is more suitable for an environment like Jupiter notebook.

## Creating placeholders

It’s a best practice to create placeholder before variable assignments when using TensorFlow. Here we’ll create placeholders to inputs (“Xs”) and outputs (“Ys”).

Placeholder “X”: Represent the ‘space’ allocated input or the images.

- Each input has 784 pixels distributed by a 28 width x 28 height matrix.
- The ‘shape’ argument defines the tensor size by its dimensions.