A Review on Color Recognition using Deep Learning and Different Image Segmentation Methods

Page 1

International Research Journal of Engineering and Technology (IRJET)

e-ISSN: 2395-0056

Volume: 09 Issue: 02 | Feb 2022

p-ISSN: 2395-0072

www.irjet.net

A Review on Color Recognition using Deep Learning and Different Image Segmentation Methods Chirag Mahesh Sahasrabudhe B. Tech Student, Dept. of Computer Science and Technology, MIT World Peace University, Pune, India. ---------------------------------------------------------------------***---------------------------------------------------------------------Abstract – With the publication of backpropagation algorithm paper by Geoffrey Hinton, deep learning has got the boost. In this paper, we talk about a deep learning model can be used to recognize various colors and impact of different segmentation methods on the color recognition.

2.3 Watershed Image segmentation method. As the name suggests, watershed segmentation is somewhat similar to geographical water shedding [5]. In this technique, the image is seen as a topographic landscape with ridges and valleys. The elevation points or values are the brightness of each pixel.

Key Words: Color recognition, CNN, Deep learning, Otsu’s method, ReLU activation function, Watershed segmentation transfer learning.

2.4 Otsu’s method for Image segmentation. In Otsu’s threshold method [6], we iterate through all the possible values of the pixels and calculate a measure of spread for every pixel. The pixel which is in the foreground can be distinguished from the pixel in background by assigning a class level. Black label can be used for background pixels and whereas white for foreground features. Generally, grayscale histogram is passed to algorithms.

2. TERMINOLOGIES USED. 2.1 Deep Learning. Deep learning is a subset of machine learning which is further a subset of artificial intelligence. When fed with huge amount of raw data, deep learning can discover patterns in the given data. Further, the multi-layers of deep learning also known as neural networks, can recognize similar patterns and hence, segregate them into different classes. One advantage of deep learning over traditional machine learning algorithms is that, we can’t give raw data (such as .csv) directly to the machine learning algorithm. Before giving input, we have to do preprocessing of the data. But we can give raw input to deep learning directly. Some of the examples of deep learning algorithms are convolutional neural network, recurrent neural network, generative adversarial networks etc. [1][2].

2.5 Adaptive Boosting. Adaboost (short for Adaptive boosting) algorithm was first discussed by Schapire and Freund, in 1997[7]. It works on the concept of Majority voting. It is an ensemble type of learning. The [8] common way to use adaptive boosting technique is with a decision tree. An adaboost with a decision tree is also known as a conventional adaboost. A tree with just one node and two leaves is known as a stump. Stump only works on one variable hence; they are also known as weak learners. The errors made by the first stump influence the output of the second stump. Stumps vary in their sizes. So, in final classification voting, some stumps get more influence (say) than the others. In adaboost every sample of the dataset is assigned with a sample weight. This sample weight indicates the importance of that sample to be correctly classified.

2.2 Convolutional neural networks. Similar to traditional neural networks, [3] CNN has three different types of layers namely, input, hidden and output layer. The difference here is that, the input given to CNN is an image or pixel matrix. The most important part of CNN is the kernel which is a 2D matrix of N X N size. In this matrix, each point has its own weight. The kernel size generally taken is of 2 X 2 matrix size.

Sample weight is calculated as follows:

Another characteristic which makes CNN technique to achieve higher accuracy results is large local receptive fields. [4] The receptive field size increases as the network becomes deeper and complex or a pooling layer is added to the network. A CNN works on a large receptive field (for example 48 X 48) as on other hand traditional one’s work on small receptive field such as 16 X 16.

© 2022, IRJET

|

Impact Factor value: 7.529

The stump having lowest Gini index or Gini impurity is taken as the first stump for classification purposes. 2.6 Rectified Linear Unit (ReLU) Activation function. This activation function preserves the properties of linear models because it is a linear function thus, making it easy to optimize [9]. ReLU function, performs on the threshold value. If the input element is less than the threshold value,

|

ISO 9001:2008 Certified Journal

|

Page 578


Turn static files into dynamic content formats.

Create a flipbook
Issuu converts static files into: digital portfolios, online yearbooks, online catalogs, digital photo albums and more. Sign up and create your flipbook.