# A main component to design fish recognition and classification system architecture

**Published:** **Last Edited:**

This essay has been submitted by a student. This is not an example of the work written by our professional essay writers.

### 1.1 Introduction

In the previous chapter, the architecture and the approaches of object recognition and classification system were shown in details. Moreover, the features of shape characters of fish that will be used for classification stage are provided. Therefore, this chapter focuses on some background of literature approaches on related work and concepts in the field of object recognition and classification. In particular, a main component to design fish recognition and classification system architecture is used; it will show these experiments history of development in several cases.

The following literature review is divided into four main sections. The fish recognition and classification, first aspect is covered. The second aspect relates with image segmentation techniques to segment underwater images are presented. Investigates most of the feature extraction and selection by shape representation and description, the third aspect is applied. Finally, the classifiers technique for object recognition and classification in aspect of support vector machine is reported.

### 1.2 Fish Recognition and Classification

Recently, there were many researchers who attempted to design and apply the interaction between an underwater environment and learning techniques to develop the recognition and classification system in order to classify the fish. Therefore, Castignolles et al., (1994) used off-line method detection with static thresholds to segment the images that recorded by S-video tapes and enhance image contrast by using background lighting. Furthermore, to recognize the species a Bayes classifier was tested after extract twelve geometrical features from fish images. However, this method needs control on the light of background, determine the value of threshold and multiple imaging. Moreover, where fish are lined up close to each other, the applications tend to be impractical for the real-time.

The moment-invariant for features extraction is fast and very easy to implement. Therefore, Zion et al. (1999) stated the features extraction from dead fish tails by using moment-invariants in order to identify of species. Moreover, the image area is used to estimate fish mass. Furthermore, the accuracy of 99%, 93% and 93%, respectively, for grey mullet, St. Peter's fish and carp is got for identification of fish species. Therefore, later Zion et al., (2000) tested this method with live fish swimming in clean water. The accuracies were 100%, 91% and 91%, respectively for fish species identification. However, the features of the tail in the image which were extracted by the moment-invariant are strongly affects by the water opaqueness and fish motion. This method needs clear environments and all features appear clearly.

An automatic system to select the desirable features for recognition and classification object is needed. Therefore, Chan et al., (1999) developed a 3D point distribution model (PDM) in order to extract the lateral length measurement automatically from the images without an extensive user interaction to locate individual landmark points on the fish by using an n-tuple classifier as a tool to initiate the model. Moreover, the WISARD architecture is used as a Look-Up Table (LUT) which holds information about the pattern that the classifier tries to recognize, in order to assess the performance and usefulness of the n-tuple classifier in the application of fish recognition. However, this method needs to fix the pre-defined threshold value, amount of prior knowledge for the fish and the bigger training set.

Determine the landmarks as tips of snout or fins for fish are very important to recognize the fish. Therefore, Cardin & Friedland (1999) stated the morphometric analysis by biometric interpretation of fish homologous landmarks as tips of snout or fins for fish stock discrimination. However, they did not refer to algorithms for determining landmarks and the external points are not satisfactory because their locations are subjective.

From other aspect, Cardin (2000) reviewed the landmarks of shape by using morph metric methods for stock identification of fish. Moreover, Winans (1987) used the fins points, extremities point and arbitrarily landmarks in order to identify the fish from those points. Therefore, the attachment of fin membranes were found to be more effective for finfish group discrimination than landmarks located on extremities. Furthermore, Bookstein, (1990) stated the homologous landmarks were found to be more effective in describing shape than other arbitrarily located landmarks. However, these methods should be considered fish sample size, life history, stage of development and the features discriminating power.

Fourier descriptor for geometric features description is very famous algorithm. Therefore, Cadieux et al., (2000) stated the Fourier descriptors of silhouette contours, the geometric features described by seven of moment-invariants stated by Hu (1962) are developed in order to count fish by species from fish ways mounted next to river. Therefore, the 78% of accuracy is achieved by using a majority vote of three classification methods. However, this method needs sensors that generate silhouette contours as the fish swim between them and the hardware based on a commercial biomass counter.

The manual measurement for the landmarks points is more accurate to identify the object. Therefore, Tillett et al., (2000) proposed the modification of point distribution model (PDM) in order to segmented fish images by means is proposed. Moreover, the edge and its proximity in order to attract landmarks are considered. Furthermore, the average accuracy of 95% by estimating fish length to manual measurement is compared. However, this method required manual placement of the PDM in an initial position close to the centre of the fish, thereby affecting the accuracy of the final fitting. Also, neighboring fish images forced the PDM away from the correct edges and fish whose orientation was very different from the initial PDM or were smaller than the initial values could not be correctly fitted.

The combining of more than one classifier is important to get more accuracy to classify the objects. Therefore, Cedieux et al., (2000) proposed intelligent system by combining the result of three classifiers in order to recognize the fish. Therefore, Byes maximum quantification classifier, a learning vector classifier and One-class-One-Network of neural network classifier are used by analysis algorithm of an infrared silhouette sensor to acquire the fish and the majority vote. Moreover, the results depended on at least two from three classifiers should be show the same result. However, this method needs other approach for feature selection in order to improve the recognition performance and to optimize the selection of relevant characteristics for fish classification. Moreover, it needs more computational to identify and classify the object.

Detection, representation the features of object and then the classification are the main steps for any recognition and classification system. Therefore, Tidd & Wilder (2001) stated a machine vision system to detect and classify fish in an estuary by using a video sync signal to drive and direct the strobe lighting through a fiber bundle into a 30 cmÃ-30 cmÃ-30 cm field of view in a water tank. Moreover, to segment fish images and eliminate partial fish segments, the window-based segmentation algorithm and an aspect ratio are used by means of the segment aspect ratio and a length test. Furthermore, Bayes classifier is used to classify three fish species from extracted fish image area and aspect ratio. However, this method is tested on only 10 images of each of the species, and needs more computation. Moreover, they concluded that the system and method have the potential to operate in situ.

The monitoring objects in underwater is difficult problem. Therefore, Rife & Rock (2001) proposed Remotely Operated Vehicles (ROV) in order to follow the marine animal in underwater. However, this method needs continuous hours of the pieces movements.

Locating the critical points of object is very important to determine the length, weight and the area of the objects. Therefore, Martinez et al., (2003) stated an underwater stereo vision system is used to calculate the weight of fish from their length by using a prior knowledge of the species in order to find points of the fish image and linking them with real-world coordinates. Moreover, in order to find caudal fin points and the tip of the head, the template matching with several templates is used. Furthermore, accuracy of 95% and 96% for estimated fish weight is reported. However, this method needs a prior knowledge of the species, critical points to calculate the length and only used to find the weight.

The shape of object is very important feature to recognize and identify the objects. Therefore, Lee et al., (2003) developed automated Fish Recognition and Monitoring (FIRM) system, as shape analysis algorithm in order to locate critical landmark points by using a curvature function analysis. Moreover, the contour is extracted based on these landmark points. Furthermore, from this information species classification, species composition, densities, fish condition, size, and timing of migrations can be estimated. However, this method utilizes high-resolution images and determines the location for the critical points of fish shape.

In a conventional n-tuple classifier, the n-tuple is formed by selecting multiple sets of n distinct locations from the pattern space. Therefore, Tillett & Lines (2004) proposed an n-tuple binary pattern classifier with the difference between two successive frames in order to locate the initial fish image for detecting the fish head. Moreover, the dead fish hanging in a tank are used to estimate the mean mass. However, the estimation accuracy was low for live fish images due to poorer imaging conditions and larger fish population density.

The different features can used together to classify the object. Therefore, Chambah et al., (2004) proposed Automatic Color Equalization (ACE) in order to recognize the fish spaces. Furthermore, the segmentation by using background subtraction was presented. The geometric features, color features, texture features and motion features are used. Then, Bayes classifier is used to classify the selected fishes to one of the learned species. However, this method depends on the color features that need lightness constancy and color constancy to extract visual information from the environment efficaciously.

The semi-local invariants recognition is based on the idea that a direct search for visual correspondence is the key to successful recognition. Therefore, Lin et al., (2005) proposed neighbor pattern classifier by semi-local invariants recognition to recognize the fish. Moreover, when they compare it with integral invariants, they found it less mismatching. Furthermore, they compare wavelet-based invariants with summation invariants and found it has more strong immunity to noise. However, this method needs some critical point of the fish shape.

The Bayesian filter was originally intended for statistical recognition techniques, and is known to be a very effective approach. Therefore, Erikson et al. (2005) proposed fish tracking by using Bayesian filtering technique. Moreover, this method models fish as an ellipse having eight parameters. However, this method considers only counting the fish without looking into its type. Furthermore, the fish may be having varying in number of the parameters.

From other aspect, Lee et al., (2008) stated several shape descriptors, such as Fourier descriptors, polygon approximation and line segments in order to categorize the fish by using contour representation that extracted from their critical landmark points. However, the main difficulty of this method is that landmark points sometimes cannot be located very accurately. Moreover, it needs a high quality image for analysis.

### Table 1.1: Critical Analysis of Relevant Approaches

Author |
Algorithm |
Remarks |

Castignolles et al. 1994 |
Off-line method |
This method needs control on the light of background, determine the value of threshold. Moreover, where fish are lined up close to each other, the applications tend to be impractical for the real-time. |

Zion et al., 1999 |
Moment-invariants |
The features of the tail in the image which were extracted by the moment-invariant are strongly affects by the water opaqueness and fish motion. Therefore, this method needs clear environments and all features appear clearly. |

Chan et al. 1999 |
PDM |
this method needs to fix the pre-defined threshold value, amount of prior knowledge for the fish and the bigger training set. |

Cardin and Friedland 1999 |
Morphometric analysis |
They did not refer to algorithms for determining landmarks and the external points are not satisfactory because their locations are subjective. |

Cardin 2000 |
Develop Morphometric analysis |
These methods should be considered fish sample size, life history, stage of development and the features' discriminating power. |

Cadieux et al. 2000 |
Fourier descriptor |
This method needs sensors that generate silhouette contours as the fish swim between them and the hardware based on a commercial biomass counter. |

Tillett et al. 2000 |
Modify PDM |
This method required manual placement of the PDM in an initial position close to the centre of the fish, thereby affecting the accuracy of the final fitting. Also, neighboring fish images forced the PDM away from the correct edges and fish whose orientation was very different from the initial PDM or were smaller than the initial values could not be correctly fitted. |

Cedieux et al. 2000 |
Intelligent System |
This method needs other approach for feature selection in order to improve the recognition performance and to optimize the selection of relevant characteristics for fish classification. Moreover, it needs more computational to identify and classify the object. |

Tidd and Wilder 2001 |
Machine Vision System |
This method is tested on only 10 images of each of the species, and needs more computation. Moreover, they concluded that the system and method have the potential to operate in situ. |

Rife and Rock 2001 |
ROV |
This method needs continuous hours of the pieces movements. |

Martinez et al., 2003 |
Template Matching |
This method needs a prior knowledge of the species, critical points to calculate the length and only used to find the weight. |

Lee et al. 2003 |
FIRM |
This method utilizes high-resolution images and determines the location for the critical points of fish shape. |

Tillett and Lines 2004 |
n-tuple |
The estimation accuracy was low for live fish images due to poorer imaging conditions and larger fish population density. |

Chambah et. al. 2004 |
ACE |
This method depends on the color features that need lightness constancy and color constancy to extract visual information from the environment efficaciously. |

Lin et al., 2005 |
Neighbor Pattern Classifier |
This method needs some critical point of the fish shape. |

Erikson et al. 2005 |
Bayesian Filtering Technique |
This method considers only counting the fish without looking into its type. Furthermore, the fish may be having varying in number of the parameters. |

Lee et al. 2008 |
Several Shape Descriptors |
The main difficulty of this method is that landmark points sometimes cannot be located very accurately. Moreover, it needs a high quality image for analysis. |

### 1.3 Image Segmentation Techniques

Basically, there are different techniques that would help to solve the image segmentation problems. Therefore, Jeon et al., (2006) categorized these techniques into, thresholding approaches, contour approaches, region approaches, clustering approaches and other optimization approaches using a Bayesian framework, neural networks. Moreover, the clustering techniques can be categorized into two general groups: partitional and hierarchical clustering algorithms. Furthermore, partitional clustering algorithms such as K-means and EM clustering are widely used in many applications such as data mining, compression, image segmentation and machine learning (Coleman & Andrews 1979; Carpineto & Romano 1996; Jain et al., 1999; Zhang 2002a; Omran et al., 2006). Therefore, this research will focus on the literature review relates with image segmentation techniques to segment fish of underwater images by using k-means algorithm and background subtraction approaches.

### 1.3.1 K-Means Algorithm for Image Segmentation

In general, the standard K-means clustering algorithm is employed in order to cluster a given dataset into k groups. Therefore, the standard K-means algorithm consists of four steps: Initialization, Classification, Centroid computation and Convergence condition. Moreover, several methods attempt to improve the standard K-means algorithm related to several aspects associated to each of the algorithm steps. Furthermore, regarding the computational of the algorithm the steps that need more improvements are initialization and convergence condition (Amir 2007 & Joaquín et al., 2007). Therefore, the following sections will be focused on this step in order to represent and address the review for this step.

### 1.3.1.1 The Initialization Step of K-Means Algorithm

Basically, the earliest reference to initialize the K-means algorithm was by Forgy in 1965 that choose points randomly and used as the seeds. Therefore, MacQueen, introduced to determine a set of cluster seeds by using an online learning strategy (MacQueen 1967 & Stephen 2007). However, this method can be choosing the point near a cluster centre or outlying point. Moreover, repeating the runs is the increased time taken to obtain a solution.

The approach in order to divide the dataset to classes without prior knowledge of classes is required. Therefore, Tou & Gonzales (1974) suggested the Simple Cluster Seeking (SCS) method by Calculating the distance between the first instance in the database and the next point in the database, if it is greater than some threshold then select it as the second seed, otherwise move to the next instance in the database and repeat until K seeds are chosen. However, this method depends on the value of threshold, the order of pattern vectors to be processed and repeating the runs is the increased time taken to reach the seeds chosen.

For optimal partition of dataset which can achieve better variation equalization than standard. Therefore, Linde et al., (1980) proposed a Binary Splitting (BS) method, based on the first run for K = 1, Then split into two clusters until convergence is reached and the cycle of split and converge is repeated until a fixed number of clusters is reached, or until each cluster contains only one point. However, this method increased the computational complexity by split and the algorithm must be run again.

Good initial seeds for clustering algorithm are significant in order to rapidly converge to the global optimal structure. Therefore, Kaufman & Rousseeuw (1990) suggested selecting the first seed as the most centrally located instance, then the next seed selected based on the greatest reduction in the distortion and continue until K seeds are chosen. However, this method needs more computation in choosing each seed.

In order to select the optimal seed artificial intelligence (AI) is used. Therefore, Babu & Murty (1993) and Jain et al., (1996) proposed a method by using genetic algorithms based on the various seed selections as population, and then the fitness of each seed selection is assessed by running the K-means algorithm until convergence and then calculating the Distortion value, in order to select of near optimal seed. However, this method should be run K-means for each solution of each generation. Moreover, a genetic algorithms result depends on the choice of population size, and crossover and mutation probabilities.

Enhancement approach in order to improve the clustering quality and overcome computational complexity is required. Therefore, Huang & Harris (1993) stated the Direct Search Binary Splitting (DSBS) method, based on Principle Component Analysis (PCA), in order to enhance splitting step in Binary Splitting algorithm. However, this method also required more computational to reach k seeds chosen.

Calculating the distance between all points of dataset in order to select the seed is used. Therefore, Katsavounidis et al. (1994) proposed the algorithm as the KKZ algorithm based on preferably one on the edge of the data as the first seed. Then, chosen the second seed based on the point which is furthest from the first seed. Moreover, choosing the furthest point from its nearest seed is repeated until K seeds are chosen. However, this method obvious pitfall from any noise in the data as preferably seed.

In order to increase the speed of the algorithm based on divide the whole input domain into subspaces is required. Therefore, Daoud & Roberts (1996) proposed approach to divide the whole input domain into two disjoint volumes, and then this subspace is assumed that the points are randomly distributed and that the seeds will be placed on a regular grid. However, this methods refers at the end into randomly choose.

The mean of the any dataset is important value in order to estimate the seed depends on it. Therefore, Thiesson et al., (1997) suggested approach to calculate the mean of the entire dataset based on randomly running time of the algorithm to produce the K seeds. However, this method uses the random way to repeat the steps until reach the desirable clusters.

In order to find better clustering initialization of k-means algorithm, Forgy's method is used. Therefore, Bradley & Fayyad (1998) presented a technique that begins by randomly breaking the data into 10, or so, subsets. Then it performs a K-means clustering on each of the ten subsets, all starting at the same set of initial seeds, which are chosen using Forgy's method. However, this method needs to determine the size of the subset and used the same initial seed for each subset.

A way of reducing the time complexity of initialization for k-means algorithm calculation is to use structures like k-d trees. Therefore, Likas et al., (2003) stated a global K-means algorithm which aims to gradually increase the number of seeds until K is found, by using the kd-tree to create K buckets and use the centroids of each bucket as seeds. However, this method needs to test the results to reach the best number of clusters.

The performance of iterative clustering algorithms depends highly on initial cluster centers. Therefore, Mitra et al., (2002) and Khan& Ahmad (2004) proposed a Cluster Centre Initialization Method (CCIA) based on the Density-based Multi Scale Data Condensation (DBMSDC) by estimating the density of the dataset at a point, and then sorting the points according to their density and examining each of the attributes individually to extract a list of possible seed locations. The process is repeated until a desired number of points remain. However, this method depends on other approach to reach the desired seeds, which lead to more computation complexity.

On the other read, in order to reduce the time complexity of initialization for k-means algorithm calculation is to use structures like k-d trees. Therefore, Stephen & Conor (2007) presented a technique for initializing the K-means algorithm based on incorporate kd-trees in order to obtain density estimates of the dataset. And then by using the distance and the density information sequentially to select K seeds. However, this method occasionally failed to provide the lowest value of distortion.

### Table 1.2: Critical Analysis of Relevant Approaches

Author |
Algorithm |
Remarks |

Forgy 1965 and MacQueen 1967 |
Random initial K-means |
This method can be choosing the point near a cluster centre or outlying point. Moreover, repeating the runs is the increased time taken to obtain a solution. |

Tou and Gonzales 1974 |
SCS |
This method depends on the value of threshold, the order of pattern vectors to be processed and repeating the runs is the increased time taken to reach the seeds chosen. |

Linde et al., 1980 |
BS |
This method increased the computational complexity by split and the algorithm must be run again. |

Kaufman and Rousseeuw 1990 |
Selecting the first seed. |
This method needs more computation in choosing each seed. |

Babu and Murty 1993 |
GA |
This method should be run K-means for each solution of each generation. Moreover, a genetic algorithms result depends on the choice of population size, and crossover and mutation probabilities. |

Huang and Harris 1993 |
DSBS |
This method also required more computational to reach k seeds chosen. |

Katsavounidis et al. 1994 |
KKZ |
This method obvious pitfall from any noise in the data as preferably seed. |

Daoud and Roberts 1996 |
two disjoint volumes |
This methods refers at the end into randomly choose. |

Thiesson et al. 1997 |
the mean of dataset |
This method uses the random way to repeat the steps until reach the desirable clusters. |

Bradley and Fayyad 1998 |
randomly breaking technique |
This method needs to determine the size of subset and the same initial seed for each subset. |

Likas et al. 2003 |
Global K-means |
This method needs to test the results to reach the best number of clusters. |

Khan and Ahmad 2004 |
CCIA |
This method depends on other approach to reach the desired seeds, which lead to more computation complexity. |

Stephen and Conor 2007 |
kd-trees |
This method occasionally failed to provide the lowest value of distortion. |

### 1.3.2 Background Subtraction for Image Segmentation

The basic approach for automatic object detection and segmentation methods is the background subtraction. Moreover, it is a commonly used class of techniques for segmenting out objects of a scene for different applications. Therefore, Wren et al., (1997) proposed running Gaussian Average based on ideally fitting a Gaussian probability density function on the last n pixel's values in order to model the background independently at each pixel location. Moreover, to increase the speed the standard deviation is computed. Therefore, the advantage of the running average is given by the low memory requirement instead of the buffer with the last n pixel values are used. However, the empirical weight as a tradeoff between stability and quick update is often chosen.

The detection of objects is usually approached by background subtraction based on multi-valued background. Therefore, Stauffer & Grimson (1999) proposed the multi-valued background model in order to describe the foreground and the background values. Moreover, the probability of observing a certain pixel value at specific time by means of a mixture of Gaussians is described. However, this method needs assigning the new observed pixel value to the best matching distribution and estimating the updated model parameters.

Density estimators can be a valuable component in an application like in the use of object tracking. Therefore, Elgammal et al. (2000) proposed a non-parametric model based on Kernel Density Estimation (KDE) by using the last n background values, in order to model the background distribution. Moreover, the sum of Gaussian kernels centered as one sample data by the most recent n background values as background is given. However, complete model estimation also requires the estimation of summation of Gaussian kernels.

The eigen-decomposition methods are computationally demanding by involving the computation of each eigenvector and corresponding eigenvalues. Therefore, Oliver et al., (2000) proposed eigen backgrounds approach based on eigenvalues decomposition by using the whole image instead of blocks of image. Moreover, this method can be improving its efficiency, but depend on the images used for the training set. However, this method not explicitly specified what images should be part of the initial sample, and whether and how such a model should be updated over time.

In order to generate and select of a plurality of temporal pixel samples derived from incoming image, the temporal median filter is used. Therefore, Lo & Velastin (2001) proposed temporal median filter based on the median value of the last n frames as the background model. Moreover, Cucchiara et al. (2003) developed the temporal median filter by computing the last n frames, sub-sampled frames and the time of the last computed median value in order to compute the median on a special set of values. However, the disadvantage of the temporal median filter approach, the computation by a buffer with the recent pixel values is required. Moreover, the median filter does not provide a deviation measure for adapting the subtraction threshold.

The information of the difference frames is accumulated, in order to construct a reliable background image. Therefore, Seki et al., (2003) proposed the background subtraction based on co-occurrence of image variations. Moreover, it works on blocks of N x N pixels treated as an N2 component vector, instead of working at pixel resolution. However, this method offers good accuracy against reasonable time and memory complexity. Furthermore, a certain update rate would be needed to cope with more extended illumination changes.

Background modeling of a moving object requires sequential density estimation. Therefore, Piccardi & Jan (2004) proposed Sequential Kernel Density Approximation (SKDA) by using some computational optimizations in order to mitigate the computational drawback of Mean-shift vector techniques. However, SKDA is an approximation of KDE which proves almost as accurate, but mitigates the memory requirement.

### Table 1.3: Critical Analysis of Relevant Approaches

Author |
Algorithm |
Remarks |

Wren et al. 1997 |
RGA |
The empirical weight as a tradeoff between stability and quick update is often chosen. |

Stauffer and Grimson 1999 |
MVBM |
This method needs assigning the new observed pixel value to the best matching distribution and estimating the updated model parameters. |

Elgammal et al. 2000 |
KDE |
Complete model estimation also requires the estimation of summation of Gaussian kernels. |

Oliver et al. 2000 |
Eigen Backgrounds |
This method not explicitly specified what images should be part of the initial sample, and whether and how such a model should be updated over time. |

Lo and Velastin 2001 |
T M F |
The disadvantage of the temporal median filter approach, the computation by a buffer with the recent pixel values is required. |

Cucchiara et al. 2003 |
developed TMF |
The median filter does not provide a deviation measure for adapting the subtraction threshold. |

Seki et al., 2003 |
CoIV |
This method offers good accuracy against reasonable time and memory complexity. Furthermore, a certain update rate would be needed to cope with more extended illumination changes. |

Piccardi and Jan 2004 |
SKDA |
SKDA is an approximation of KDE which proves almost as accurate, but mitigates the memory requirement. |

### 1.4 Features Extraction and Representation Techniques

Generally, feature extraction is an important step to pattern recognition and machine learning problems. Moreover, to recognize and classify an object in an image, some features out of the image must be first extracted. Therefore, image content may include both visual features and semantic content features. Furthermore, general visual features include color, texture and shape. Among these features, shape is the most important visual feature because it represents significant regions or relevant objects in images and it is one of the basic features used to describe image content.

### 1.4.1 Chain Code for shape representation

A binary shape may naturally be represented by its contour as a connected chain of pixel. Therefore, Freeman (1961) introduced the first chain code scheme in 1961 that is known as Freeman Chain Code (FCC), in order to efficient representation of a contour. Moreover, the codes involve 4-connected and 8-connected in order to follow the boundary in counter clockwise manner and keeps track of the direction from one contour pixel to the next. However, this method is very sensitive to noise as the errors cumulative, lack of the spatial information of the image content, and it is not invariant in starting point and scaling of the contour.

The less computation to represent the shape from their contour is needed. Therefore, Papert (1973) presented one of the simplest chain codes by using just two codes a right turn and left turn when following the contour shape. However, this method is not sufficient to represent the shape from different angles.

The efficient representation of shapes from their contour is needed. Therefore, Freeman (1974) proposed Differential Chain Code (DCC), in order to improve the efficiency and to solve the problems of sensitive to noise and not rotationally invariant by chain code. Moreover, the object can be rotate in 90-degree increments and still compare by subtracting each element of the chain code from the previous one and taking the result modulo n, where n is the connectivity. Moreover, Bons & Kegel (1977) introduced the differential chain coding, by calculating the difference between the two links and assigning a variable length code to the result based on the correlation between two successive links. However, these methods don't get around the inherent sensitivity of chain codes to rotation on the discrete pixel grid.

For better compression and smooth contours that help to improve the efficiency of DCC is required. Therefore, Kaneko & Okudaira (1985) proposed an algorithm for coding smooth contours based on divide the shape into a sequence of straight line segments. Furthermore, Eden & Kocher (1985) introduced three basis symbols to represent different combinations of pairs of adjacent links, in order to improve the efficiency. Moreover, Lu & Dunham (1991) used Markov models to describe the relationship between successive chain links. However, this method is a data hog and depends on the Markov property.

The statistical method to overcome the limitations of the translation and scale invariant for shape descriptor is required. Therefore, Iivarinen (1996) proposed Chain Code Histogram (CCH), in order to describe the shape with invariant translation and scale by giving just the approximation of the object's shape and the similar object can be grouped together. However, these methods only shows the probabilities for the different directions present in the contour. In other words, this method didn't take the direction distribution in a chain code into account. Therefore, this approximation causes some mistakes in image retrieval process.

Extracting interesting shape properties are required. Therefore, Bribiesca (1999) introduced a new chain code, called Vertex Chain Code (VCC), in order to represent the shape with invariant under translation, rotation and starting point based on represent shapes composed of triangular, rectangular, and hexagonal cells. However, this method has the main limitation that only allow rotation invariance over multiples of 45 or 90 degrees, in dependence of the metric used between neighbors.

The natural objects that exist in 3D structure and the 3D objects are digitalized and represented by constant orthogonal straight line segments. Therefore, Bribiesca (2000) proposed chain code for representing 3D curves by only considers relative direction changes that invariant under translation and rotation. However, in order to give information to each sample point of 3D surface, it is necessary to establish the mapping relationship between the planar image and the 3D surface. Therefore, it is usually computationally expensive.

The object can be recognized from different angles and viewpoints. Therefore, Wang & Xie (2004) introduced Minimize Sum Statistical Direction Code (MSSDC), in order to solve the problem of being invariant to translation, rotation, and proportional to scaling. However, this method cannot preserve information on the exact shape of a contour, because it only shows the probabilities for the different directions present in the contour.

The shape representation by discrete curve is needed. Therefore, Cruz & Dagnino 2005) proposed 3OT approach by using only three of the five original symbols. It is similar to 3D chain code approach proposed by Bribiesca (2000). However, this method used only to represent 2D binary shapes without holes.

Adaptation the concept of chain codes and Vertex Chain Code in order to enhance the efficiency is reported. Therefore, Liu & Zalik (2005) adopted chain code with Huffman coding based on the most frequently used eight-directional Freeman chain code. However, this method used a chain moves in a direction defined by a code element depends on the probability.

Shape representation by vertex chain code in order to improve the efficiency is required. Therefore, Liu et al., (2007) developed, Extended Vertex Chain Code (E_VCC) and Variable-length Vertex Chain Code (V_VCC), based on rectangular cells of vertex chain code. However, the length of the E_VCC code is not longer than the original VCC and V_VCC depends on the probability.

In other aspect Liu et al., (2007) developed Compressed Vertex Chain Code (C_VCC) consists of five codes and using Huffman coding concept. However, this method is determined the Huffman codes by extensive statistical investigation. It is worth noticing that the statistical data depends slightly on used resolution. Moreover, this dependence is not so drastic that they could cause any change in the derived Huffman codes.

### Table 1.4: Critical Analysis of Relevant Approaches

Author |
Algorithm |
Remarks |

Freeman 1961 |
FCC |
This method is very sensitive to noise as the errors cumulative, lack of the spatial information of the image content, and it is not invariant in starting point and scaling of the contour. |

Papert 1973 |
SCC |
This method is not sufficient to represent the shape from different angles. |

Freeman 1974 and Bons and Kegel 1977 |
DCC |
These methods don't get around the inherent sensitivity of chain codes to rotation on the discrete pixel grid. |

Kaneko and Okudaira, 1985, 1991 |
DCC with Markov Model. |
This method is a data hog and depends on the Markov property. |

Iivarinen 1996 |
CCH |
This method only shows the probabilities for the different directions present in the contour. In other words, this method didn't take the direction distribution in a chain code into account. Therefore, this approximation causes some mistakes in image retrieval process. |

Bribiesca 1999 |
VCC |
This method has the main limitation that only allow rotation invariance over multiples of 45 or 90 degrees, in dependence of the metric used between neighbors. |

Bribiesca 2000 |
CC for 3D |
In order to give information to each sample point of 3D surface, it is necessary to establish the mapping relationship between the planar image and the 3D surface. Therefore, it is usually computationally expensive |

Wang and Xie, 2004 |
MSSDC |
this method cannot preserve information on the exact shape of a contour, because it only shows the probabilities for the different directions present in the contour. |

Cruz and Dagnino 2005 |
3OT |
This method used only to represent 2D binary shapes without holes. |

Liu and Zalik 2005 |
CC with Huffman coding |
this method used a chain moves in a direction defined by a code element depends on the probability. |

Liu et al. 2007 |
E_VCC and V_VCC |
The length of the E_VCC code is not longer than the original VCC and V_VCC depends on the probability. |

Liu et al. 2007 |
C_VCC |
this method is determined the Huffman codes by extensive statistical investigation. It is worth noticing that the statistical data depends slightly on used resolution. Moreover, this dependence is not so drastic that they could cause any change in the derived Huffman codes. |

### 1.5 Summary

This chapter is a review of the literature dealing with some works in recognition and classification systems are applied to solve fish recognition problems. The literature contains works dealing with three approaches: image segmentation approaches, feature extraction approach and classification approach.

Therefore, image segmentation approaches have been demonstrated to be an effective way of developing underwater image segmentation. There have been many attempts to apply different image segmentation algorithm such as K-means algorithm and background subtraction.

On the other hand, a number of researchers who use feature extraction in order to represent and describe shape are mentioned. These works demonstrate that chain code is a good approach for it is capable of improving the efficient representation of shapes from their boundary.

However, several SVM- based on object classification is reviewed. Considering these works, SVM is good for object classification, due to its efficient interaction with the environment that gives learning ability to classify the fish from their features such as shape feature.

### Reference:

[1] Castignolles, N., Cattoen, M., Larinier, M., 1994. Identification and counting of live fish by image analysis. In: Rajala, S., Stevenson, R.L. (Eds.), Image and Video Processing II, Proc. SPIE 2182, pp. 200-209.

[2] Zion, B., Shklyar, A., Karplus, I., 1999. Sorting fish by computer vision. Comput. Elect. Agric. 23, 175-187.

[3] Zion, B., Shklyar, A., Karplus, I., 2000. In-vivo fish sorting by computer vision. Aquacult. Eng. 22 (3), 165-179.

[4] D.Chan, S. Hockaday, R. D. Tillett, L. G. Ross, "Factors Affecting the Training of a WISARD Classifier for Monitoring Fish Underwater" British Machine Vision Conference 1999.

[5] Cardin, S., Friedland, K., 1999. The utility of image processing techniques for morphometric analysis and stock identification. Fish. Res. 43, 129-139.

[6] Cardin, S., 2000. Advances in morphometric identification of fishery stocks. Rev. Fish Biol. Fish. 10, 91-112.

[7] Bookstein, F.L., 1990. Introduction to methods for landmark data. In: Rohlf, F.J., Bookstein, F.L. (Eds.), Proceedings of the Michigan Morphometrics Workshop, vol. 2. University of Michigan, Mus. Zool. Spec, pp. 215-226.

[8] Winans, G.A., 1987. Using morphometric and meristic characters for identifying stock of fish. In: Kumpf, H.E., Vaught, R.N., Grimes, C.B., Jhonason, A.G., Nakamura, E.L. (Eds.), Proceeding of Stock IdentificationWorkshop, vol. 199. NOAA Tech. Mem. NMFS-SEFC, pp. 135-146.

[9] Cadieux, S., Lalonde, F., Michaud, F., 2000. Intelligent system for automated fish sorting and counting. In: Proceeding IEEE/RSJ, International Conference on Intelligent Robots and Systems (IROS), Takamatsu, Japan.

[10] Hu, M., 1962. Visual pattern recognition by moment invariants. IRE Trans. Inf. Theory IT-8, 179-187.

[11] Tillett, R., McFarlane, N., Lines, J., 2000. Estimating dimensions of free-swimming fish using 3D point distribution models. Comput. Vis. Image Underst. 79, 123-141.

[12] Cadieux, S.; Michaud, F.; Lalonde, F. "Intelligent system for automated fish sorting and counting". Intelligent Robots and Systems, 2000. (IROS 2000) Proceedings IEEE/RSJ International Conference on 31 Oct.-5 Nov. 2000 Page(s):1279 - 1284 vol.2.

[13] Tidd, R.A., Wilder, J., 2001. Fish detection and classification system. J. Elect. Imag. 10 (1), 283-288.

[14] Rife, J. and S.M. Rock (2001).A pilot-aid for ROV-based tracking of gelatinous animals in the midwater. In: MTS/IEEE Oceans 2001: An Ocean Odyssey.Honolulu, Hawaii.MTS/IEEE, p.1,137-1,144.

[15] Martinez de Dios, J.R., Serna, C., Ellero, A., 2003. Computer vision and robotics techniques in fish farms. Robotica 21, 233-243.

[16] D. J. Lee, S. Redd, R. Schoenberger, X. Xu, and P. Zhan. An automated fish species classification and migration monitoring system. In Proc. 29th Annual Conf. IEEE Ind. Elect. Soc., pages 1080-1085, November 2003.

[17] Tillett, R., Lines, J., 2004. Monitoring salmon biomass with stereo cameras underwater. In: Proceedings of the EurAgEng2004 Conference, Leuven, Belgium.

[18] Chambah M, Semani D, Renouf A, Courtellemont P, Rizzi A (2004) Underwater color constancy: enhancement of automatic live fish recognition. Proceedings of the SPIE 5293:157-168

[19] Erikson F.Morais, Mario F.M.Campos, Flavio L.C. Padua and Rodrigo L. Carceroni, "Particle Filter-Based Predictive Tracking for Robust Fish Counting" Computer Graphics and Image Processing, 2005. SIBGRAPI 2005. 18th Brazilian Symposium on 09-12 Oct. 2005 Page(s):367 - 374

[20] W.Y.Lin, N.Boston, and Y.H.Hu, "Summation invariant and its applications to shape recognition." Appeared in Proceedings of ICASSP 2005

[21] D.-J. Lee et al.: Contour Matching for Fish Species Recognition and Migration Monitoring, Studies in Computational Intelligence (SCI) 122, 183-207 (2008)[22]

[23] B. Jeon, Y. Yung and K. Hong "Image segmentation by unsupervised sparse clustering, " pattern recognition letters 27science direct,(2006) 1650-1664

[24] M. G. H. Omran, A. Salman and A. P. Engelbrecht "Dynamic clustering using particle swarm optimization with application in image segmentation", Pattern Anal Applic (2006) 8: 332-344

[25] Zhang, Y. J. (2002a). "Image engineering and related publications" International Journal of Image and Graphics,(2002a) 2(3), 441-452.

[26] G. B. Coleman, H. C. Andrews (1979) "Image segmentation by clustering", Proc IEEE 67:773-785.

[27] A. K. Jain, M. N. Murty, P. J. Flynn, 1999. "Data clustering: A review", ACM Comput. Surveys 31 (3), 264-323.

[28] C. Carpineto, G. Romano (1996) "A lattice conceptual clustering system and its application to browsing retrieval", Mach Learn 24(2):95-122[29]

[30] Joaquín Pérez O, Rodolfo Pazos R, Laura Cruz R, Gerardo Reyes S, Rosy Basave T, and Héctor Fraire H, "Improving the Efficiency and Efficacy of the K-means Clustering Algorithm Through a New Convergence Condition", ICCSA 2007, LNCS 4707, Part III, pp. 674-682, Springer-Verlag Berlin Heidelberg 2007

[31] Amir Ahmad , Lipika Dey, A k-mean clustering algorithm for mixed numeric and categorical data. Science direct. Data & Knowledge Engineering 63 (2007) 503-527

[32] MacQueen, J.B., 1967. Some methods for classification and analysis of multivariate observation. In: Le Cam, L.M., Neyman, J. (Eds.), University of California

[33] Tou, J., Gonzales, R., 1974. Pattern Recognition Principles. Addison-Wesley, Reading, MA.

[34] Linde, Y., Buzo, A., Gray, R.M., 1980. An algorithm for vector quantizer design. IEEE Trans. Commun. 28, 84-95.

[35] Kaufman, L., Rousseeuw, P.J., 1990. Finding Groups in Data. An Introduction to Cluster Analysis. Wiley, Canada.

[36] Babu, G.P., Murty, M.N., 1993. A near-optimal initial seed value selection in k-means algorithm using a genetic algorithm. Pattern Recognition Lett. 14 (10), 763-769.

[37] JAIN, A. K. AND FLYNN, P. J. 1996. Image segmentation using clustering. In Advances in Image Understanding: A Festschrift for Azriel Rosenfeld, N. Ahuja and K. Bowyer, Eds, IEEE Press, Piscataway, NJ.

[38] Huang, C., Harris, R., 1993. A comparison of several codebook generation approaches. IEEE Trans. Image Process. 2 (1), 108-112.

[39] Katsavounidis, I., Kuo, C.C.J., Zhen, Z., 1994. A new initialization technique for generalized lloyd iteration. Signal Process. Lett. IEEE 1 (10), 144-146.

[40] Daoud, M.B.A., Roberts, S.A., 1996. New methods for the initialization of clusters. Pattern Recognition Lett. 17 (5), 451-45.

[41] Thiesson, B., Meck, B., Chickering, C., Heckerman, D., 1997. Learning mixtures of bayesian networks. Microsoft Technical Report TR-97-30, Redmond, WA.

[42] Bradley, P.S., Fayyad, U.M., 1998. Refining initial points for k-means clustering. In: Proc. 15th Internat. Conf. on Machine Learning. Morgan Kaufmann, San Francisco, CA, pp. 91-99. Available from: http://citeseer.ist.psu.edu/bradley98refining.html

[43] Likas, A., Vlassis, N., Verbeek, J.J., 2003. The global k-means clustering algorithm. Pattern Recognition 36, 451-461.

[44] Khan, S.S., Ahmad, A., 2004. Cluster center initialization algorithm for k means clustering. Pattern Recognition Lett. 25 (11), 1293-1302.

[45] Stephen J. Redmond *, Conor Heneghan "A method for initialising the K-means clustering algorithm using kd-trees" 2007 Pattern Recognition Letters 28 (2007) 965-973

[46] C. Wren, A. Azarhayejani, T. Darrell, and A.P. Pentland, "Pfinder: real-time tracking of the human body," IEEE Trans. on Patfern Anal. and Machine Infell., vol. 19, no. 7, pp. 78g785, 1997.

[47] C. Stauffer and W.E.L. Grimson, "Adaptive background mixture models for real-time tracking," Proc. IEEE CVPR 1999, pp. 24&252, June 1999.

[48] A. Elgammal, D. Hanvood, and L.S. Davis, "Nonparametric model for background subtraction," Proc. ECCV 2000, pp. 751-767, June 2000.

[49] N.M. Oliver, B. Rosario, and A.P. Pentland, "A Bayesian computer vision system for modeling human interactions," IEEE Trans. on Paftern Anal. and Machine Zntell., vol. 22, no. 8, pp. 831-843,2000.

[50] B.P.L. Lo and S.A. Velastin, "Automatic congestion detection system for underground platforms," Proc. ISIMP2001, pp. 158-161, May2001.

[51] R. Cucchiara, C. Grana, M. Piccardi, and A. Prati, "Detecting moving objects, ghosts, and shadows in video streams," ZEEE Tram on Paftem Anal. and Machine Infell,, vol. 25, no. 10, pp. 1337-1442,2003.

[52] M. Seki, T. Wada, H. Fujiwara, and K. Sumi, "Background subtraction based on cooccurrence of image variations," Proc. CVPR 2003, Vol. 2, pp. 65-72,2003.

[53] M. Piccardi, T. Jan, "Efficient mean-shift background subtraction", to appear in Proc. of IEEE 2004 KIP, Singapore, Oct. 2004.[54]

[55] Freeman H, Computer Processing of Line- Drawing Images, ACM Computing Surveys 6, 1974, 57-97

[56] Papert S, Uses of Technology to Enhance Education, Technical Report 298, Al Lab, MIT, 1973

[57] H. Freeman, Computer processing of line drawing images, Comput. Surv. 6 (March 1974) 57-97.

[58] J. H. Bons and A. Kegel, "On the Digital Processing and Transmission of Handwriting and Sketching," Proceedings of EUROCON 77, 1977, pp. 880-890

[59] T. Kaneko, M. Okudaira, Encoding of arbitrary curves based on chain code representation, IEEE Trans. Commun. 33 (June 1985) 697-707.

[60] C. Lu, J. Dunham, Highly efficient coding schemes for contour line based on chain code representations, IEEE Trans. Commun. 39 (October 1991) 1511-1514.

[61] M. Eden, M. Kocher, On the performance of a contour coding algorithm in the context of image coding. Part 1: Contour segment coding, Signal Process. 8 (1985) 381-386.

[62] Iivarinen J, and Visa A, Shape Recognition of irregular Objects, Intelligent Robots and Computer Vision XV: Algorithms, Techniques, Active Vision, and Material Handlings, SPIE, 1996, 25-32

[63] Bribiesca E, A New Chain Code. Pattern Recognition, Vol. 32, Issue 2, 1999, 235-251

[64] Bribiesca E, A Chain Code for Representing 3D Curves, Pattern Recognition, Vol 33, 2000, 755-765

[65] X.L. Wang and K.L. Xie, "A novel direction chain code-based image retrieval", In Proceedings of the Fourth International Conference on Computer and Information Technology (CIT'04), 2004, pp. 190-193.

[66] Cruz S H and Dagnino R M R, Compressing bi-Level Images by Means of a 3-bit Chain Code, SPIE Opt. Eng. 44, 2005, 1-8

[67] Liu Yong Kui, Zalik Borut, An Efficient Chain Code with Huffman Coding, Pattern Reognition, Vol 38, 2005, 553-557

[68] Liu Yong Kiu, Wei Wei, Wang Peng Jie, Zalik Borut, Compressed Vertex Chain Codes, Pattern Recognition, Vol 40, 2007, 2908-2913

[69] Liu Yong Kiu, Wei Wei, Wang Peng Jie, Zalik Borut, Compressed Vertex Chain Codes, Pattern Recognition, Vol 40, 2007, 2908-2913