Text Regions Based On Measurement Computer Science Essay

Published:

This essay has been submitted by a student. This is not an example of the work written by our professional essay writers.

The text regions based on measurement have discrete characteristics from the non text regions based methods on slope strength and consistency properties. Basically a region-based method consists of two stages: 1) To evaluate the text detection in local image regions using differentiation, and 2) Local text regions into text blocks and text verification to discard non-text regions for future processing using text localization method. An earlier method proposed by Jin Hyung Kim et al. [2] Kim support vector machines (SVMs) are used as the texture classifier well even in 2D or 3D dimensional spaces in the absence of a rich set of training and also further extractor is required to decrease the dimensionality of the quality component. SVMs extract their own architecture using kernel function in this paper Lyu et al. [3] text edges to detect candidate of various scales with a sobel operator and also using threshold procedure to luminance changes is used to filter out non-text edges and the clustered text regions into text lines by definite profile projection analysis(PPA). In the paper Chen and Yuille [5] To increase up text detection proposed methods using a cascade Boost classifier, the selected from a feature pool containing gray-level, gradient and edge features in the weak learners. In this method the detected text regions are combined into text blocks are alignment by the local binarization. This method performs capitalistic is more than 10 times faster than the other methods. The observations based on the connected component methods have distinct mathematical features, spatial neighbouring components and statistical relationships. These methods usually consist of three stages: 1) Connected component extraction to segment candidate text from images; 2) CCA to filter out non-text components using heuristic rules or classifiers; and 3) post-processing to cluster the text components into text blocks. [6] Stroke width transformation method (SWT) to measure the appropriate height and width of the text in the images. It's used to reduce the unwanted interference in the non-stroke regions in the images. In the paper Ali Mosleh. [7] Histogram of gradient basis calculation is used to calculate horizontal and vertical projections in the image. It reduces the background by calculating the threshold intensity of the text alone in the text blocks. In the paper Marco Pedersoli. [9] In the paper Mohammadreza Yadollahi. Experimental results evaluated on the Neural network classifier which performance training and testing methods. Training competition dataset show that our approach yields higher precision and performance compared with state-of-the-art methods.

ABSTRACT

Text detection and localization in Multi-color images is important for content-based image analysis. This conflict is occurring due to the various backgrounds, the inhomogeneous illumination, and the variations of text font, size and line orientation. Detecting text in multi-colour images is an important prerequisite. The RBG image is converted into YUV image, after that the multidimensional filter is used to reduce the noise in the YUV image. Canny edge detection is used to measure the continuity of the edges in the images. A efficient text detection is proposed using stroke width transformation method based on contours which can effectively remove the interference of non-stroke edges in complex background and the importance of recent feature (inter-frame feature), in the part of caption extraction(detection, localization).The horizontal and vertical histogram basis is used to calculate the luminance and chrominance which defines the background. Moreover the morphological operation which removes non text areas in the boundaries. Since some background pixels can also have the similar colour, some false stroke areas or character pixels are possible to appear in the output image, which will degrade the recognition rate of OCR (optical character recognition). It exploits the temporal homogeneity of colour of text pixels to filter out some background pixels with similar colour. OCR software enables us to successfully extract the text from an image and convert it into an editable text document. Experimental results evaluated on the Neural network classifier which performance training and testing methods. Training competition dataset show that our approach yields higher precision and performance compared with forefront methods. The experimental results demonstrate the proposed method will provides efficient result than the existing technique.

INTRODUCTION

In recent years the current use of analysing techniques receives intensive attention for the increasing use of digital image capturing devices, such as mobile phones, digital camera, PDAs, static image and dynamic images etc. Text information has inspired great interests, among all the content in an image, since both human and computer can easily understand the wide applications like sign detection and translation, mobile text recognition, license plate reading, content-based web image search and so on, define an integrated image text detection and information extraction system (TDIE, shown in Fig.1) with six stages: text detection, image segmentation, text localization, text extraction, text enhancement and recognition(ORC). ). Among these stages, text detection and localization are complicated to the overall system performance. In multiple methods it have been proposed to address video image and video text detection and localization difficulties, and some of them have concurred impressive results for specific applications. However, speed and precise of the text detection and localization in Multi-colour images is still a competition due to the variations of text font, size, shape, colour and lining orientation, and it is often influenced by complicated background, brighter changes, image distortion and degrading. The existing methods can be roughly categorized into two groups: region-based method and connected component (CC)-based method. Region-based methods define to detect and separate localize text regions by texture analysis. Currently, the vector separated from each local region is fed into a classifier for estimating the likelihood of text then neighbouring text regions are combined to create text blocks. Because text regions from non-text content to detect and localize texts have distinct textural properties accurately even when the images are noisy. On the other hand, Connected component based methods segment an image are difficulties to extracting texts into a set of Connected component group to successively from small Connected component to larger ones, then the final Connected component can be either text or background by analysing their mathematical characteristics. Problem description of the existing methods has reported devastating localization performance; there still remain several problems to solve. For region-based methods, the speed is intensively down and the performance is sensitive to text line orientation. On the other hand, Connected component based methods cannot segment the text components preciously without exact knowledge of text position and scale. Moreover, designing speed and authentic connected component analyser is defiant since there are many non-text regions which are easily perplexed with texts when analysed individually. The proposed method is a combination of two algorithms. In the literature survey many algorithms were developed for segmentation. But they aren't benefit for various types of the text detection images. So we go for a combination of three algorithms that will segment the text image. The proposed method consists of six modules. They are Pre-processing, Edge detection, Stroke width Transformation, Image segmentation, Text localization, Optical character recognition and neural network classifier which consist of both testing and training methods for the future extraction. Multi-Dimensional filter is used to discard the unwanted noise in the YUV transformed image. Edge detection is mentioned by using canny edge detection detector is used to measure edges of the overall boundaries of the horizontal and vertical axis present in the YUV transformed image. Morphological operations are defined by moving a structuring element over a binary image to be modified in such a way that it is centred over an image pixel at some point. The process of removing certain details in an image which is smaller than certain preference shape is called Morphological image processing and the preference shape is called structuring element. Text components are clustered into text lines by words practically with an energy reduction method. Moreover the morphological operation which removes non text areas in the boundaries. Since some background pixels can also have the similar colour, some false stroke areas or character pixels are possible to appear in the output image, which will degrade the recognition rate of OCR (optical character recognition). It exploits the temporal homogeneity of colour of text pixels to filter out some background pixels with similar colour. OCR software enables us to successfully extract the text from an image and convert it into an editable text document. Experimental results evaluated on the Neural network classifier which performance training and testing methods. Training competition dataset show that our approach yields higher precision and performance compared with forefront methods. The experimental results demonstrate the proposed method will provides efficient result than the existing technique.

III.SYSTEM OVERVIEW

It is mentioned earlier that region-based methods or connected component based method can't be able to detect and localize the text, it need some special work based on the local and global information to specify the text detection and recognition task. It's noted that region-based methods and connected component based methods are interrelating. In this the region based method is used to extract the local texture information to segment the image orderly while connected component method can filter out the non-text components and localize the text. The RBG image is converted into YUV image, after that the multidimensional filter is used to reduce the noise in the YUV image. Canny edge detection is used to measure the continuity of the edges in the images. A efficient text detection is proposed using stroke width transformation method based on contours which can effectively remove the interference of non-stroke edges in complex background and the importance of recent feature (inter-frame feature), in the part of caption extraction(detection, localization).The horizontal and vertical histogram basis is used to calculate the luminance and chrominance which defines the background. Moreover the morphological operation which removes non text areas in the boundaries. Since some background pixels can also have the similar colour, some false stroke areas or character pixels are possible to appear in the output image, which will degrade the recognition rate of OCR (optical character recognition). It exploits the temporal homogeneity of colour of text pixels to filter out some background pixels with similar colour. OCR software enables us to successfully extract the text from an image and convert it into an editable text document. Experimental results evaluated on the Neural network classifier which performance training and testing methods. Training competition dataset show that our approach yields higher precision and performance compared with forefront methods. The experimental results demonstrate the proposed method will provides efficient result than the existing technique and the required output will displayed in text notepad file.

IV.PRE-PROCESSING

The pre processing is the process of making an image that is suitable for the next level .It performs filtering of noise and other artefacts in the image and sharpening the edges in the image. RBG to YUV transformation and Resharpening also take place here. After that the pre processing of data uses some Rule-based filtering (inference) system so the image is enhanced. But the possibilities for the noise in Multi-colour image are very less. Here we are using the Multi-dimensional filter for the noise removal.

4.1 RBG Color model

A RBG colour image is an image in which each pixel is specified by three values one each for the red component, blue component, and green components of the pixel scalar. Here each pixel consists of intensity values. The intensity value for single or double arrays, values range from 0 to1 for eight bit integer and values range from zero to two fifty five for sixteen bit integer, the subentries values for sixteen bit integer values range from zero to sixty five thousand five thirty five bits. In RGB colour model, each colour appears in its primary spectral components. The colour of a pixel is made up of three components; red, and blue (RGB), described by their corresponding intensities. Colour component are also known as colour channels or colour planes (components). In the RGB colour model, a colour image can be represented by the intensity function. The intensity of each colour channel is usually stored using eight bits, which indicates that the quantization level is 256. That is, a pixel in a colours image requires a total storage of 24 bits. A 24 bit memory can express as 224 =256Ã-256Ã-256=16777216 distinct colours. The number of colours should adequately meet the display effect of most images. Such images may be called true colour images, where each information of pixel is kept by using a 24-bit memory.

V. EDGE DETECTION

Edges are conspicuous local changes of intensity in an image. Edges typically occur on the boundary of two separate regions between images.

5.1 Goal of edge detection

Produce a line drawing of an edge from an image of that scene. Efficient features can be separate out from the edges of an image (corners, lines, curves).These features are used by top-level computer vision algorithms (e.g., recognition).

5.2 Causes intensity changes

Various physical events cause extremity changes a) Mathematical events1) Object boundary (incoherence in depth and surface colour and texture).2) Bottom layer (incoherence in surface orientation and surface colour and texture).b) Non-mathematical events1) Direct reflection of light, such as a mirror. 2) Shadows. 3) Inter-reflections.

5.3 Steps of edge detection

1) Smoothing: Reduce as much noise as possible, without eliminating the true edges.2) Apply a filter to improve the amount of the edges in the image .3) Detection: determine which edge pixels should be removed as noise and which should be detained. 4) Localization: determine the exact location of an edge (sub-pixel resolution might be required for some applications, that is, innovate the location of an edge to better than the spacing between two pixels). Edge ticking and lining are usually required in this step. Most edge detection methods work on the appropriative conditions that an edge occurs where there is incoherence in the intensity function. In a continuous image of pixels we can calculate the gradient value by taking the variation of greyscale values between adjacent pixels.

YUV transformation

YUV transformed color image is an encoded part of dynamic image and static image. Its main purpose is to reduce the bandwidth of the chrominance component and obtain the luminance component alone in the dynamic and static image. In this YUV models identify the color image were the luminance (brightness) component is define as (Y) components and the chrominance (color) component is define as (UV) components. YUV component specifies the analog encoding and digital encoding of color information in dynamic and static images. YUV signals are significantly developed from Red, Green and Blue color source. Measured values of R, G, and B are defined to produce Y', and measure the overall (brightness) luminance. U and V are computed as scaled differences between Y' and the B and R values.YUV is computed from RGB as follows:

The value of WR is define as a range of 0.2126 and the value of WB is define as a range of 0.0722.Y values are conventionally shifted and aligned value ranges from sixteen to two thirty five moreover than using complete range of zero to two fifty five.

VI. IMAGE SEGMENTATION

Nib lack's local binarization algorithm is adopted due to its high efficiency then the CCs from the gray-level image and non-sensitivity to image degrading. The formula to binaries each pixel x is defined as

The black and white image are converting into a gray scale image using binarization process. A pixel intensity range between 0 to 255 levels contains in a gray scale image. Gray thresholding is done by using the process of binarization. It is also defined as the sets of all pixels above defined a value to white and the below rest of pixels to black in the image and threshold value to binaries the image is very important to the global value for all suitable images. For instance, if threshold value is chosen as X for an image then the pixels of that image having gray scale intensity range equalling X will be binaries to a value 0 and other pixels will be given a value 1. Thus greyscale image will be converted as black & white image. In a black & white image, each pixel has a value of either 0 or 1. The black pixel (digit 1) refers to foreground of the image and white pixel (digit 0) refers to background of the image.

VII. MORPOLOGICAL OPERATION

Morphological operations are defined by moving a structuring element over a binary image to be modified is centred over an image pixel at some point. The process of removing certain details in an image then the smaller certain reference shape is called Morphological image processing and the reference shape is called structuring element. The structuring element in a morphological operation plays an important role with its different shape and size. Shape and size are defined by 0's & 1's. This process is similar to a convolution process. When a structuring element of any size containing complement of 0's & 1's is placed on a binary image, between the structuring element and underlying binary image will be performed a logical operation at each pixel position. The binary result of that logical operation is stored in the output image at that pixel position. Black and white image are converting into a gray scale image using binarization process. A pixel intensity range between 0 to 255 levels contains in a gray scale image. Gray thresholding is done by using the process of binarization. It is also defined as the sets of all pixels above defined a value to white and the below rest of pixels to black in the image and threshold value to binaries the image is very important to the global value for all suitable images. For instance, if threshold value is chosen as X for an image then the pixels of that image having gray scale intensity range equalling X will be binaries to a value 0 and other pixels will be given a value 1. Thus greyscale image will be converted as black & white image. In a black & white image, each pixel has a value of either 0 or 1. The black pixel (digit 1) refers to foreground of the image and white pixel (digit 0) refers to background of the image.

Multi-dimensional filter

4.3 Multi-dimensional filter

The main goal of the multi-dimensional filter is to run through the signal process which discards the unwanted noise present in the dynamic and static images. The multi-dimensional filter runs through the process step by step, alternating each step with the conventional of filter with the operated unstable image. The design of multidimensional filters used for digital group forming conventionally requires regular sampling of the unstable data. Regular sampling in the context of this work means for example spatial sampling on a rectangular or a hexagonal grid, although other regular patterns can be used as well.

4.4 Discarding Noise by Multi-dimensional Filtering

Multi-dimensional filtering is similar to using a suitable filter, in that each o/p pixel value is set to an average of the pixel values in the neighbourhood of the responding i/p pixel. Moreover, the multi-dimensional filter, the value of an o/p pixel is expressed by the median of the neighbouring pixels, either by the mean. The median is much less sensitive than the mean compared to intense values (called statistics). Therefore Multi-dimensional filtering is able to discard these statistics without decreasing the sharpness of the image

Stroke width transformation

In order to find the text from the image, first the image is transferred into edges using canny edge detection method. The threshold value is chosen, that edges of the letters are not diminished while avoiding or generating many edges. Stroke width transformation (SWT) is performed to calculate the approximate width and height of possible stroke where each pixel belongs to starting edge points and tracing along the gradient. In this the letter can have brighter (or) darker color compared to its background. Stroke width transformation (SWT) has to perform by both the forward edge and backward edge along with the gradient edges. The stroke width transformation is a local image which computes with the pixel width that containing the stroke of the likely pixel. Stroke width transformation (SWT) values are more complex in the corners; it's calculated by measuring the height and width of the text in the image. The Stroke width transformation is used to measure the approximate width and height of the text aligned in each rows and column directions in the image.

Canny edge detection

Canny edge detector is the optimal and most widely used algorithm for edge detection. Canny edge detector provides robust edge detection, localization and linking. Compared to other edge detector canny edge detection has better continuity and low contrast lined edges when compared to sobel edge detector and prewit edge detector. Due to the low contrast line edges the text can be easily obtained from the background images. This can have more flexibility and better management which involves the kernel function in canny edge detection algorithm. At each and every edge of the lines is calculated for the output pixels at a particular row. We need input pixels calculate at the row's below and above the text region boundaries. Thus the output at the first and the last rows are undefined and the same happens in case of columns. To incorporate this, the output width and height and the output buffer position changes after each steps.

Writing Services

Essay Writing
Service

Find out how the very best essay writing service can help you accomplish more and achieve higher marks today.

Assignment Writing Service

From complicated assignments to tricky tasks, our experts can tackle virtually any question thrown at them.

Dissertation Writing Service

A dissertation (also known as a thesis or research project) is probably the most important piece of work for any student! From full dissertations to individual chapters, we’re on hand to support you.

Coursework Writing Service

Our expert qualified writers can help you get your coursework right first time, every time.

Dissertation Proposal Service

The first step to completing a dissertation is to create a proposal that talks about what you wish to do. Our experts can design suitable methodologies - perfect to help you get started with a dissertation.

Report Writing
Service

Reports for any audience. Perfectly structured, professionally written, and tailored to suit your exact requirements.

Essay Skeleton Answer Service

If you’re just looking for some help to get started on an essay, our outline service provides you with a perfect essay plan.

Marking & Proofreading Service

Not sure if your work is hitting the mark? Struggling to get feedback from your lecturer? Our premium marking service was created just for you - get the feedback you deserve now.

Exam Revision
Service

Exams can be one of the most stressful experiences you’ll ever have! Revision is key, and we’re here to help. With custom created revision notes and exam answers, you’ll never feel underprepared again.