Practical Recognition System for Text Printed on Clear ReflectedMaterial

Text embedded in an image contains useful information for applications in the medical, industrial, commercial, and research fields. While many systems have been designed to correctly identify text in images, no work addressing the recognition of degraded text on clear plastic has been found. This paper posits novel methods and an apparatus for extracting text from an image with the practical assumption: (a) poor background contrast, (b) white, curved, and/or differing fonts or character width between sets of images, (c) dotted text printed on curved reflective material, and/or (d) touching characters. Methods were evaluated using a total of 100 unique test images containing a variety of texts captured from water bottles. These tests averaged a processing time of ∼10 seconds (using MATLAB R2008A on an HP 8510 W with 4 G of RAM and 2.3 GHz of processor speed), and experimental results yielded an average recognition rate of 90 to 93% using customized systems generated by the proposed development.


Introduction
Recognition of degraded characters is a challenging problem in the field of image processing and optical character recognition (OCR).The accuracy and the efficiency of OCR applications are dependent upon the quality of the input image [1][2][3].Security applications and data processing have increased the interest in this area dramatically.Therefore, the ability to replicate and distribute extracted data becomes more important [4,5].
In [6], Jung et al. presented a survey of the term text information extraction (TIE) within an image, by assuming that there is no prior knowledge such as location, orientation, number of characters, font, color, and so on.They also noted that: (a) the variations of text due to differences in size, style, orientation, and alignment, as well as low image contrast and complex background make the problem of automatic text extraction extremely challenging; (b) variety of approaches to TIE from images and video have been proposed for specific applications, such as page segmentation, address block location, license plate location, and content-based image/video indexing.In spite of such extensive studies, it still remains laborious to design a general-purpose OCR system [5] for the reason that there is an abundance of possible sources of variation when extracting text from a shaded or textured background, from low-contrast or complex images, or from images having variations in font size, style, color, and orientation [6].These variations make the problem of automatic TIE extremely arduous and convoluted.
Babu et al. in [14] proposed degraded character recognition from historical printed books using a coupling of vertical and horizontal hidden Markov models.Other studies focused on text localization methods which directly measured pixel-by-pixel differences between two images and quantified them as the average of the squares using a log scale [15].Despite their relative simplicity, these algorithms have the drawback of being unable to assess similarity across different distortion types [16] and possess a greater computational complexity [17].Prior researches have also examined image invariants that handle scaled, rotated, skewed, blurred, and noise-distorted images [18,19].One recent research study dealt with degraded character recognition [20], while another research has focused on single characters [21].Another study has been done on shape-based invariants, which are most commonly used for character recognition [22].Methods of text extraction from colored (TEFC) images are presented in [23].
The process of extracting text from binary images has been presented in [17,24,25].The research by Likforman-Sulem and Sigelle illustrated an example of how to convert a gray image to a binary image and how to locate the text area using connected component techniques [20].Some enhanced methods used morphology [26], while other approaches used algorithms with performance measures.Other researchers employed skew correction and normalization operations for enhancements and parallel classification [27][28][29].
These approaches have been adopted for common applications such as locating an address on postal mail or a courtesy amount on checks mainly because of the simplicity in implementation [30].The current researches are focusing on developing OCR systems for limited domain applications such as postal address reading [31], check sorting [32], tax reading [33], and office automation for text entry [6] and others.However, there are several practical security applications where one needs to have an optical dotted lines/ character recognition system scanned from the reflection of light and the curvature of water bottles images (see Figure 1).
Since the quality of clear reflected material images is much lower than printed or handwritten texts images that employ commonly used character segmentation and recognition algorithms, it is complicated to gain the high success rates by directly applying them to these images.Figure 1 illustrates some bottle text images.It continues to be a challenging problem for an automatic recognition system to detect text from these images because they may face unexpected conditions such as closing loops, spurious branches, and shiny or raised text.Other conditions could be related to the surface on which the text is printed, such as curvature, indentation, or matte finishes.As a result, new OCR tools should naturally be developed to use the unique properties of text in clear reflected material images.
This paper presents an application-oriented dotted text localization and character segmentation algorithms in clear reflected material images.We have successfully applied the proposed methods in developing text localization and character segmentation algorithms on curved plastic bottles images including Ozarka and Dasani bottled water brand names.
The rest of paper is organized as follows: Section 2 explicates the proposed system and covers the image capturing device (camera), enhancement, text localization and binarization; segmentation and Section 3 explains classification and recognition flow which includes the training set, feature vectors, and classification; Section 4 exhibits computer simulation results and comparison to other systems; Section 5 gives some concluding remarks and perspectives.

Proposed System
This section summarizes the proposed system along with components therein.System overview is summarized by the block diagram shown in Figure 2. It includes image capturing, preprocessing, localization, filling, rotation segmentation, and recognition.

Image Capturing.
The capturing device selection satisfies the need for recognition application.The device resolution, focus, aperture, shutter speed, and price are the main factors for selection.Experimental testing shows that camera with a single lens reflex (SLR) has an extremely short shutter speed, high resolution, and excellent close-up capability.In addition of low cost, one of the advantages of this kind of camera is that a user can visualize the image and send it remotely to computer station for analyses.This reduces the amount of memory storage on the device.

Preprocessing.
The goal of preprocessing is to reduce irrelevant information such as noise and increase both contrast and sharpness of images.Different enhancement options are proposed to empower the system and to enable portability.Accuracy and speed of recognition stand as two of the highest concerns in any type of enhancement.The measure of enhancement method is shown in (1).It checks and compares between different options where the original image is divided into m blocks with (k1 × k2 size) before processing each block.The variable α, I (min k, l) w and I (max k, l) w are the maximum and minimum intensity value in a given block, C is a small constant equal to 0.0001 to avoid dividing by 0.
This measure used for choosing the best parameters among set of images for the local enhancement algorithms.The process of enhancement goes through the three stages; the first one is a local one which is applied on the original image, followed by global, and finally background removal method.

Algorithm for Local Image Enhancement.
Local image enhancement: the logarithmic model as presented in [26,34] can be used for both contrast and sharpness of the image.It effectively enhances details in very dark or very bright areas (adjust image sharpness) of the image.The following are the proposed formulas used for computation: where f and g are the generated vectors from (2), γ(M) = AM + B, A and B are constant parameters [35], and f (i, j) and f (i, j) are the normalized original image and the normalized enhanced image.The arithmetic mean of a (3×3) window of the original image is a(i, j), where the parameters α and β controls the contrast.Therefore, from (2) when β > 1, the sharpness of the image increases as β increases, conversely if β < 1, the resulting image is blurred.This flexibility is not available with many conventional methods including histogram equalization.The flexibility and quality of a logarithmic image-processing algorithm make it a very effective technique for text recognition applications.Figure 3   and β = 8.5.The results demonstrate that the algorithm works better for the gray images.Using larger window size will generate an image with sharper edges, which may be desirable for edge detection applications.On the other hand, some of the details of the original image get lost as the window size increases.Larger window size has a disadvantage as the algorithm becomes more computational and time consuming.

Algorithm for Global Image Enhancement.
The proposed global enhancement method is based on converting a typical 24-bit RGB colored image to a grayscale image by performing six conversions methods.Note that each pixel component (e.g., R in RGB) ranges in value from 0 to 255 and is represented by an eight-bit binary number.As well, a digital color image may be easily converted from one color coordinate system to another.This method differs from what was presented by Adolf and Agaian in [23,36] by factoring color intensity log function and applying logarithmic operations.
This paper employs the following conversions methods to map and generate gray images out of red (R), green (G), and blue (B) colors using different weight factors (a K for red, b K for green, C K for blue, where k = 1 : 6).
To avoid division by zero, the corresponding value of the color component of the received image is biased upward by one.A crucial aspect to the text extraction process is the proposed conversion of the received color image into a grayscale in a manner that maximizes the contrast between any text in the image and the rest of the image.The proposed process uses different combinations of weight factors and logarithmic functions before the gray images are converted to binary images using a defined threshold (T) as shown below.
The following conversion process applies with respect to pixels in the following manner.
(i) The first gray image (C 1i j ) is based on the average of red (R i j ) and green (G i j ) components, as in (4).The "i" and "j" represent the row and column pixel location for all images.(ii) The second gray image is a result of dividing the minimum of the three color components values by the average of the color components values, as in (5).
(iii) The third image is a result of dividing the minimum values of the three color components by the maximum values of the color components, as in (6).
(iv) The fourth set selected consists of 6 colors images, which results from all possible combinations of the colors by dividing one of the values by the other values, as in (7), and using the log for the other set of colors.
One means one of the colors (take log function), another means another log function of another color as follows.The fifth set of gray images (6 images) is a result of dividing a sum of a combination of two values of (R, G, and B) by the sum of all of the values, as in (8).The other newly proposed set of images (6 totals) was also generated based on C4 i j .This new set is defined as shown in the ( 9), where = 1 : 6.
For C1 and C4n images, various weight factors are used, and the optimal factors are selected for each set of images experimentally.A nested loop is implemented to find the best combination factors of a, b, and c.The best output image is selected or used, based on the highest intensity difference between text and background.Table 1 shows set of factors for the images we used.For all generated grayscale image as discussed is converted to a new image C new as shown in (10), with an optimal factor of p = 8.This factor is needed so that the nearest values (close pixel intensity) are spread out.
Figure 4 shows samples of enhancement results.The last column in the figure shows the best result (in terms of image quality and intensity difference between text and background).In comparison, the new methods illustrate that conventional enhancement (histogram, linearization, and local deviations) does not yield successful results even for fairly clean images, as shown in Figure 5.
To summarize, the following algorithm is what is been used as default system enhancement.
Background Removal Enhancements Algorithm.This is the last step for enhancement.The steps for the proposed algorithm are as follows.Input Image (output of global enhancement) Step 1: Examine adjacent pixels.If the pixel is significantly darker than the adjacent pixels, it is converted to black; otherwise it is converted to white (I nbkgd ).Using this method the new value is determined based upon local conditions and not on a single value.
The new image (I new ) is generated using background image (I nbkgd ) which is generated from (I grad ) image.
Step 2: In this final step, the enhanced image is generated by combining the newly generated image (I new2 = I gray + I new ) with the average Image (I avg = (I new + I gray )/2).
Output Image: Enhanced image.
Figure 6 demonstrates the proposed background removal enhancements example.The background removal is best suited for this system and is used as a default enhancement.

Finding the Region of Interest (Text Localization).
In this section, a text location algorithm with candidate location is presented.Intensity of a plastic image is a main information source for text detection, but it maintains sensitivity to lighting variations.On the other hand, the gradient of the intensity (edge) is less sensitive to lighting changes.An edgebased feature is used for detection phase with the following assumptions.
(i) The text is designed with high contrast to its background in both color and intensity images.
(ii) The characters in the same context have almost the same size in most of the cases.Most approaches for text identification refer to gray or binary document images.Only recently, some techniques have been proposed for text identification and extraction in color documents.Recently, Wang et al. [37] proposed a color text image binarization technique based on a color quantization procedure and a binary texture analysis; however, this technique neglects the identification of the text region stage that does not include a page layout analysis technique.Thillou and Gosselin [38] proposed a color binarization technique for complex camera-based images based on wavelet denoising and a color clustering with Kmeans; however, in this approach the technique does not include any text extraction stage and is applied only for document images containing already detected text.
The proposed approach uses edge detection and pixel intensity for localizing the edges of the text.After the edges are computed, the number of edges in the x and y directions is calculated.If the result is higher than a certain threshold, it will be considered as a text area.Then, each text area will be binarized using the luminance values.The noise around the digit will greatly affect this method, so a high quality preprocessing step is crucial for the localization approach.Having the right threshold eradicates noise and improves the binary image quality.In this research, gray image is used as the default color map for water bottles.Binarization is based on gray threshold values for a localized area.
Edge detection algorithms have been chosen for the water bottle system.Figure 7  area, (c) is the enhanced image, and (D-I) represents Gaussian, Sobel, Laplacian, log, average, and unsharp detection, respectively.Finding the region of interest algorithm is listed below and shown in Figure 8. Figure 9 show examples of test cases for slabs and water bottles, respectively.

Input:
The candidates' image with text.
Step 1: The original image is converted to gray.This conversion is done using the standard conversion formula that is used for calculating the effective luminance of a pixel as shown in (11).
Step 2: Enhancing the gray image and calculating the sum of column of inverted image.
Step 3: Applying the log (laplacian of Gaussian) edge detection vertically and horizontally.
Step 4: Identifying the connected regions.This is based on scanning the image, pixel-by-pixel from top to bottom and left to right in order to identify connected pixel regions, that is, regions of adjacent pixels which share the same set of intensity values.We use the procedure below for detecting text region of horizontal and vertical direction.
Horizontal = Sum of horizontal edges in selected region Selected region size , Vertical = Sum of vertical edges in selected region Image width . ( Step 5: Calculating local threshold to get a binary image. The original image is segmented into 6 images to get more accurate threshold.The threshold (T) is defined by ( 14) in which the result of multiplying the optimal factor with the standard deviation of the gray image is subtracted from the maximum intensity value of the gray image.
Output: The created region of interest.
2.6.Text Segmentation Approach.This section details the new algorithm for rotation, segmentation, and filling.Before applying the segmentation approach, the rotation and curved text handling approaches are discussed.
After the completion of the edge detection process, the utilization of the connected region procedure occurs before the binary image is generated.
2.6.1.Image Rotation.The goal in this process is to use geometry tools and operations to automatically change the orientation of the images appropriate for classification.Dealing with degraded dotted text on clear plastic poses a challenge for the segmentation of dotted and angled text.To address this problem, the image is binarized and rotated (if necessary).Then, text filling, localization, and segmentation procedures are applied.In the proposed system, the localized binary text image is used as input to the system.
The text in the image can be at different angles.Below is the algorithm pseudocode flow for determining the angle of rotation.Figure 10 summarizes a proposed rotation flow with an example.

Input: The candidates binarized image
Step 1: The algorithm identifies the location of the first nonbackground pixel (white, binary value of "1") in each binary image column starting from the bottom.
Step 2: It then identifies the location of the first nonbackground pixel (white pixel "1") from the top of each column in the previously generated image.Step 3: Next, it identifies the four corners (top, left, bottom, and right), where the text touches.
Step 4: The rotation slope (angle φ) is determined using these four corners and φ is calculated as: Once the angle φ is known, the text is rotated back to the horizontal orientation and the image is refurbished by using the select character procedure to remove all background outside the text.

Output Image: Rotated image
The rotation methods have been tested for two sets of water bottle images where the text is rotated with different angles.Tables 2 and 3 illustrate the results obtained when testing for orientation or rotation invariance for 50 Dasani images and 40 Ozarka images.Each set of images is selected for different bottles, for example set 1 has the bottle with clear day light, and set 2 during night time.

Proposed New Curved Text Handling Approach.
Text images can have curved text from the curvature of the bottle or the printing of the characters.Figure 11 shows the algorithm along with an example wherein the original data is curved in the text image.The idea is to segment the image based on the black column between characters and then crop only text pixels; the boundary of the segmented character image is then touching the characters from four sides.

Segmentation. Extraction of degraded text in an image
remains problematic for current algorithms to resolve.One such problem, as mentioned in [2], is the occurrence of touching characters, and this invalidates the assumption that character repetition patterns in the input text match that of a language mode.Figure 12(a) shows the new algorithm for segmentation from a binary image (text image).The segmentations steps are

Input (Binary image after rotation)
Step 1: Extract text image (text touches the 4 corners) using curved text handling approach.
Step 2: Segment vertically: Based on number of lines, split the image vertically where H is the image height and j is the number of lines.
Step 3: Segment horizontally: for each segmented image in Step 2, segment the image Image (j) based on number of black columns.

Number of characters
where W( j) is the width of each line image and N is the number of black column between characters.Step 4: In this step, every segmented character is checked.If it is larger than a normalized character with by 15% (this value is predetermined based on the experiment), then it is segmented again based on the normalized width, which means the characters are connected.Output image: Characters.

ISRN Machine Vision
The proposed segmentation in a localized binary text image is based on the black column between characters, the character width, and the text box width extracted from the text image.Based on the width of the new text image, the width of each separated image is reviewed.If the width of a segmented character is greater than a normalized width, then the segmented character is known to be composed of several characters and needs further segmentation.A new value for the normalized character width is chosen based on average width of the segmented characters.Figure 12(b) shows segmentation of solid-connected characters from binary image.
2.9.Proposed Fill Dotted Characters.Dotted characters in the segmented images complicate the recognition process.Morphological operations such as dilation, imfill [39], and other methods of interpolation using spline curves or spline fills [39,40] are evaluated and tested to generate solid characters.The larger the gap, the more likely these fill methods yield an erratic result.These methods did not rectify the problem because the character holes/dots are not spaced close enough together.The proposed fill algorithm is applied after segmentation to solidify characters.The proposed fill algorithm is shown in Figure 13.The proposed method is based on the shift and combine operation.
The extracted character image goes through three shift operations: (1) shift up m pixels, (2) shift down with l pixel, and (3) shift left n pixels.Subsequently, a logical check of the original image with the three-shifted images occurs before combining them.The values of l, m, and n are chosen for specific systems after experimental results.Figure 14 shows results of segmentation with filling using Ozarka bottles, this example shows segmentation with filling after an image is split into N images,where N is the number of lines.
The segmentation methods have been tested for the two sets of images that total 90 images.Table 4 displays the results obtained from Ozarka images with a segmentation rate of 96.5%, while Table 5 displays the results obtained from Dasani images with a segmentation rate of 98.1%.The  ( Table 6 summarizes recognition using the proposed fill algorithm, the morphological fill algorithm, no fill algorithm, and spline fill.

Proposed Classification and Recognition Flow
The proposed recognition process uses a structural method for identifying each character.After skeletonization of the extracted character, the feature vector is extracted and compared to the database.The objective of the feature extraction phase is an organized pull out of a set of important features that reduce redundancy in the word image while preserving the key information for recognition.The first feature set is based on global and geometric properties.
The second feature set is based on the analysis of the local properties.
Figure 15 details the recognition algorithm.An advantage of this structural method is its ability to describe the structure of a pattern explicitly.The recognition step extracts the image invariant features and applies the character recognition algorithm using set of binary support vector machine (SVM) classifiers.This is a widely used discriminative classification algorithm (51).
3.1.Feature Vectors.After isolating the characters in an image, a set of properties for each of these characters is determined.
The new set of features for single characters based on their geometric properties.The character image is segmented as shown in Figure 16 into blocks.Each block has its own feature value.The solid lines show bigger blocks and the dotted lines show the smaller blocks.The feature value is calculated in (19), where m and n are the vertical and horizontal group numbers.There values are selected based on experimental result where m and n equal 40.As the process is driven by character properties, the total number of block feature vector (BFV) is based on 26 groups per character; each group is generated based on different areas as seen in (20).A comparison was done from each set of features (all generated groups) and those which varied by less than 30% were removed.Groups: Gr(1), Gr(2),. ..,Gr(n).
The set of groups generated is regrouped, Figure 17 shows the pseudocode used to regroup characters based on the extracted features and an appropriate slope within the same group is reached.This approach demonstrates that characteristics of similar characters/shapes have distinguishable feature vector values.For example, (S and 8) (O, Q), and (L, I, F) are shown in Figure 18.This figure shows that for very close shapes like (O, Q), segmented from different images, the features from set (h1) and set (h2) exhibit an acceptable difference to distinguish between similar shapes.
An example of the regrouping method of feature vector results is shown in Figure 19(a) and all groups are shown in Figure 19(b).The line shows the sum of the features used to distinguish characters from each other based on the slope.
Features vectors were selected to minimize the processing time while maintaining a high accuracy rate.To maximize the recognition accuracy, three other existing set methods were used along with the proposed approach: (1) the first one is based on using 4 levels of the Haar transform [41] with a total of 6 features, (2) the second one is based on skeleton lines of the image (49) with 55 features, and (3) the third one is based on features for the area with 79 features [42].Using a total number of 155 features per character, a target feature vector matrix is built for each system.Having a generalized target matrix for all images can slow and reduce accuracy.

Recognition.
Having many different type feature vectors, the recognition is a multiclass problem.Since SVM supports only two-class recognition, a multiclass system is constructed by combining two-class SVMs as mentioned in [43].
Let {x i , y i }, i = 1, 2, . .., k, y i ∈ {1, −1} and x i ∈ IR n be the training samples where the training vector is x i and y i its corresponding target value.For input pattern x, the decision function of binary classifier is shown in (13). where: where k is the number of learning patterns, the target value of learning pattern y i , x i , b is the bias, and K(x, x i ) is a kernel function which high-dimensional feature space that equals Polynomial kernel which is shown in (23) and the Gaussian radial basis functions kernel which is shown in (24).
In (24), if p is chosen as 1, the polynomial kernel is called linear and if p is chosen as 2, it is called a quadratic kernel.(This approach can be seen in [43]).In our study, for a Gaussian radial basis function kernel, the kernel width, σ 2 is estimated from the variance of the sample vectors.
The SVM is chosen in this research for its processing speed and flexibility in using various distance methods.The classifier is tested using a different set of data with varying rules (nearest, random, consensus) and varying distances (Euclidean, cityblock, cosine, correlation).For the concrete slab and Ozarka test case systems, the highest accuracy results as shown in Table 7, where the Euclidean distance with the nearest rule gives best results (95% matching).

Computer Simulations and Comparison
Experiments were carried out on an unconstrained Dasani bottles for both Ozarka plastic bottled water images as well as concrete slabs, which were captured at a resolution of 300 dpi.The water bottle text consists of 2 lines, each with 15 characters for Ozarka and each with varying numbers of characters for Dasani.Methods were evaluated using a     21 and 22.These were used as test case images from [6].Images in Figure 23 were used for slab test cases.The system was built using MATLAB R2008A installed in a Compaq 8510W workstation with processor speed of 2.3 GHz and 4G of RAM.Table 8 shows the summary results obtained from each set of pictures (systems) when tested for text localization and orientation.The segmentation results for all systems are shown in Table 9.The results in this table demonstrate impressive accuracy, especially when applying the fill algorithm.The results for slab images with varying line pictures are shown in Table 10.The best accuracy with dotted characters was found with the Ozarka application.At a 93% accuracy rate, the elapsed time was 18.94 seconds.Dasani water bottles yielded an average accuracy of 90%, with an elapsed time of 29.5 seconds.Meanwhile, for the concrete slabs, the segmentation and recognition accuracy rate was an average of 98%.
The fill algorithm increased the accuracy by 10-20%.The greatest improvement after using the fill algorithm was for the Dasani bottles where the accuracy rate increased from 80% to 99%.The accuracy rate for Ozarka increased from 91% to 97%.Since the text in the slab images is solid lines, the fill algorithm did not affect the result.
All known research did not discuss the white text printed on reflected material, so comparison is applied to work presented in this research with relevant systems presented in other literatures [11][12][13][44][45][46][47][48][49][50][51].A comparison to their systems (as they reported), mostly from the last five years, as shown in Table 11 demonstrates that the presented system produced better results especially when compared with other related,same language, recognition systems as in [12,[44][45][46][47][48]51].

Conclusion
In this paper, we presented a complete system for automatic detection and recognition of dotted text for application to clear reflected material images.This paper also poses novel tools for extracting text from an image with the following qualities: (a) poor background contrast, (b) white, curved, and/or differing fonts or character width between sets of images, (c) dotted text printed on curved reflective material, and (d) touching characters.Several improvements to current mechanisms were presented, including methods for: (i) text detection and segmentation of touching and/or closely printed characters, (ii) filling of dotted text, (iii) rotation and curvature handling.
We have successfully applied the proposed methods in developing text localization and character segmentation  algorithms on Ozarka and Dasani curved plastic bottled water images.Methods were evaluated using a total of 100 test images containing a variety of texts captured from water bottles.The overall results yielded a recognition rate of 93% accuracy for images captured under the specified conditions.The filling algorithm improved text recognition by more than 20% compared to fill methods presented in the literature.This method shows almost 16% improvements over spline technique and other morphological operations.The main strength of the proposed system lies in its training phase, which does not require any manual segmentation of the data to train the character models.
These tests averaged a processing time of ∼10 seconds (using MATLAB R2008A on an HP 8510W with 4G of RAM and 2.3 GHz of processor speed), and experimental results yielded an average recognition rate of 90% to 93% using customized systems generated by the proposed development.We consider that our approach achieves good performance given that the data correspond to real text bottle images; however, it is hard to compare completely with other approaches since there is no similar investigation.

Figure 1 :
Figure 1: Images white text on white bottle.
displays the proposed algorithm example test case results using a window size of 10 × 10 with α = 5

Figure 3 :
Figure 3: Enhancement test case images (1st row is original, 2nd row is enhanced image).

Figure 4 :
Figure 4: Samples results based on color image algorithm.

Figure 5 :Figure 6 :
Figure 5: Test cases for nonworking enhancement using conventional methods.

Figure 7 :
Figure 7: Edge detection process test case for water bottles.

Figure 13 :After filling line 2 Figure 14 :
Figure 13: Character fill algorithm with an example.
segmentation and recognition equation (18) demonstrates the evaluation rates.Segmentation = Total number of correct segmented characters Total number of images in the text per image , Recognition = Total number of correctly recognized character Total number of detected character in an image .

Figure 16 :
Figure 16: New feature vector group creation example.

3. 3 .
Proposed Training Procedure.The recognition algorithm relies on a set of learned characters and their properties.It compares the features in the segmented image file to the features in the learned set.
Figure 20 shows the training database procedure.Each image can have up to 999 vectors in the training set.The input or the trained image for each character in the training set is either resized or assumed to be 24 by 42 pixels to satisfy the procedure, regardless of whether or not the image is distorted.The training databases include vectors for images of all characters and numbers along with distorted characters for system training.Training images are normalized to a fixed size.

Figure 19 :
Figure 19: Sorted data and character regroup for feature vectors.

Figure 21 :
Figure 21: Sample of Ozkara text case study images.

Figure 22 :
Figure 22: Sample of Dasani text case study images.

Table 2 :
Rotation results/success rate for Dasani.

Table 3 :
Rotation results/success rate for Ozarka.

Table 6 :
Recognition rate comparisons with fill/no fill and spline.

Table 7 :
Classify distance and rule test.

Table 8 :
Systems rotation accuracy with different method.

Table 9 :
System segmentation accuracy with different method.

Table 10 :
Recognition accuracy for slab imaging test case.