Monday, March 30, 2020

License Plate Recognition System Essay Example

License Plate Recognition System Essay Project Report On LICENSE PLATE RECOGNITION SYSTEM PROJECT GROUP MEMBERS A. NISHANTH J. VISHWESH NACHIKET VASANT VAIDYA NAVEEN SUKUMAR TAPPETA R. ANAND UNDER THE GUIDANCE OF PROF. S. R. SATHE DEPARTMENT OF COMPUTER SCIENCE VISVESVARAYA NATIONAL INSTITUTE OF TECHNOLOGY NAGPUR 2008-2009 DEPARTMENT OF COMPUTER SCIENCE VISVESVARAYA NATIONAL INSTITUTE OF TECHNOLOGY NAGPUR 2008-09 CERTIFICATE This Is To Certify That A. Nishanth J. Vishwesh Nachiket Vasant Vaidya Naveen Sukumar Tappeta R. Anand Have Successfully Completed The Project Titled LICENSE PLATE RECOGNITION SYSTEM during the academic year 2008-2009. This dissertation is towards the partial fulfillment in the requirement for the degree of BACHELOR OF TECHNOLOGY (COMPUTER SCIENCE AND ENGINEERING), V. N. I. T, Nagpur. DR. S. R. SATHE DR. O. G. KAKDE (PROJECT GUIDE) (HEAD OF THE DEPT) ACKNOWLEGEMENT This project would not have taken its shape, nor would it have been successfully completed without the support and encouragement of our project guide Prof. S. R. Sathe, who presented us with an invaluable opportunity to work on this project. We take this opportunity to express our sincere gratitude to him. We are extremely indebted to him for the same. We wish to express our heartfelt gratitude to the Department of Computer Science, and the Head of Department, Dr. O. G. Kakde for granting us full freedom towards the utilization of all the facilities in the department. This project has been a very good experience for all of us, which helped us to work together as a good team. We are sure that the knowledge and experience gathered by this will stand us in good stead in our future. We will write a custom essay sample on License Plate Recognition System specifically for you for only $16.38 $13.9/page Order now We will write a custom essay sample on License Plate Recognition System specifically for you FOR ONLY $16.38 $13.9/page Hire Writer We will write a custom essay sample on License Plate Recognition System specifically for you FOR ONLY $16.38 $13.9/page Hire Writer In this thesis a Number Plate Recognition System for Indian License Plates has been explicated. The System comprises of 4 modules for each of the following: The extraction of a region of interest (ROI) containing a car, the extraction of the license plate candidates from these ROIs, the segmentation of the characters from the best candidate and finally using Optical Character Recognition (OCR) on the segmented characters. The results are fed to a grammar checking module after which the license plate number is obtained. The algorithm used to generate the ROIs is a weighted histogram method, the license plate extraction uses vertical edge detection and image morphology, the character segmentation is done using a simple connected component analysis along with heuristics and finally the OCR is implemented using the novel Hierarchical Temporal Memory (HTM) Framework. This thesis proposes a solution to: the problem of License Plate Localization in images with a complicated background, the problem of increasing the effectiveness of morphology and edge based approaches as they are very sensitive to noisy edges and the problem of recognition of characters that have varied size, rotation and have a lot of noise. The system has two components viz. a component that runs on Matlab 7. 6 that performs the image processing and a component on python2. 5. 2 that runs the Numenta Platform for Intelligent Computing(Nupic) for the OCR. The combination of algorithms that we have used proves very effective and has been applied successfully over a test database. ABSTRACT | INTRODUCTION| | | | | INTRODUCTION License plate recognition is a mass surveillance method that uses optical character recognition on images to read the license plates on vehicles. There are three key components to any NPR system: License Plate extraction, Character Segmentation, Optical Character Segmentation. Analyses on the performance of a number of techniques which are used in Number Plate Recognition are discussed below (Refer [1]). LICENSE PLATE LOCALIZATION TECHNIQUES Binary Image Processing Techniques Using Edge and Morphology Based Approaches These techniques are sensitive to noisy edges, however hybrid techniques in this area coupled with prior system information such as distance constraints from the car boost system accuracies to as high as 99. 6% in [2]. In this thesis, a method is proposed to boost the accuracies of these Image processing techniques when applied to images with complicated backgrounds by selecting region of interests. Image Transformations A gabor filter based method is used for texture analysis and license plate detection. These methods are computationally expensive and slow for images with large analysis. In the method that uses Hough transform (HT), edges in the input image are detected first. Then, HT is applied to detect the LP regions. The execution time of the HT requires too much computation when applied to a binary image with great number of pixels. Methods Based On Color and Templates The solutions based on color currently available do not provide a high degree of accuracy in natural scenery, since color is not stable when the lighting conditions change. In addition, as these methods are color based, they are country specific. Methods based on Templates have little effect on Indian License Plates due to rampant non-standardization. The hierarchy introduced in the License Plate Recognition system that has been developed in this thesis viz. Image Candidates containing Car Candidates containing License Plates Enhances the Edge and Morphology based techniques by drastically reducing the number of false Number Plate candidates and increases the accuracy of Plate Localization. OPTICAL CHARACTER RECOGNITION TECHNIQUES Pattern Matching It is suitable for single font, not-rotated, fixed size characters only. It is reported that 90% of central processing unit time was consumed for the computation of the cross-correlation measures between the various templates and the relative sub-image. Hidden Markov Models (HMMs) The disadvantage was the complex procedure of preprocessing and parameterization. It gives a result of 95. 7%. It has a restriction on effective distance of plate recognition system. Hybrid Approach This uses statistical and structural recognition method. It achieves robustness and a high recognition performance. Its success rate is 95. 41%. Neural Networks Multilayered Feed Forward Neural Networks Approach The network has to be trained for many training cycles to attain the good performance. The number of hidden layers as well as number of respective neurons has to be defined by trial and error approach only. This approach cannot handle the noisy part. Self-Organized Neural Networks Based On Kohonen’s The feature maps are implemented to handle noisy, deformed, broken, or incomplete characters acquired from License Plates that were bent and/or tilted with respect to the camera. The method focuses on accuracy at the cost of increased complexity and execution speed. Thus it may be concluded that on analyzing some earlier used techniques for Optical Character Recognition (OCR), many methods involving HMMs , self-organized neural networks based on Kohonen’s, though very robust, have high computational costs involved. We propose a new technique for OCR on Numentas HTM framework. It achieves a high level of rotation and scale invariance in recognition and the hierarchical structure has an added advantage of memory efficiency while making invariant representations of characters. ORGANIZATION The thesis is organized as follows: The First chapter gives an introduction to the architecture of a general License Plate Recognition System and lists the components of the software system that has been developed for this thesis. It also lists the assumptions and the system parameters. The second chapter provides an explanation to the Image Processing and Segmentation techniques that have been used. The third chapter illustrates the process involved in extracting the Regions of Interest (ROI) containing the car, Localization of the number plate from the ROI and describes the segmentation techniques to get the individual characters. The third chapter illustrates the concepts of Hierarchical Temporal Memory (HTM). The fourth chapter describes the HTM Learning Algorithms. The fifth chapter deals with the HTM based Optical Character Recognition (OCR) and Grammar Check Modules implemented for the system. The last chapter lists the results that were obtained by applying the system on various images, conclusion and the future work. | CHAPTER-I| SYSTEM ARCHITECTURE| | | | | CHAPTER-I 1. SYSTEM ARCHITECTURE A Number Plate Recognition system comprises of software (Image Processing and Character Recognition) and hardware components (Custom License Plate Capture Cameras) as shown in the figure. Images from the acquisition device are processed by the Software components and the results are logged or the results could be used to trigger actions such as: opening a gate to a restricted area. Figure 1. 1 The system proposed in the following thesis has been designed specifically for deployment at sites such as: Site Access Control, Car Parks, Freight Logistics Companies, Toll Booths, Airports, Hotels, Industrial Estates, Contract Car Parking, Banks and Stadiums. In this thesis we have focused on the implementation of the software component of a License plate Recognition System and the algorithms proposed allow the system to be used with a variety of low and high quality acquisition devices subject to the set constraints. . 1 SOFTWARE ARCHITECTURE The system has two components namely, the Matlab component and the Python Component. The Matlab Component retrieves images from the Image Database. First, the ROI containing the car are extracted by the Car Candidate Generation Module. Second, the License Plate Localization and Segmentation Module localize the License Plate from the ROI, segment the characters and pass them to Python Component through the Inter Process Communicat ion Module. The Python Component applies the HTM framework based OCR on the characters segmented earlier and passes the results onto a Grammar Check Module. The Grammar Check Module returns a single result to the Matlab Component via Inter Process Communication Module. The results are finally displayed and can be further processed. The following pipeline illustrates the Software Architecture of the system: License Plate Recognition System Image Database Python Component HTM based OCR Grammar Checking Module Inter Process Communication Matlab Component Car Candidate Generation License Plate Localization and Character Segmentation Inter process Communication Display Results Figure 1. 2 1. 1. 1 Assumptions made by the System * The System is designed to provide best results when applied to number plates corresponding to the rules stated below: â€Å"On June 1, 2005, the Government of India introduced High Security Registration (HSR) number plates which are tamper proof. All new motorized road vehicles that came into the market after that need to adhere to the new plates, while existing vehicles have been given two years to comply. Features incorporated include the number plate having a patented chromium hologram; a laser numbering containing the alpha-numeric identification of both the testing agency and manufacturers and a retro-reflective film bearing a verification inscription India at a 45-degree inclination. The characters are embossed on the plate for better visibility. The letters IND are printed in a light shade of blue on the observers left side under the hologram. † Figure 1. 3 * Even though the system works for multiple cars, we focus on retrieving the number plate of the prominent car. The grammar used is based on only Indian license plates and can be extended to other countries also. 1. 1. 2 Parameters taken by the System * Image resolution used: 1200 x 1600 * The images are acquired from a distance between 3 to 6 meters and a height of 2 meters * The input to the system is an RGB image * The output is a string containing the registration number | CHAPTER-II| IMAGE PROCESSING AND SEGMENTATIO N TECHNIQUES| | | | CHAPTER-II 2. IMAGE PROCESSING AND SEGMENTATION TECHNIQUES 2. 1 IMAGE ENHANCEMENT Image Enhancement involves the processing of an image that is better in quality in comparison to the raw image. The following sections explain the various spatial filters that are used in the system. Spatial Filtering involves convolution of an image with a mask. The filter masks are called convolution masks or kernels. The response R of an m x n mask at any point (x,y) in an image is given by R= i=1mnwizi where the w’s are the mask coefficients, the z’s are the values of the image gray levels corresponding to those coefficients and mn is the total number of coefficients in the mask. 2. 1. 1 Average Filter R= 1121i=1121zi Figure 2. 2 Figure 2. 1 The equation above shows an 11 x 11 smoothing filter which is the average of the gray levels of the pixels in the 11 x 11 neighborhood defined by the mask. 2. 1. 2 Median Filter It is necessary to perform a high degree of noise reduction in an image before performing higher-level processing steps, such as edge detection. The median filter is a non-linear digital filtering technique, used to remove noise from images or other signals. It examines a sample of the input and decides if it is representative of the signal. This is performed using a window consisting of an odd number of samples. The values in the window are sorted into numerical order; the median value, the sample in the center of the window, is selected as the output. The oldest sample is discarded, a new sample acquired, and the calculation repeats. Figure 2. 3 For example, suppose that a 3 x 3 neighborhood has values (10, 20, 20, 20, 15, 20, 20, 25, 100). These values are sorted as (10, 15, 20, 20, 20, 20, 20, 25, 100), which results in a median of 20. Thus, the principal function of median filters is to force points with distinct gray levels to be more like their neighbors. In fact, isolated clusters of pixels that are light or dark with respect to their neighbors, and whose area is less than n2/2 (one-half the filter area), are eliminated by an n x n median filter. In this case eliminated means forced to the median intensity of the neighbors. Larger clusters are affected considerably less. Figure 2. 4 2. 2 IMAGE SEGMENTATION In computer vision, segmentation refers to the process of partitioning a digital image into multiple segments (sets of pixels) (Also known as super pixels). The goal of segmentation is to simplify and/or change the representation of an image into something that is more meaningful and easier to analyze. Image segmentation is typically used to locate objects and boundaries (lines, curves, etc. ) in images. More precisely, image segmentation is the process of assigning a label to every pixel in an image such that pixels with the same label share certain visual characteristics. 2. 2. 1 IMAGE MORPHOLOGY AND DILATION Mathematical morphology (MM) is a theory and technique for the analysis and processing of geometrical structures, based on set theory, lattice theory, topology, and random functions. MM is most commonly applied to digital images, but it can be employed as well on graphs, surface meshes, solids, and many other spatial structures. Dilation operation is fundamental to morphological processing. With A and B as sets in Z, the dilation of A by B, denoted A? B is defined as A? B= zBz? A ? ? This equation is based on obtaining the reflection of B about its origin and shifting this reflection by z. The dilation of A by B then is the set of all displacements, Z, such that B and A overlap by at least one element. Set B is commonly referred to as the structuring element in dilation. Figure 2. 5 2. 2. EDGE DETECTION The sobel operator calculates the gradient of the image intensity at each point, giving the direction of the largest possible increase from light to dark and the rate of change in that direction. The result shows how abruptly or smoothly the image changes at that point, and therefore how likely it is that that part of the image represents an edge, as well as how t hat edge is likely to be oriented The gradient of a two-variable function (the image intensity function) is at each image point a 2D vector with the components given by the derivatives in the horizontal and vertical directions. At each image point, the gradient vector points in the direction of largest possible intensity increase, and the length of the gradient vector corresponds to the rate of change in that direction. The result of the Sobel operator at an image point which is in a region of constant image intensity is a zero vector and at a point on an edge is a vector which points across the edge, from darker to brighter values Mathematical Representation: Mathematically, the operator uses two 3? 3 kernels which are convolved with the original image to calculate approximations of the derivatives one for horizontal changes, and one for vertical. If we define A as the source image, and Gx and Gy are two images which at each point contain the horizontal and vertical derivative approximations, the computations are as follows: Gy=+1+2+1000-1-2-1*A and Gx=+10-1+20-2+10-1*A where * here denotes the 2-dimensional convolution operation. The x-coordinate is here defined as increasing in the right-direction, and the y-coordinate is defined as increasing in the down-direction. At each point in the image, the resulting gradient approximations can be combined to give the gradient magnitude, using: G=Gx2+Gy2 Using this information, we can also calculate the gradients direction: ? =arctanGxGy where, for example, ? is 0 for a vertical edge which is darker on the left side. Figure 2. 6 Figure 2. 7 2. 2. 3 EXTRACTION OF BLOB PROPERTIES USING MOMENTS Image moments are certain particular weighted averages (moments) of the image pixels intensities, or functions of those moments, usually chosen to have some attractive property or interpretation. For a 2-D continuous function f(x,y) the moment (sometimes called raw moment) of order (p + q) is defined as Mpq=- - xpyqfx,ydx dy or p, q = 0,1,2, Adapting this to scalar (greytone) image with pixel intensities I(x,y), raw image moments Mij are calculated by Mij=xyxiyjIx,y Area: Area (for binary images) or sum of grey level (for greytone images): M00 Centroid: Centroid: x,y=M10M00,M01M00 Orientation and Length of Major and Minor Axis: Information about image orientation can be derived by first using the second order central moments to construct a covari ance matrix. ?20=? 20? 00=M20M00-x2 ?02=? 02? 00=M02M00-y2 ?11=? 11? 00=M11M00-xy2 The covariance matrix of the image I(x,y) is now covIx,y=? 20? 11? 11? 02 The eigenvectors of this matrix correspond to the major and minor axes of the image intensity, so the orientation can thus be extracted from the angle of the eigenvector associated with the largest eigenvalue. It can be shown that this angle ? is given by the following formula: ? =12arctan2? 11? 20-? 02 The eigenvalues of the covariance matrix can easily be shown to be ? i=? 20+? 022 ±4? 112+? 20-? 0222 and are proportional to the squared length of the eigenvector axes. The relative difference in magnitude of the eigenvalues are thus an indication of the eccentricity of the image, or how elongated it is. The eccentricity is 1-? 2? Bounding Box: The extreme (min and max) x and y values in the blob are recorded and a set of bounding box co-ordinates namely (xmin , ymin) and (xmax , ymax) are generated. The bounding box is the smallest box that completely encloses the image. | CHAPTER-III| LOCALIZATION AND SEGMENTATION| | | | CHAPTER-III 3. LOCALIZATION AND SEGMENTATION 3. 1 INTRODUCT ION To localize the license plate three classes of techniques are made use of, viz. Morphology-based techniques, Edge-based techniques and Histogram-based techniques. Later, the Character- Segmentation is achieved using Connected Component Analysis and by applying Heuristics. The functioning of this part of the system can be further divided into 3 sub-sections. The first deals with Car Candidate Generation, the second deals with License Plate Localization and the third with Character Segmentation. 3. 2 EXTRACTING THE CAR-CANDIDATE REGION The Car-Candidate-Generation is done to focus on the car area in the given image to remove false candidates like name-boards, and noisy edges etc. The following steps were used to extract the car candidates: 1. Figure 3. 1 Figure 3. 2 The RGB Image is converted to a Grayscale image using: I = 0. 2989 * R + 0. 870 * G + 0. 1140 * B. 2. Figure 3. 3 An averaging filter is applied on the Grayscale image by convolving it with a 1111 mask. 3. Figure 3. 4 The Grayscale Image is subtracted from the Averaged image to yield an edge image. The effect induced by Averaged Image Grayscale Image is as follows: Values in the grayscale image higher than the average value (in the 11 x 11 neighborhood) are eliminated while calculating the difference as negative values are converted to 0. Values close to the average value are eliminated while thresholding to convert this image to a black and white image. Thus only pixels with values significantly lower than the average in the 11 x 11 neighborhood are retained. All low intensity parts of the edge in a grayscale image have a high value in the averaged image and hence in the differenced image they retain a good magnitude. 4. Figure 3. 5 The Otsu’s Threshold is used to convert the edge image into a black and white image. 5. Figure 3. 6 The Black and White Image is labeled resulting in each connected component to have a number/label associated with it. The area of these blobs is computed. 6. Figure 3. 7 A Weighted Vertical Histogram is computed for the labeled image. For every column in the Labeled Image, we extract the unique labels. The sum of the areas corresponding to these unique labels is the value of the Histogram for that column. 7. Figure 3. 8 Figure 3. 9 The peaks having magnitude above a threshold (V) and having a distance of VThresh between them are clustered to form vertical strips containing potential car candidates. Depending on the threshold value, eliminate the false car candidates. 8. Figure 3. 10 A weighted horizontal histogram is computed on the vertical strips thus obtained after eliminating the false candidates. 9. Figure 3. 12 Figure 3. 11 The peaks having magnitude above a threshold (H) and having a distance of HThresh between them are clustered to form rectangular strips assuming a threshold of one-third of the average in a strip, and obtain the car candidates. 3. 3 LICENSE PLATE LOCALIZATION The following pipeline explains the steps involved in the localization Once the car candidate is obtained; we localize the number plate in the following manner: 1. Grayscale image corresponding to the potential car candidate region is extracted. 2. To remove Salt and Pepper noise, median filter with a 33 mask is applied. It is useful to reduce speckle noise and salt and pepper noise. Its edge-preserving nature makes it useful in cases where edge blurring is undesirable. 3. Vertical edge detection is done on the image using sobel operator. This yields a grayscale image which is then subjected to a threshold based on RMS estimate of noise. 4. The image is dilated using a rectangular structuring element of size [2 15]. Figure 3. 13 5. Figure 3. 14 The area and bounding box of the connected components are then computed by first labeling the image and then extracting the blob properties. 6. Figure 3. 15 Area and Aspect Ratio heuristics are applied on connected components in the Dilated Image, to get Candidate License Plate (CLP) regions, and the corresponding regions are extracted from the edge image. 3. 4 CHARACTER SEGMENTATION Now the number plate candidates thus obtained are subjected to Character Segmentation by the following steps: The Connected Component Analysis is performed to obtain bounding box of each character. The connected component analysis algorithm is applied to the processed images. So we get the bounding rectangle of the object and the number of the object pixels in these rectangles. The following heuristics are applied to eliminate fake License Plate Candidates: * First the height heuristic is applied such that if the height of the bounding box of each object in the candidate license plate region is at least 0. 4 times the height of the Minor Axis Length of the CLP it was a part of, only then can it be a character. * Second the width heuristic is applied such that if the width of the bounding box of each object in the candidate license plate region is less than 0. 125 times the length of the Major Axis length of the CLP it was a part of, only then can it be a character. Any CLP retrieved from the dilated image should have at least four such objects stated in the above two points to qualify to be final candidate. * Lastly, an equation of a line passing through center of the plate is calculated using the Centroid and the Orientation of the plate candidate. All centroids of the Connected Components in the plate should be less than minimum perpendicular distance (Mi nDist) from the line. This heuristic further eliminates fake candidates. The other segmented regions eliminated, as they don’t qualify to be a character and may just be noise. Figure 3. 16 Thus the segmented characters are obtained. | CHAPTER-IV| HIERARCHICAL TEMPORAL MEMORY (HTM)| | | | CHAPTER-IV 4. HIERARCHICAL TEMPORAL MEMORY (HTM) 4. 1 WHAT IS HTM? Hierarchical Temporal Memory (HTM) is a technology that replicates the structural and algorithmic properties of human brain such as visual pattern recognition, understanding spoken language, recognizing and manipulating objects by touch. HTMs are not programmed and do not execute different algorithms for different problems. Instead, HTMs â€Å"learn† how to solve problems. HTMs are trained by exposing them to sensory data and the capability of the HTM is determined largely by what it has learnt. HTMs are organized as a tree-shaped hierarchy of nodes, where each node implements a common learning and memory function. HTM memory is hierarchical in both space and time to capture and model the structure of the world. HTMs perform the following four basic functions regardless of the particular problem they are applied to: * Discover causes in the world * Infer causes of novel input * Make predictions * Direct behavior . 1. 1 Discover causes in the world Figure 4. 1 Left box in the figure 4. 1 represents a world the HTM is to learn about. The world consists of objects and their relationships. The objects in the world are physical such as cars, people, and buildings. The right box in Figure 4. 1 represents an HTM. It interfaces to its world through one or more senses shown in the middle of the figure. The senses sample some attribute of the world such as light or touch, though the senses used by an HTM do not need to be the same senses humans have. Typically the senses don’t directly detect the objects in the world. Senses typically present an array of data to the HTM, where each element in the array is a measurement of some small attribute of the world. From an HTM’s perspective, there are two essential characteristics of sensory data. First, the sensory data must measure something that is directly or indirectly impacted by the causes in the world. Second, the sensory data must change and flow continuously through time, while the causes underlying the sensory data remain relatively stable. The temporal aspect of sensory data can come from movements or changes of the objects in the world or it can come from movement of the sensory system itself through the world. The HTM’s output is manifest as a set of probabilities for each of the learned causes. This moment-to-moment distribution of possible causes is called a â€Å"belief†. If an HTM knows about ten causes in the world, it will have ten variables representing those causes. The value of these variables – its belief – is what the HTM believes is happening in its world at that instant. Typical HTMs will know about many causes, and as you will see, HTMs actually learn a hierarchy of causes. 4. 1. 2 Infer causes of novel input After an HTM network was trained with the set of training data, the network would be ready for the inference and it gives the output as the probability vector with the highest probability as the category to which the object belongs to. 4. 1. 3 Make predictions HTMs consist of a hierarchy of memory nodes where each node learns causes and forms beliefs. Part of the learning algorithm performed by each node is to store likely sequences of patterns. By combining memory of likely sequences with current input, each node has the ability to make predictions of what is likely to happen next. An entire HTM, being a collection of nodes, also makes predictions. Just as an HTM can infer the causes of novel input, it also can make predictions about novel events. Predicting the future of novel events is the essence of creativity and planning. Leaving the details of how this works for later, we can state now what prediction can be used for. There are several uses for prediction in an HTM, including priming, imagination and planning, and generating behavior. Priming When an HTM predicts what is likely to happen next, the prediction can act as what is called a â€Å"prior probability†, meaning it biases the system to infer the predicted causes. For example, if an HTM were processing text or spoken language, it would automatically predict what sounds, words, and ideas are likely to occur next. This prediction helps the system understand noisy or missing data. Imagination and Planning HTMs automatically predict and anticipate what is likely to happen next. Instead of using these predictions for priming, an HTM’s predictions can be fed back into the HTM as a substitute for sensory data. This process is what humans do when they think. Thinking, imagining, planning the future, and silently rehearsing in our heads are all the same, and achieved by following a series of predictions. HTMs can do this as well. Imagining the future can be valuable in many applications. For example, a car may be equipped with an HTM to monitor nearby traffic, to drive accordingly. 4. 1. 4 Direct behavior An HTM that has learned the causes in its world, and how those causes behave over time, has in essence created a model of its world. Now suppose an HTM is attached to a system which physically interacts with the world. What is important is that the system can move its sensors through its world and/or manipulate objects in its world. In such a system, the HTM can learn to generate complex goal-oriented behavior. As the HTM discovers the causes in its world, it learns to represent its built-in behaviors just as it learns to represent the behaviors of objects in the outside world. From the HTM’s perspective, the system it is connected to is just another object in the world. Through an associative memory mechanism, the HTM-based representations of the built-in behaviors are paired with the mechanisms creating the built-in behaviors themselves. . 2 SIGNIFICANCE OF THE CONCEPT OF HIERARCHY The following reasons explain the concepts to introduce a hierarchical structure: Shared representations lead to generalizat

Saturday, March 7, 2020

Longevity essays

Longevity essays Almost everyone dreams of living forever. It seems so appealing to be able to live life for a very long time, happily and healthily, and going out and enjoying everything that life has to offer once or twice. A lot of people think that its a gift from God to live a long time. But what happens when you have to watch all of your friends and family and all of those who are dear to you are dying all around you. Then it becomes a curse. Some may argue that everyone close to them can live long with them, but what about their family and friends? Realistically speaking, is everyone on the planet going to be able to get his or her hands on a longevity supplement so that everyone can live forever? Unfortunately, the answer to that is no. Many scientists, researchers, and even authors are trying to learn more about longevity and ways to live longer. Author Ben Bova has a new book, Immortality, which is one of the many recent volumes that explain how science is beginning to discover more about aging and death and how to prolong life. "Biologists are learning what causes aging in the cells of your body," he says. He reminds us that the human body has about 100 trillion cells and when they cease to function and reproduce, we do as well. Biologists have learned that each cell has a biological clock that tells the cell when to stop dividing and reproducing and when to die. In 1961, California's Leonard Hayflick discovered that most cells in our bodies divide and reproduce themselves about 50 times before slowing down and dying. It's called the Hayflick Limit. Hayflick insists that despite science's best efforts, humans will only live to be 120, which is the longest a human being has ever lived. "When you drive your brand new automobile off of the automobile showroom floor, you have a clear understanding of the potential of that automobile to live for a certain number of years before you have to buy a new one. That's longevity determination...