• Refine Query
  • Source
  • Publication year
  • to
  • Language
  • 80
  • 6
  • 3
  • 3
  • 3
  • 3
  • 3
  • 3
  • 2
  • 2
  • 2
  • 2
  • 1
  • 1
  • 1
  • Tagged with
  • 123
  • 123
  • 123
  • 55
  • 52
  • 37
  • 28
  • 25
  • 24
  • 23
  • 23
  • 21
  • 20
  • 17
  • 16
  • About
  • The Global ETD Search service is a free service for researchers to find electronic theses and dissertations. This service is provided by the Networked Digital Library of Theses and Dissertations.
    Our metadata is collected from universities around the world. If you manage a university/consortium/country archive and want to be added, details can be found on the NDLTD website.
91

Hybrid segmentation on slant & skewed deformation text in natural scene images / Hybrid segmentation on slant and skewed deformation text in natural scene images

Fei, Xiao Lei January 2010 (has links)
University of Macau / Faculty of Science and Technology / Department of Computer and Information Science
92

Gabor filter parameter optimization for multi-textured images : a case study on water body extraction from satellite imagery.

Pillay, Maldean. January 2012 (has links)
The analysis and identification of texture is a key area in image processing and computer vision. One of the most prominent texture analysis algorithms is the Gabor Filter. These filters are used by convolving an image with a family of self similar filters or wavelets through the selection of a suitable number of scales and orientations, which are responsible for aiding in the identification of textures of differing coarseness and directions respectively. While extensively used in a variety of applications, including, biometrics such as iris and facial recognition, their effectiveness depend largely on the manual selection of different parameters values, i.e. the centre frequency, the number of scales and orientations, and the standard deviations. Previous studies have been conducted on how to determine optimal values. However the results are sometimes inconsistent and even contradictory. Furthermore, the selection of the mask size and tile size used in the convolution process has received little attention, presumably since they are image set dependent. This research attempts to verify specific claims made in previous studies about the influence of the number of scales and orientations, but also to investigate the variation of the filter mask size and tile size for water body extraction from satellite imagery. Optical satellite imagery may contain texture samples that are conceptually the same (belong to the same class), but are structurally different or differ due to changes in illumination, i.e. a texture may appear completely different when the intensity or position of a light source changes. A systematic testing of the effects of varying the parameter values on optical satellite imagery is conducted. Experiments are designed to verify claims made about the influence of varying the scales and orientations within predetermined ranges, but also to show the considerable changes in classification accuracy when varying the filter mask and tile size. Heuristic techniques such as Genetic Algorithms (GA) can be used to find optimum solutions in application domains where an enumeration approach is not feasible. Hence, the effectiveness of a GA to automate the process of determining optimum Gabor filter parameter values for a given image dataset is also investigated. The results of the research can be used to facilitate the selection of Gabor filter parameters for applications that involve multi-textured image segmentation or classification, and specifically to guide the selection of appropriate filter mask and tile sizes for automated analysis of satellite imagery. / Thesis (M.Sc.)-University of KwaZulu-Natal, Durban, 2012.
93

Voice input for the disabled /

Holmes, William Paul. January 1987 (has links) (PDF)
Thesis (M. Eng. Sc.)--University of Adelaide, 1987. / Typescript. Includes a copy of a paper presented at TADSEM '85 --Australian Seminar on Devices for Expressive Communication and Environmental Control, co-authored by the author. Includes bibliographical references (leaves [115-121]).
94

A new class of convolutional neural networks based on shunting inhibition with applications to visual pattern recognition

Tivive, Fok Hing Chi. January 2006 (has links)
Thesis (Ph.D.)--University of Wollongong, 2006. / Typescript. Includes bibliographical references: leaf 208-226.
95

API för att tolka och ta fram information från kvitton

Sanfer, Jonathan January 2018 (has links)
Denna rapport redogör för skapandet av ett API som kan extrahera information från bilder på kvitton. Informationen som APIet skulle kunna ta fram var organisationsnummer, datum, tid, summa och moms. Här ingår även en fördjupning om tekniken OCR (optical character recognition) som omvandlar bilder och dokument till text. Examensarbetet utfördes åt Flex Applications AB. Examensarbetet utfördes åt Flex Applications AB. / This report describes the creation of an API that can extract information from pictures of receipts. Registration number, date, time, sum and tax are the information that the API was going to be able to deliver. In this thesis there is also a deepening of the technology OCR (optical character recognition) that transforms pictures and documents to text. The thesis was performed for Flex Applications AB.
96

A Possibilistic Approach To Handwritten Script Identification Via Morphological Methods For Pattern Representation

Ghosh, Debashis 04 1900 (has links) (PDF)
No description available.
97

Detekce objektu ve videosekvencích / Object Detection in Video Sequences

Šebela, Miroslav January 2010 (has links)
The thesis consists of three parts. Theoretical description of digital image processing, optical character recognition and design of system for car licence plate recognition (LPR) in image or video sequence. Theoretical part describes image representation, smoothing, methods used for blob segmentation and proposed are two methods for optical character recognition (OCR). Concern of practical part is to find solution and design procedure for LPR system included OCR. The design contain image pre-processing, blob segmentation, object detection based on its properties and OCR. Proposed solution use grayscale trasformation, histogram processing, thresholding, connected component,region recognition based on its patern and properties. Implemented is also optical recognition method of licence plate where acquired values are compared with database used to manage entry of vehicles into object.
98

OCR of hand-written transcriptions of hieroglyphic text

Nederhof, Mark-Jan January 2016 (has links)
Encoding hieroglyphic texts is time-consuming. If a text already exists as hand-written transcription, there is an alternative, namely OCR. Off-the-shelf OCR systems seem difficult to adapt to the peculiarities of Ancient Egyptian. Presented is a proof-of-concept tool that was designed to digitize texts of Urkunden IV in the hand-writing of Kurt Sethe. It automatically recognizes signs and produces a normalized encoding, suitable for storage in a database, or for printing on a screen or on paper, requiring little manual correction. The encoding of hieroglyphic text is RES (Revised Encoding Scheme) rather than (common dialects of) MdC (Manuel de Codage). Earlier papers argued against MdC and in favour of RES for corpus development. Arguments in favour of RES include longevity of the encoding, as its semantics are font-independent. The present study provides evidence that RES is also much preferable to MdC in the context of OCR. With a well-understood parsing technique, relative positioning of scanned signs can be straightforwardly mapped to suitable primitives of the encoding.
99

Scale Invariant Object Recognition Using Cortical Computational Models and a Robotic Platform

Voils, Danny 01 January 2012 (has links)
This paper proposes an end-to-end, scale invariant, visual object recognition system, composed of computational components that mimic the cortex in the brain. The system uses a two stage process. The first stage is a filter that extracts scale invariant features from the visual field. The second stage uses inference based spacio-temporal analysis of these features to identify objects in the visual field. The proposed model combines Numenta's Hierarchical Temporal Memory (HTM), with HMAX developed by MIT's Brain and Cognitive Science Department. While these two biologically inspired paradigms are based on what is known about the visual cortex, HTM and HMAX tackle the overall object recognition problem from different directions. Image pyramid based methods like HMAX make explicit use of scale, but have no sense of time. HTM, on the other hand, only indirectly tackles scale, but makes explicit use of time. By combining HTM and HMAX, both scale and time are addressed. In this paper, I show that HTM and HMAX can be combined to make a com- plete cortex inspired object recognition model that explicitly uses both scale and time to recognize objects in temporal sequences of images. Additionally, through experimentation, I examine several variations of HMAX and its
100

Ensemble Methods for Historical Machine-Printed Document Recognition

Lund, William B. 03 April 2014 (has links) (PDF)
The usefulness of digitized documents is directly related to the quality of the extracted text. Optical Character Recognition (OCR) has reached a point where well-formatted and clean machine- printed documents are easily recognizable by current commercial OCR products; however, older or degraded machine-printed documents present problems to OCR engines resulting in word error rates (WER) that severely limit either automated or manual use of the extracted text. Major archives of historical machine-printed documents are being assembled around the globe, requiring an accurate transcription of the text for the automated creation of descriptive metadata, full-text searching, and information extraction. Given document images to be transcribed, ensemble recognition methods with multiple sources of evidence from the original document image and information sources external to the document have been shown in this and related work to improve output. This research introduces new methods of evidence extraction, feature engineering, and evidence combination to correct errors from state-of-the-art OCR engines. This work also investigates the success and failure of ensemble methods in the OCR error correction task, as well as the conditions under which these ensemble recognition methods reduce the Word Error Rate (WER), improving the quality of the OCR transcription, showing that the average document word error rate can be reduced below the WER of a state-of-the-art commercial OCR system by between 7.4% and 28.6% depending on the test corpus and methods. This research on OCR error correction contributes within the larger field of ensemble methods as follows. Four unique corpora for OCR error correction are introduced: The Eisenhower Communiqués, a collection of typewritten documents from 1944 to 1945; The Nineteenth Century Mormon Articles Newspaper Index from 1831 to 1900; and two synthetic corpora based on the Enron (2001) and the Reuters (1997) datasets. The Reverse Dijkstra Heuristic is introduced as a novel admissible heuristic for the A* exact alignment algorithm. The impact of the heuristic is a dramatic reduction in the number of nodes processed during text alignment as compared to the baseline method. From the aligned text, the method developed here creates a lattice of competing hypotheses for word tokens. In contrast to much of the work in this field, the word token lattice is created from a character alignment, preserving split and merged tokens within the hypothesis columns of the lattice. This alignment method more explicitly identifies competing word hypotheses which may otherwise have been split apart by a word alignment. Lastly, this research explores, in order of increasing contribution to word error rate reduction: voting among hypotheses, decision lists based on an in-domain training set, ensemble recognition methods with novel feature sets, multiple binarizations of the same document image, and training on synthetic document images.

Page generated in 0.1346 seconds