Spelling suggestions: "subject:"computer algorithms."" "subject:"aomputer algorithms.""
331 |
Design and evaluation of a memory architecture for a parallel matrix processor array / Nicholas M. Betts.Betts, Nicholas M. January 2000 (has links)
CD-ROM in pocket on back end paper. / Bibliography: leaves 254-259. / xiv, 259 leaves : ill. ; 30 cm + 1 computer optical disc (4 3/4 in.) / Title page, contents and abstract only. The complete thesis in print form is available from the University Library. / Proposes a specialized matrix processor architecture that targets numerically intensive algorithms that can be cast in matrix terms. / Thesis (Ph.D.)--University of Adelaide, Dept. of Electrical and Electronic Engineering, Advisory Centre for University Education, 2000
|
332 |
Design and evaluation of a memory architecture for a parallel matrix processor array / Nicholas M. Betts.Betts, Nicholas M. January 2000 (has links)
CD-ROM in pocket on back end paper. / Bibliography: leaves 254-259. / xiv, 259 leaves : ill. ; 30 cm + 1 computer optical disc (4 3/4 in.) / Title page, contents and abstract only. The complete thesis in print form is available from the University Library. / Proposes a specialized matrix processor architecture that targets numerically intensive algorithms that can be cast in matrix terms. / Thesis (Ph.D.)--University of Adelaide, Dept. of Electrical and Electronic Engineering, Advisory Centre for University Education, 2000
|
333 |
An efficient algorithm for extracting Boolean functions from linear threshold gates, and a synthetic decompositional approach to extracting Boolean functions from feedforward neural networks with arbitrary transfer functionsPeh, Lawrence T. W. January 2000 (has links)
[Formulae and special characters can only be approximated here. Please see the pdf version of the Abstract for an accurate reproduction.] Artificial neural networks are universal function approximators that represent functions subsymbolically by weights, thresholds and network topology. Naturally, the representation remains the same regardless of the problem domain. Suppose a network is applied to a symbolic domain. It is difficult for a human to dynamically construct the symbolic function from the neural representation. It is also difficult to retrain networks on perturbed training vectors, to resume training with different training sets, to form a new neuron by combining trained neurons, and to reason with trained neurons. Even the original training set does not provide a symbolic representation of the function implemented by the trained network because the set may be incomplete or inconsistent, and the training phase may terminate with residual errors. The symbolic information in the network would be more useful if it is available in the language of the problem domain. Algorithms that translate the subsymbolic neural representation to a symbolic representation are called extraction algorithms. I argue that extraction algorithms that operate on single-output, layered feedforward networks are sufficient to analyse the class of multiple-output networks with arbitrary connections, including recurrent networks. The translucency dimensions of the ADT taxonomy for feedforward networks classifies extraction approaches as pedagogical, eclectic, or decompositional. Pedagogical and eclectic approaches typically use a symbolic learning algorithm that takes the network’s input-output behaviour as its raw data. Both approaches construct a set of input patterns and observe the network’s output for each pattern. Eclectic and pedagogical approaches construct the input patterns respectively with and without reference to the network’s internal information. These approaches are suitable for approximating the network’s function using a probably-approximately-correct (PAC) or similar framework, but they are unsuitable for constructing the network’s complete function. Decompositional approaches use internal information from a network more directly to produce the network’s function in symbolic form. Decompositional algorithms have two components. The first component is a core extraction algorithm that operates on a single neuron that is assumed to implement a symbolic function. The second component provides the superstructure for the first. It consists of a decomposition rule for producing such neurons and a recomposition rule for symbolically aggregating the extracted functions into the symbolic function of the network. This thesis makes contributions to both components for Boolean extraction. I introduce a relatively efficient core algorithm called WSX based on a novel Boolean form called BvF. The algorithm has a worst case complexity of O(2 to power of n divided by the square root of n) for a neuron with n inputs, but in all cases, its complexity can also be expressed as O(l) with an O(n) precalculation phase, where l is the length of the extracted expression in terms of the number of symbols it contains. I extend WSX for approximate extraction (AWSX) by introducing an interval about the neuron’s threshold. Assuming that the input patterns far from the threshold are more symbolically significant to the neuron than those near the threshold, ASWX ignores the neuron’s mappings for the symbolically input patterns, remapping them as convenient for efficiency. In experiments, this dramatically decreased extraction time while retaining most of the neuron’s mappings for the training set. Synthetic decomposition is this thesis’ contribution to the second component of decompositional extraction. Classical decomposition decomposes the network into its constituent neurons. By extracting symbolic functions from these neurons, classical decomposition assumes that the neurons implement symbolic functions, or that approximating the subsymbolic computation in the neurons with symbolic computation does not significantly affect the network’s symbolic function. I show experimentally that this assumption does not always hold. Instead of decomposing a network into its constituent neurons, synthetic decomposition uses constraints in the network that have the same functional form as neurons that implement Boolean functions; these neurons are called synthetic neurons. I present a starting point for constructing synthetic decompositional algorithms, and proceed to construct two such algorithms, each with a different strategy for decomposition and recomposition. One of the algorithms, ACX, works for networks with arbitrary monotonic transfer functions, so long as an inverse exists for the functions. It also has an elegant geometric interpretation that leads to meaningful approximations. I also show that ACX can be extended to layered networks with any number of layers.
|
334 |
Algorithms for distributed caching and aggregationTiwari, Mitul. January 1900 (has links)
Thesis (Ph. D.)--University of Texas at Austin, 2007. / Vita. Includes bibliographical references.
|
335 |
On multiple sequence alignmentWang, Shu, January 1900 (has links)
Thesis (Ph. D.)--University of Texas at Austin, 2007. / Vita. Includes bibliographical references.
|
336 |
Fast algorithms for computing statistics under interval uncertainty with applications to computer science and to electrical and computer engineering /Xiang, Gang, January 2007 (has links)
Thesis (Ph. D.)--University of Texas at El Paso, 2007. / Title from title screen. Vita. CD-ROM. Includes bibliographical references. Also available online.
|
337 |
The use of interval-related expert knowledge in processing 2-D and 3-D data with an emphasis on applications to geosciences and biosciences /Araiza, Roberto, January 2007 (has links)
Thesis (Ph. D.)--University of Texas at El Paso, 2007. / Title from title screen. Vita. CD-ROM. Includes bibliographical references. Also available online.
|
338 |
Modeling wildland fire radiance in synthetic remote sensing scenes /Wang, Zhen. January 2007 (has links)
Thesis (Ph.D.)--Rochester Institute of Technology, 2007. / Typescript. Includes bibliographical references (leaves 155-167).
|
339 |
Computerized algorithms to score P1 wave characteristics in the cortical auditory evoked potentials of children with cochlear implantsWood, Jim, January 2007 (has links)
Thesis (M.S.)--University of Texas at El Paso, 2007. / Title from title screen. Vita. CD-ROM. Includes bibliographical references. Also available online.
|
340 |
On the error analysis and implementation of some eigenvalue decomposition and similar value decomposition algorithms /Ren, Huan. January 1900 (has links)
Thesis (Ph. D.)--University of California, Berkeley, 1996. / Cover title. "June 1997" [i.e. 1996]. Includes bibliographical references.
|
Page generated in 0.0456 seconds