• Refine Query
  • Source
  • Publication year
  • to
  • Language
  • 588
  • 44
  • 39
  • 37
  • 9
  • 5
  • 3
  • 3
  • 3
  • 3
  • 3
  • 3
  • 2
  • 1
  • 1
  • Tagged with
  • 768
  • 768
  • 185
  • 174
  • 156
  • 135
  • 119
  • 82
  • 71
  • 66
  • 63
  • 63
  • 59
  • 55
  • 48
  • About
  • The Global ETD Search service is a free service for researchers to find electronic theses and dissertations. This service is provided by the Networked Digital Library of Theses and Dissertations.
    Our metadata is collected from universities around the world. If you manage a university/consortium/country archive and want to be added, details can be found on the NDLTD website.
331

Design and evaluation of a memory architecture for a parallel matrix processor array / Nicholas M. Betts.

Betts, Nicholas M. January 2000 (has links)
CD-ROM in pocket on back end paper. / Bibliography: leaves 254-259. / xiv, 259 leaves : ill. ; 30 cm + 1 computer optical disc (4 3/4 in.) / Title page, contents and abstract only. The complete thesis in print form is available from the University Library. / Proposes a specialized matrix processor architecture that targets numerically intensive algorithms that can be cast in matrix terms. / Thesis (Ph.D.)--University of Adelaide, Dept. of Electrical and Electronic Engineering, Advisory Centre for University Education, 2000
332

Design and evaluation of a memory architecture for a parallel matrix processor array / Nicholas M. Betts.

Betts, Nicholas M. January 2000 (has links)
CD-ROM in pocket on back end paper. / Bibliography: leaves 254-259. / xiv, 259 leaves : ill. ; 30 cm + 1 computer optical disc (4 3/4 in.) / Title page, contents and abstract only. The complete thesis in print form is available from the University Library. / Proposes a specialized matrix processor architecture that targets numerically intensive algorithms that can be cast in matrix terms. / Thesis (Ph.D.)--University of Adelaide, Dept. of Electrical and Electronic Engineering, Advisory Centre for University Education, 2000
333

An efficient algorithm for extracting Boolean functions from linear threshold gates, and a synthetic decompositional approach to extracting Boolean functions from feedforward neural networks with arbitrary transfer functions

Peh, Lawrence T. W. January 2000 (has links)
[Formulae and special characters can only be approximated here. Please see the pdf version of the Abstract for an accurate reproduction.] Artificial neural networks are universal function approximators that represent functions subsymbolically by weights, thresholds and network topology. Naturally, the representation remains the same regardless of the problem domain. Suppose a network is applied to a symbolic domain. It is difficult for a human to dynamically construct the symbolic function from the neural representation. It is also difficult to retrain networks on perturbed training vectors, to resume training with different training sets, to form a new neuron by combining trained neurons, and to reason with trained neurons. Even the original training set does not provide a symbolic representation of the function implemented by the trained network because the set may be incomplete or inconsistent, and the training phase may terminate with residual errors. The symbolic information in the network would be more useful if it is available in the language of the problem domain. Algorithms that translate the subsymbolic neural representation to a symbolic representation are called extraction algorithms. I argue that extraction algorithms that operate on single-output, layered feedforward networks are sufficient to analyse the class of multiple-output networks with arbitrary connections, including recurrent networks. The translucency dimensions of the ADT taxonomy for feedforward networks classifies extraction approaches as pedagogical, eclectic, or decompositional. Pedagogical and eclectic approaches typically use a symbolic learning algorithm that takes the network’s input-output behaviour as its raw data. Both approaches construct a set of input patterns and observe the network’s output for each pattern. Eclectic and pedagogical approaches construct the input patterns respectively with and without reference to the network’s internal information. These approaches are suitable for approximating the network’s function using a probably-approximately-correct (PAC) or similar framework, but they are unsuitable for constructing the network’s complete function. Decompositional approaches use internal information from a network more directly to produce the network’s function in symbolic form. Decompositional algorithms have two components. The first component is a core extraction algorithm that operates on a single neuron that is assumed to implement a symbolic function. The second component provides the superstructure for the first. It consists of a decomposition rule for producing such neurons and a recomposition rule for symbolically aggregating the extracted functions into the symbolic function of the network. This thesis makes contributions to both components for Boolean extraction. I introduce a relatively efficient core algorithm called WSX based on a novel Boolean form called BvF. The algorithm has a worst case complexity of O(2 to power of n divided by the square root of n) for a neuron with n inputs, but in all cases, its complexity can also be expressed as O(l) with an O(n) precalculation phase, where l is the length of the extracted expression in terms of the number of symbols it contains. I extend WSX for approximate extraction (AWSX) by introducing an interval about the neuron’s threshold. Assuming that the input patterns far from the threshold are more symbolically significant to the neuron than those near the threshold, ASWX ignores the neuron’s mappings for the symbolically input patterns, remapping them as convenient for efficiency. In experiments, this dramatically decreased extraction time while retaining most of the neuron’s mappings for the training set. Synthetic decomposition is this thesis’ contribution to the second component of decompositional extraction. Classical decomposition decomposes the network into its constituent neurons. By extracting symbolic functions from these neurons, classical decomposition assumes that the neurons implement symbolic functions, or that approximating the subsymbolic computation in the neurons with symbolic computation does not significantly affect the network’s symbolic function. I show experimentally that this assumption does not always hold. Instead of decomposing a network into its constituent neurons, synthetic decomposition uses constraints in the network that have the same functional form as neurons that implement Boolean functions; these neurons are called synthetic neurons. I present a starting point for constructing synthetic decompositional algorithms, and proceed to construct two such algorithms, each with a different strategy for decomposition and recomposition. One of the algorithms, ACX, works for networks with arbitrary monotonic transfer functions, so long as an inverse exists for the functions. It also has an elegant geometric interpretation that leads to meaningful approximations. I also show that ACX can be extended to layered networks with any number of layers.
334

Algorithms for distributed caching and aggregation

Tiwari, Mitul. January 1900 (has links)
Thesis (Ph. D.)--University of Texas at Austin, 2007. / Vita. Includes bibliographical references.
335

On multiple sequence alignment

Wang, Shu, January 1900 (has links)
Thesis (Ph. D.)--University of Texas at Austin, 2007. / Vita. Includes bibliographical references.
336

Fast algorithms for computing statistics under interval uncertainty with applications to computer science and to electrical and computer engineering /

Xiang, Gang, January 2007 (has links)
Thesis (Ph. D.)--University of Texas at El Paso, 2007. / Title from title screen. Vita. CD-ROM. Includes bibliographical references. Also available online.
337

The use of interval-related expert knowledge in processing 2-D and 3-D data with an emphasis on applications to geosciences and biosciences /

Araiza, Roberto, January 2007 (has links)
Thesis (Ph. D.)--University of Texas at El Paso, 2007. / Title from title screen. Vita. CD-ROM. Includes bibliographical references. Also available online.
338

Modeling wildland fire radiance in synthetic remote sensing scenes /

Wang, Zhen. January 2007 (has links)
Thesis (Ph.D.)--Rochester Institute of Technology, 2007. / Typescript. Includes bibliographical references (leaves 155-167).
339

Computerized algorithms to score P1 wave characteristics in the cortical auditory evoked potentials of children with cochlear implants

Wood, Jim, January 2007 (has links)
Thesis (M.S.)--University of Texas at El Paso, 2007. / Title from title screen. Vita. CD-ROM. Includes bibliographical references. Also available online.
340

On the error analysis and implementation of some eigenvalue decomposition and similar value decomposition algorithms /

Ren, Huan. January 1900 (has links)
Thesis (Ph. D.)--University of California, Berkeley, 1996. / Cover title. "June 1997" [i.e. 1996]. Includes bibliographical references.

Page generated in 0.0456 seconds