Spelling suggestions: "subject:"forminformation systems"" "subject:"informationation systems""
401 |
Critical GIS : theorizing an emerging scienceSchuurman, Nadine Cato 05 1900 (has links)
This research takes as its starting point the past decade of critiques from human geographers, and proposes
an alternate model for appraisals of technology. The first section begins with an analysis of the bases and
motivation of external assessments of GIS. A historiographical account reveals that the critical impulse
among human geographers was not static, but evolved to incorporate greater subtlety based on cooperation
with GIS scholars. Critiques from human geographers, nevertheless, had a profound impact on the
discipline, and practitioners of GIS frequently felt that their perspectives on issues including the roots of
GIS, its epistemological bases, and its ethics had been undervalued by critics. A re-analysis of critiques,
from the perspectives of GIS practitioners, investigates objections to critical accounts of the technology.
The second half of the research builds upon existing critiques and responses to them, but asks the question,
"is there a more constructive means to engage with technology, from a theoretical perspective?" Two
contemporary research questions in GIS are investigated, as a means of establishing a preliminary
methodology for critique that engages with GIS at a conceptual, as well as a technical level. Factors that
have influenced the progress of automated generalization are examined in some detail. The argument is
made that both social and digital parameters define the technology, and it is unproductive to focus on one at
the expense of the other. The second research question concerns data models and the extent to which fields
and objects are inevitable. The case is made that a web of historical and scientific justification has
prevented researchers from seeking alternatives to the atomic and plenum views of space. Finally, an
appeal is made for continued theoretical examination of the technology as part of an effort to develop
geographic information science. / Arts, Faculty of / Geography, Department of / Graduate
|
402 |
The estimation of digitizing error and its propagation results in GIS and application to habitat mappingChen, Zhong-Zhong 01 January 1995 (has links)
In order to identify and quantify sources of digitizing error, and to understand how error is carried through a map overlay, three experiments were performed. First, four operators repeatedly digitized eight randomly distributed points eight times. Data were analyzed using parametric statistics and by an error models so that average point digitizing error, the operator error (random error and bias), and the machine error (random error) were estimated; the operators' digitizing characteristics were statistically analyzed. In the second experiment I designed and created a standard coverage with the ARC/INFO GENERATE command. The coverage consists of a set of special geometric entities: a series of differently sized circles, differently shaped triangles and rectangles drawn by PC ARCPLOT. These figures were designed to test the effect of the original map polygons' characteristics on both machine error and operator error, such as area and perimeter, figure shape and geometric entities combination, line curvature, number of vertices selected for representing a line, position of a geometric entity on digitizing board. Several operators repeatedly digitized the coverage six or more times. The operator error (area error and perimeter error) were obtained by subtracting the standard coverage from a digitized coverage. The machine error (area error and perimeter error) were obtained by subtracting the theoretical true coverage from the standard coverage. The results were analyzed statistically. The causation of errors and the operators' digitizing characteristics were further discussed. The third experiment addressed how digitizing errors are propagated through map overlay. In this experiment the digitized coverages created in the second experiment by each operator were overlaid with the ARC/INFO UNION command. Area error, perimeter error, and the numbers of the spurious polygons were collected. The means, sums, maximum, minimum, and standard deviation of area error and perimeter error were obtained. The relationship between area error, perimeter error, and number of spurious polygons of the overlaid coverage were analyzed. This study: (1) focuses on position, as opposed to attribute, error; (2) examines errors in vector-based, not raster-based, GIS; and (3) examines errors caused during the digitizing process, and their propagation through map overlay. The digitizing method is point mode, not stream mode. Results were applied to error management and error reduction to: (1) create a theoretical model which can be used to check quality of the vector source coverages, and to lead users to correctly utilize the GIS data, to prevent them from making unnecessary mistakes; (2) identify some rules to properly use the ARC/INFO ELIMINATE command, and to set MMU (minimum mapping unit) for a particular project.
|
403 |
Execution performance issues in full-text information retrievalBrown, Eric William 01 January 1996 (has links)
The task of an information retrieval system is to identify documents that will satisfy a user's information need. Effective fulfillment of this task has long been an active area of research, leading to sophisticated retrieval models for representing information content in documents and queries and measuring similarity between the two. The maturity and proven effectiveness of these systems has resulted in demand for increased capacity, performance, scalability, and functionality, especially as information retrieval is integrated into more traditional database management environments. In this dissertation we explore a number of functionality and performance issues in information retrieval. First, we consider creation and modification of the document collection, concentrating on management of the inverted file index. An inverted file architecture based on a persistent object store is described and experimental results are presented for inverted file creation and modification. Our architecture provides performance that scales well with document collection size and the database features supported by the persistent object store provide many solutions to issues that arise during integration of information retrieval into more general database environments. We then turn to query evaluation speed and introduce a new optimization technique for statistical ranking retrieval systems that support structured queries. Experimental results from a variety of query sets show that execution time can be reduced by more than 50% with no noticeable impact on retrieval effectiveness, making these more complex retrieval models attractive alternatives for environments that demand high performance.
|
404 |
Adaptive query modification in a probabilistic information retrieval modelHaines, David Leon 01 January 1996 (has links)
There is a vast amount of information available with the aid of computers. It is now far easier to make information available on a CD-ROM or on the Internet than it is to find specific information to fill someone's need. To expect all users to be experts in negotiating the vast amount of available data is unrealistic. Information retrieval systems are designed to help users sort through this sea of text and find the documents that best meet their needs. Information retrieval systems search for documents that match a user's information need based on some user-supplied representation of that need. One important consideration is that the naive users, the ones who most need help, are unlikely to be able to express their need in the best possible way. The specification of the user's query is a difficult task for the user to do well and for the system to understand completely. One important source of information about the user's need is a collection of example documents that illustrate how the user's need can be met. These documents not only provide more information than the user could possibly specify directly, they are also often possible to obtain at a low cost. In this dissertation, a probabilistic theory of how to utilize information available in example documents to automatically improve a user's query and to thereby improve the effectiveness of the information retrieval system is described. This has been done by extending the inference network model of information retrieval developed by Turtle and Croft (47) by adding the mechanism of annotated inference networks and by providing methods to measure and control the contribution of individual components of a query. The research described here not only provides a sound theoretical understanding of how to extract information from example documents but also suggests methods that lead to practical improvements in performance.
|
405 |
Supporting connection mobility in wireless networksRamjee, Ramachandran 01 January 1997 (has links)
A multimedia connection in a wireless network typically utilizes three important network resources: wireless link resources, wired link resources and network server resources. When the users participating in the connection are mobile, these resources must be reallocated as the users move in a manner so that the connection is not disrupted. This dissertation contributes a set of algorithms for supporting connection mobility through efficient and, in certain cases, optimal use of these network resources. In the first part of this thesis, we examine various techniques for allocating wireless channel resources to connections. We define three important practical problems in channel allocation faced by network engineers. We then derive new and optimal admission control policies for each of these problems. We further show that the optimal policies provide significant performance gains over other previously proposed policies. We also develop computationally-efficient algorithms for deploying these optimal policies in real-time at the base-stations. In the second part of this thesis, we examine ways of rerouting the connections of mobile users so that the wired link resources are utilized efficiently. We propose, implement, and experimentally and analytically evaluate the performance of several connection rerouting schemes. Our study shows that one of our schemes is particularly well suited for performing connection rerouting. This scheme operates in two phases: a real-time phase where a reroute operation is executed without causing any disruption to user traffic, and a non-real-time phase where more efficient reroutes are effected. In the third and final part of this thesis, we examine ways of efficiently utilizing the computational resources in the network. We study policies for migrating user agents, which act as proxies for mobile users, as users move. We show that two simple threshold policies that we propose, a Count policy which limits the number of agents in each server and a Distance policy which gives preference to migration of agents that are farther away from their users, deliver excellent performance across a wide range of system parameters and configurations.
|
406 |
Solving the word mismatch problem through automatic text analysisXu, Jinxi 01 January 1997 (has links)
Information Retrieval (IR) is concerned with locating documents that are relevant for a user's information need or query from a large collection of documents. A fundamental problem for information retrieval is word mismatch. A query is usually a short and incomplete description of the underlying information need. The users of IR systems and the authors of the documents often use different words to refer to the same concepts. This thesis addresses the word mismatch problem through automatic text analysis. We investigate two text analysis techniques, corpus analysis and local context analysis, and apply them in two domains of word mismatch, stemming and general query expansion. Experimental results show that these techniques can result in more effective retrieval.
|
407 |
A language modeling approach to information retrievalPonte, Jay Michael 01 January 1998 (has links)
In today's world, there is no shortage of information. However, for a specific information need, only a small subset of all of the available information will be useful. The field of information retrieval (IR) is the study of methods to provide users with that small subset of information relevant to their needs and to do so in a timely fashion. Information sources can take many forms, but this thesis will focus on text based information systems and investigate problems germane to the retrieval of written natural language documents. Central to these problems is the notion of "topic." In other words, what are documents about? However, topics depend on the semantics of documents and retrieval systems are not endowed with knowledge of the semantics of natural language. The approach taken in this thesis will be to make use of probabilistic language models to investigate text based information retrieval and related problems. One such problem is the prediction of topic shifts in text, the topic segmentation problem. It will be shown that probabilistic methods can be used to predict topic changes in the context of the task of new event detection. Two complementary sets of features are studied individually and then combined into a single language model. The language modeling approach allows this problem to be approached in a principled way without complex semantic modeling. Next, the problem of document retrieval in response to a user query will be investigated. Models of document indexing and document retrieval have been extensively studied over the past three decades. The integration of these two classes of models has been the goal of several researchers but it is a very difficult problem. Much of the reason for this is that the indexing component requires inferences as to the semantics of documents. Instead, an approach to retrieval based on probabilistic language modeling will be presented. Models are estimated for each document individually. The approach to modeling is non-parametric and integrates the entire retrieval process into a single model. One advantage of this approach is that collection statistics, which are used heuristically for the assignment of concept probabilities in other probabilistic models, are used directly in the estimation of language model probabilities in this approach. The language modeling approach has been implemented and tested empirically and performs very well on standard test collections and query sets. In order to improve retrieval effectiveness, IR systems use additional techniques such as relevance feedback, unsupervised query expansion and structured queries. These and other techniques are discussed in terms of the language modeling approach and empirical results are given for several of the techniques developed. These results provide further proof of concept for the use of language models for retrieval tasks.
|
408 |
Network support for applications requiring quality of service in heterogeneous environmentsFiroiu, Victor 01 January 1998 (has links)
Group communication, be it one-to-many (such as TV broadcasting) or many-to-many (such as teleconferencing) is becoming increasingly important because it enables the widespread dissemination of information (such as in today's Word Wide Web) and the collaboration between remote groups. This kind of communication can be supported efficiently in digital networks through multicasting, a technique of non-redundant simultaneous data transmission from a sender to a set of receivers. Multicast applications such as voice and video require Quality of Service guarantees (such as maximum packet delay, packet loss probability), which can be provided by reserving network resources. In this dissertation we propose solutions to several critical problems of multicasting in heterogeneous environments: differences in network resource availability, differences in receiver Quality of Service requirements, differences in network resource availability and differences in resource reservation protocols. In the first part of the dissertation we consider the problem of resource reservation for multicast sessions in the context of both network and receiver heterogeneity. We develop centralized and distributed algorithms that accommodate this heterogeneity by performing a differentiated per-link resource reservation. We apply these algorithms in the context of packetized voice and MPEG video multicast connections over wide area networks. We find that our algorithms enable a network to carry as much as a 50% more traffic compared to the case where the network does not accommodate heterogeneity. In the second part of the dissertation we present algorithms for local (link) admission control and resource reservation at an Earliest Deadline First packet scheduler that provides heterogeneous packet delay guarantees at a link. When the data transmission is characterized by piecewise linear traffic envelopes, we show that the algorithms have very low computational complexity and thus, practical applicability. In the third part of the dissertation we focus on resource reservation protocols in the heterogeneous environment of IP over ATM networks. We describe a method for establishing reservations in the ATM network for IP flows (named ATM shortcutting). This method provides better performance to IP flows by avoiding the IP processing of IP packets, and better utilization of ATM network resources. In the last part of the dissertation we quantify the improvement in utilization of IP/ATM network when using ATM shortcutting. We present methods to evaluate this benefit given an IP/ATM network topology, link capacities and traffic patterns. We use this methods in simulation experiments using random networks. These experiments indicate that in many cases ATM shortcutting brings benefits in network utilization when it decreases the average length of network routes.
|
409 |
Discriminating technologies: Personal information in an age of computer profilingElmer, Greg 01 January 2000 (has links)
The doctoral dissertation critically investigates the processes, sites and technologies of consumer profiling. Technologically speaking, the dissertation focuses on the means by which individual consumption patterns and histories (demographic and psychographic data) are automatically solicited into computerized databases and networks. In addition to providing a historical perspective on consumer profiling or solicitation technologies (from product registration cards and mail-in coupons to point-of-sale scanners, smart cards, computer databases and registration sites on the world wide web), the dissertation also attempts to theorize the spatial, international and politically discriminatory dimensions of consumer profiling. In so doing, the dissertation analyzes a number of niche marketing and consumer profiling campaigns that attempt to construct maps of targeted consumer markets in trans-national territories and cyberspace.
|
410 |
A generative theory of relevanceLavrenko, Victor 01 January 2004 (has links)
We present a new theory of relevance for the field of Information Retrieval. Relevance is viewed as a generative process, and we hypothesize that both user queries and relevant documents represent random observations from that process. Based on this view, we develop a formal retrieval model that has direct applications to a wide range of search scenarios. The new model substantially outperforms strong baselines on the tasks of ad-hoc retrieval, cross-language retrieval, handwriting retrieval, automatic image annotation, video retrieval, and topic detection and tracking. Empirical success of our approach is due to a new technique we propose for modeling exchangeable sequences of discrete random variables. The new technique represents an attractive counterpart to existing formulations, such as multinomial mixtures, pLSI and LDA: it is effective, easy to train, and makes no assumptions about the geometric structure of the data.
|
Page generated in 0.1712 seconds