• Refine Query
  • Source
  • Publication year
  • to
  • Language
  • 56
  • 23
  • 13
  • 8
  • 6
  • 5
  • 5
  • 4
  • 3
  • 2
  • 1
  • 1
  • 1
  • 1
  • Tagged with
  • 220
  • 220
  • 85
  • 73
  • 48
  • 43
  • 32
  • 25
  • 24
  • 22
  • 20
  • 18
  • 17
  • 16
  • 16
  • About
  • The Global ETD Search service is a free service for researchers to find electronic theses and dissertations. This service is provided by the Networked Digital Library of Theses and Dissertations.
    Our metadata is collected from universities around the world. If you manage a university/consortium/country archive and want to be added, details can be found on the NDLTD website.
91

Zavedení doménové struktury v rozsáhlé firmě / Implementation of the Domain Structure in a Large Company

Hošek, David January 2017 (has links)
This master`s thesis deals with domain structure. The main concern of this work is analysis and implementation of the Domain Structure in a company. Domain structure will be introduced because of the current unsatisfactory solutions and increase information security. The name of the company is not specified in this thesis.
92

Dropbox & Co, alles schon ge-cloud?

Syckor, Jens January 2014 (has links)
Cloudspeicherdienste sind zu einem Standard für den Austausch großer Datenmengen in virtuellen Gemeinschaften geworden, sowohl im privaten Umfeld als auch im öffentlichen Bereich. Einfache Bedienbarkeit sowie nahtlose Integration in Applikationen, Betriebssystemen und Endgeräten sind wesentliche Bausteine dieses Siegeszuges.
93

Self-Healing Cellular Automata to Correct Soft Errors in Defective Embedded Program Memories

Voddi, Varun 01 December 2009 (has links)
Static Random Access Memory (SRAM) cells in ultra-low power Integrated Circuits (ICs) based on nanoscale Complementary Metal Oxide Semiconductor (CMOS) devices are likely to be the most vulnerable to large-scale soft errors. Conventional error correction circuits may not be able to handle the distributed nature of such errors and are susceptible to soft errors themselves. In this thesis, a distributed error correction circuit called Self-Healing Cellular Automata (SHCA) that can repair itself is presented. A possible way to deploy a SHCA in a system of SRAM-based embedded program memories (ePM) for one type of chip multi-processors is also discussed. The SHCA is compared with conventional error correction approaches and its strengths and limitations are analyzed.
94

A Quantitative Analysis of Memory Controller Page Policies

Blackmore, Matthew 28 February 2013 (has links)
Two common goals in computing system design are increasing performance and decreasing power consumption. DRAM-based memory subsystems are a major component of both system performance and power consumption. Memory controllers employ strategies to efficiently schedule DRAM operations to reduce latency and to utilize DRAM low power modes when possible. One of the most important of these is the page policy, which determines when to close pages in DRAM. An effective DRAM memory controller page policy is important to minimizing power consumption and increasing system performance. This thesis explores the impact memory controller page policy has on performance as measured by the number of page-hits minus page-misses and estimated average memory access latency. I captured real-time DDR3 command and address memory traces for the SPEC CPU2006 benchmarks under three memory controller page policies: closed page, fixed open-page, and Intel's adaptive open-page [1]. Traces were captured using a programmable memory traffic analyzer (PMTA), a device interposed between the DIMM slot and DDR3 DIMM on the motherboard. The memory traces for each benchmark were analyzed to determine the absolute number of page-hits and page-misses that occurred. In software post-processing I simulated a theoretically perfect "oracle" page policy for each captured trace to compare the efficiency of existing policies. The SPEC CPU 2006 benchmarks under the oracle page policy for each trace exhibited an average increase in the number of page-hits minus page-misses of 280.3% and an average decrease in the average memory latency of 11.1%. Two new adaptive open-page policies are proposed and simulated using the captured memory traces. These proposed policies result in an average increase of 74.8% and 62.4% in the number of page-hits minus page-misses over Intel's adaptive open-page policy and an average decrease in the average memory latency of 3.8% and 3.4%.
95

Ranked Similarity Search of Scientific Datasets: An Information Retrieval Approach

Megler, Veronika Margaret 04 June 2014 (has links)
In the past decade, the amount of scientific data collected and generated by scientists has grown dramatically. This growth has intensified an existing problem: in large archives consisting of datasets stored in many files, formats and locations, how can scientists find data relevant to their research interests? We approach this problem in a new way: by adapting Information Retrieval techniques, developed for searching text documents, into the world of (primarily numeric) scientific data. We propose an approach that uses a blend of automated and curated methods to extract metadata from large repositories of scientific data. We then perform searches over this metadata, returning results ranked by similarity to the search criteria. We present a model of this approach, and describe a specific implementation thereof performed at an ocean-observatory data archive and now running in production. Our prototype implements scanners that extract metadata from datasets that contain different kinds of environmental observations, and a search engine with a candidate similarity measure for comparing a set of search terms to the extracted metadata. We evaluate the utility of the prototype by performing two user studies; these studies show that the approach resonates with users, and that our proposed similarity measure performs well when analyzed using standard Information Retrieval evaluation methods. We performed performance tests to explore how continued archive growth will affect our goal of interactive response, developed and applied techniques that mitigate the effects of that growth, and show that the techniques are effective. Lastly, we describe some of the research needed to extend this initial work into a true "Google for data".
96

Entropy reduction of English text using variable length grouping

Ast, Vincent Norman 01 July 1972 (has links)
It is known that the entropy of English text can be reduced by arranging the text into groups of two or more letters each. The higher the order of the grouping the greater is the entropy reduction. Using this principle in a computer text compressing system brings about difficulties, however, because the number of entries required in the translation table increases exponentially with group size. This experiment examined the possibility of using a translation table containing only selected entries of all group sizes with the expectation of obtaining a substantial entropy reduction with a relatively small table. An expression was derived that showed that the groups which should be included in the table are not necessarily those that occur frequently but rather occur more frequently than would be expected due to random occurrence. This was complicated by the fact that any grouping affects the frequency of occurrence of many other related groups. An algorithm was developed in which the table originally starts with the regular 26 letters of the alphabet and the space. Entries, which consist of letter groups, complete words, and word groups, are then added one by one based on the selection criterion. After each entry is added adjustments are made to account for the interaction of the groups. This algorithm was programmed on a computer and was run using a text sample of about 7000 words. The results showed that the entropy could easily be reduced down to 3 bits per letter with a table of less than 200 entries. With about 500 entries the entropy could be reduced to about 2.5 bits per letter. About 60% of the table was composed of letter groups, 42% of single words and 8% of word groups and indicated that the extra complications involved in handling word groups may not be worthwhile. A visual examination of the table showed that many entries were very much oriented to the particular sample. This may or may not be desirable depending on the intended use of the translating system.
97

SADDAS; a self-contained analog to digital data acquisition system.

Petersen, Walter Anton 01 January 1972 (has links)
SADDAS, a. Self-contained Analog to Digital Data Acquisition System, converts analog voltage inputs to formatted BCD (binary coded decimal digital magnetic tape. SADDAS consists of a 16 channel multiplexer, a 17 bit (4 digits + sign) 40 microsecond analog to digital converter, a 512 byte 8 bit core memory, a 30 IPS (inches per second) digital tape recorder at a density of 556 cpi (characters per inch), and a controller which integrates these instruments into a flexible and easy-to-use system. Sampling rates in excess of 360 samples per second may be used when converting seven channels of data, such as IRIG (Inter Range Instrumentation Group) analog magnetic tapes.
98

Efficient Social Network Data Query Processing on MapReduce

Liu, Liu 01 January 2013 (has links) (PDF)
Social network data analysis becomes increasingly important today. In order to improve the integration and reuse of their data, many social networks start to apply RDF to present the data. Accordingly, one common approach for social network data analysis is to employ SPARQL to query RDF data. As the sizes of social networks expand rapidly, queries need to be executed in parallel such as using the MapReduce framework. However, the state-of-the-art translation from SPARQL queries to MapReduce jobs mainly follows a two layer rule, in which SPARQL is first translated to SQL join, is not efficient. In this thesis, we introduce two primitives to enable automatic translation from SPARQL to MapReduce, and to enable efficient execution of the SPARQL queries. We use multiple-join-with-filter to substitute traditional SQL multiple join when feasible, and merge different stages in the MapReduce query workflow. The evaluation on social network benchmarks shows that these two primitives can achieve up to 2x speedup in query running time compared with the original two layer scheme.
99

Design of Multilayer Optical Media: Organic Photovoltaics and Optical Data Storage

Valle, Brent 19 August 2013 (has links)
No description available.
100

Nanoscale Surface Patterning and Applications: Using Top-Down Patterning Methods to Aid Bottom-Up Fabrication

Pearson, Anthony Craig 31 August 2012 (has links) (PDF)
Bottom-up self-assembly can be used to create structures with sub-20 nm feature sizes or materials with advanced electrical properties. Here I demonstrate processes to enable such self-assembling systems including block copolymers and DNA origami, to be integrated into nanoelectronic devices. Additionally, I present a method which utilizes the high stability and electrical conductivity of graphene, which is a material formed using a bottom-up growth process, to create archival data storage devices. Specifically, I show a technique using block copolymer micelle lithography to fabricate arrays of 5 nm gold nanoparticles, which are chemically modified with a single-stranded DNA molecule and used to chemically attach DNA origami to a surface. Next, I demonstrate a method using electron beam lithography to control location of nanoparticles templated by block copolymer micelles, which can be used to enable precise position of DNA origami on a surface. To allow fabrication of conductive structures from a DNA origami template, I show a method using site-specific attachment of gold nanoparticles to and a subsequent metallization step to form continuous nanowires. Next, I demonstrate a long-term data storage method using nanoscale graphene fuses. Top-down electron beam lithography was used to pattern atomically thin sheets of graphene into nanofuses. To program the fuses, graphene is oxidized as the temperature of the fuse is raised via joule heating under a sufficiently high applied voltage. Finally, I investigate the effect of the fuse geometry and the electrical and thermal properties of the fuse material on the programming requirements of nanoscale fuses. Programming voltages and expected fuse temperatures obtained from finite element analysis simulations and a simple analytical model were compared with fuses fabricated from tellurium, a tellurium alloy, and tungsten.

Page generated in 0.0717 seconds