Spelling suggestions: "subject:"[een] DOCUMENT"" "subject:"[enn] DOCUMENT""
581 |
Critical analysis of conversion methods between different XML dialectsRoellinghoff, Manuel 12 1900 (has links)
Thesis (MBA)--Stellenbosch University, 2002. / ENGLISH ABSTRACT: The Extensible Markup Language (XML) has become the language of choice for any
kind of data transfer between different business entities. Technologies based on XML
like the Extensible Stylesheet Language, the Extensible Linking Language and SOAP
add vital functionality and make the use of XML in many different business
application scenarios possible.
Companies that start employing XML as their format of choice for data exchange are
often faced with the problem of different XML dialects within their company or with
their business partners. This research project looks at the different solutions that are
available for this problem and evaluates their appropriateness for certain scenarios.
The first viable solution is employing the Extensible Stylesheet Language which
allows for basic transformation on the basis of stylesheets. Companies that use this
technology for simple conversion problems can either develop applications
themselves or rely on tools from third-party vendors. For more advanced conversion
needs a more flexible solution is appropriate. Custom made conversion frameworks
developed by using a fourth generation programming language like Java offer the
advantages of scalability by connecting more than one communication partner,
platform-independency and a high degree of flexibility. Another option that will payoff in the future is the adherence to one of the developing
industry XML standards. Such standards are mostly still in the development phase
and have not been adopted widely yet. But once it becomes clear which standards
are going to survive for each industry the adoption of such a standard will eliminate
the need for any customised conversion and facilitate data exchange between
business partners that adhere to the same standard.
In general it is recommended to research existing XML standards before developing
a proprietary one. When a proprietary format already exists the choice of the
conversion tools and methods depends on the complexity of the transformation.
Simple transformation needs can be solved by employing the XSLT technology while
more advanced needs require the development of a customised conversion
framework. / AFRIKAANSE OPSOMMING: Die Extensible Markup Language (XML) het die taal van keuse geword vir enige
soort van data oordraging tussen verskillende bedryfsentiteite. Tegnologie gebaseer
op XML soos die Extensible Stylesheet Language, die Extensible Linking Language
en Simple Object Access Protocol (SOAP) dra by tot betekenisvolle
waardetoevoeging en maak die gebruik van XML in 'n groot verskeidenheid van
besigheidstoepassings scenarios moontlik.
Ondernemings wat XML as hul formaat van keuse vir data oordraging meer en meer
begin gebruik, word dikwels met die probleem van verskillende XML dialekte intern
tot hul onderneming sowel as ekstern tot hul sakevennote, gekonfronteer. Hierdie
studie ontleed die verskillende oplossings wat beskikbaar is vir die probleem en
evalueer dan die toepaslikheid vir sekere scenarios.
Die eerste uitvoerbare oplossing is die gebruik van die Extensible Stylesheet
Language wat toelaat vir basiese transformasie op die grondslag van stylblaaie.
Ondernemings wat hierdie tegnologie vir eenvoudige omsettingsprobleme wil gebruik
kan óf self toepassings ontwikkel of staatmaak op gereedskap vanaf derde party
verskaffers.
Vir meer gevorderde omsettingsbehoeftes is 'n meer buigsame oplossing geskik.
Doelgemaakte omskakelingsraamwerke wat ontwikkel word deur gebruik te maak van 'n vierde generasie programmeringstaal soos JAVA bied die voordele van
skalering deur meer as een kommunikasievennoot te konnekteer,
platform onafhanklikheid en 'n hoë graad van buigsaamheid.
Nog 'n opsie wat rendement in die toekoms kan lewer is om te bly by een van die
ontwikkelende industrie XML standaarde. Sulke standaarde is meesal nog in die
ontwikkelingsfase en word nog nie tot nou algemeen aanvaar nie. As dit egter begin
duidelik word watter standaarde gaan oorleef vir elke industrie sal die aanvaarding
van sulke standaarde die nodigheid vir doelgemaakte omskakeling elimineer en sal
dit data uitruiling tussen sakevennote wat hou by dieselfde standaarde bevorder.
In die algemeen word dit aanbeveel dat bestaande XML standaarde goed nagevors
word alvorens 'n alleenreg een ontwikkel word. Wanneer 'n alleenregformaat alreeds
bestaan sal die keuse van transformasiegereedskap en metodes afhang van die
kompleksiteit van die transformasie. Eenvoudige transformeerde behoeftes kan
bevredig word deur die XSLT tegnologie te gebruik, terwyl meer gevorderde
behoeftes die ontwikkeling van 'n doelgemaakte omskakelingraamwerk sal benodig.
|
582 |
Evaluation and development of conceptual document similarity metrics with content-based recommender applicationsGouws, Stephan 12 1900 (has links)
Thesis (MScEng (Electrical and Electronic Engineering))--University of Stellenbosch, 2010. / ENGLISH ABSTRACT: The World Wide Web brought with it an unprecedented level of information overload.
Computers are very effective at processing and clustering numerical and binary data,
however, the automated conceptual clustering of natural-language data is considerably
harder to automate. Most past techniques rely on simple keyword-matching techniques
or probabilistic methods to measure semantic relatedness. However, these approaches do
not always accurately capture conceptual relatedness as measured by humans.
In this thesis we propose and evaluate the use of novel Spreading Activation (SA)
techniques for computing semantic relatedness, by modelling the article hyperlink structure
of Wikipedia as an associative network structure for knowledge representation. The
SA technique is adapted and several problems are addressed for it to function over the
Wikipedia hyperlink structure. Inter-concept and inter-document similarity metrics are
developed which make use of SA to compute the conceptual similarity between two concepts
and between two natural-language documents. We evaluate these approaches over
two document similarity datasets and achieve results which compare favourably with the
state of the art.
Furthermore, document preprocessing techniques are evaluated in terms of the performance
gain these techniques can have on the well-known cosine document similarity metric
and the Normalised Compression Distance (NCD) metric. Results indicate that a near
two-fold increase in accuracy can be achieved for NCD by applying simple preprocessing
techniques. Nonetheless, the cosine similarity metric still significantly outperforms NCD.
Finally, we show that using our Wikipedia-based method to augment the cosine vector
space model provides superior results to either in isolation. Combining the two methods
leads to an increased correlation of Pearson p = 0:72 over the Lee (2005) document similarity
dataset, which matches the reported result for the state-of-the-art Explicit Semantic
Analysis (ESA) technique, while requiring less than 10% of the Wikipedia database as
required by ESA.
As a use case for document similarity techniques, a purely content-based news-article
recommender system is designed and implemented for a large online media company.
This system is used to gather additional human-generated relevance ratings which we
use to evaluate the performance of three state-of-the-art document similarity metrics for
providing content-based document recommendations. / AFRIKAANSE OPSOMMING: Die Wêreldwye-Web het ’n vlak van inligting-oorbelading tot gevolg gehad soos nog nooit
tevore. Rekenaars is baie effektief met die verwerking en groepering van numeriese en
binêre data, maar die konsepsuele groepering van natuurlike-taal data is aansienlik moeiliker
om te outomatiseer. Tradisioneel berus sulke algoritmes op eenvoudige sleutelwoordherkenningstegnieke
of waarskynlikheidsmetodes om semantiese verwantskappe te bereken,
maar hierdie benaderings modelleer nie konsepsuele verwantskappe, soos gemeet deur
die mens, baie akkuraat nie.
In hierdie tesis stel ons die gebruik van ’n nuwe aktiverings-verspreidingstrategie (AV)
voor waarmee inter-konsep verwantskappe bereken kan word, deur die artikel skakelstruktuur
van Wikipedia te modelleer as ’n assosiatiewe netwerk. Die AV tegniek word aangepas
om te funksioneer oor die Wikipedia skakelstruktuur, en verskeie probleme wat hiermee
gepaard gaan word aangespreek. Inter-konsep en inter-dokument verwantskapsmaatstawwe
word ontwikkel wat gebruik maak van AV om die konsepsuele verwantskap tussen twee
konsepte en twee natuurlike-taal dokumente te bereken. Ons evalueer hierdie benadering
oor twee dokument-verwantskap datastelle en die resultate vergelyk goed met die van
ander toonaangewende metodes.
Verder word teks-voorverwerkingstegnieke ondersoek in terme van die moontlike verbetering
wat dit tot gevolg kan hê op die werksverrigting van die bekende kosinus vektorruimtemaatstaf
en die genormaliseerde kompressie-afstandmaatstaf (GKA). Resultate
dui daarop dat GKA se akkuraatheid byna verdubbel kan word deur gebruik te maak van
eenvoudige voorverwerkingstegnieke, maar dat die kosinus vektorruimtemaatstaf steeds
aansienlike beter resultate lewer.
Laastens wys ons dat die Wikipedia-gebasseerde metode gebruik kan word om die
vektorruimtemaatstaf aan te vul tot ’n gekombineerde maatstaf wat beter resultate lewer
as enige van die twee metodes afsonderlik. Deur die twee metodes te kombineer lei tot ’n
verhoogde korrelasie van Pearson p = 0:72 oor die Lee dokument-verwantskap datastel.
Dit is gelyk aan die gerapporteerde resultaat vir Explicit Semantic Analysis (ESA), die
huidige beste Wikipedia-gebasseerde tegniek. Ons benadering benodig egter minder as
10% van die Wikipedia databasis wat benodig word vir ESA.
As ’n toetstoepassing vir dokument-verwantskaptegnieke ontwerp en implementeer ons
’n stelsel vir ’n aanlyn media-maatskappy wat nuusartikels aanbeveel vir gebruikers, slegs
op grond van die artikels se inhoud. Joernaliste wat die stelsel gebruik ken ’n punt toe aan
elke aanbeveling en ons gebruik hierdie data om die akkuraatheid van drie toonaangewende
maatstawwe vir dokument-verwantskap te evalueer in die konteks van inhoud-gebasseerde
nuus-artikel aanbevelings.
|
583 |
Kylkontrollrapporter vid Dalslands Miljökontor från 2007-2013 : Vikten av instruktioner för likriktad handläggningFrom, Lisa January 2015 (has links)
The aim of the study was to investigate how Dalsland’s environmental office have managed the yearly reports on fluorinated greenhouse gasses and substances that deplete the ozone layer during the period 2007-2013. The environmental office’s methods, existing routines, present regulations and guidelines have been studied. The purpose of this report was to present suggestions on how the managing of these yearly reports can be improved, and to construct a document of guidelines for the environmental officers, that can be used to make the managing more efficient. The Environmental Office of Dalsland was established 1 January, 2007, when the municipalities of Bengtsfors, Dals Ed, Färgelanda and Mellerud merged their environmental offices together under a joint Environmental committee. The supervision concerning the yearly reports on fluorinated greenhouse gasses and substances that deplete the ozone layer, have varied and the organization within the environmental office have changed over time. There are areas where improvements are needed in order to insure that the auditing of the reports becomes more efficient. The operators should be offered more guidance and information about the current legislation and the demands these rules make on them. Internal routines need to be established on how to register the information that the reports contains, in order to make it easier to extract information for statistical purposes.
|
584 |
DEXTER: Generating Documents by means of computational registersOldham, Joseph D. 01 January 2000 (has links)
Software is often capable of efficiently storing and managing data on computers. However, even software systems that store and manage data efficiently often do an inadequate job of presenting data to users. A prototypical example is the display of raw data in the tabular results of SQL queries. Users may need a presentation that is sensitive to data values and sensitive to domain conventions. One way to enhance presentation is to generate documents that correctly convey the data to users, taking into account the needs of the user and the values in the data. I have designed and implemented a software approach to generating human-readable documents in a variety of domains. The software to generate a document is called a {\em computational register}, or ``register'' for short. A {\em register system} is a software package for authoring and managing individual registers. Registers generating documents in various domains may be managed by one register system. In this thesis I describe computational registers at an architectural level and discuss registers as implemented in DEXTER, my register system. Input to DEXTER registers is a set of SQL query results. DEXTER registers use a rule-based approach to create a document outline from the input. A register creates the output document by using flexible templates to express the document outline. The register approach is unique in several ways. Content determination and structural planning are carried out sequentially rather than simultaneously. Content planning itself is broken down into data re-representation followed by content selection. No advanced linguistic knowledge is required to understand the approach. Register authoring follows a course very similar to writing a single document. The internal data representation and content planning steps allow registers to use flexible templates, rather than more abstract grammar-based approaches, to render the final document, Computational registers are applicable in a variety of domains. What registers can be written is restricted not by domain, but by the original data representation. Finally, DEXTER shows that a single software suite can assist in authoring and management of a variety of registers.
|
585 |
電子病歷彙總工具之設計與實作 / Design and Implementation of a Content Aggregator for Electronic Medical Records林柏維, Lin, Bo Wei Unknown Date (has links)
臺灣電子病歷內容基本格式 (Taiwan Electronic Medical Record Template, TMT)是參考國際相關標準之後,專為臺灣本土需求而設計的電子病歷標準。為了進一步評估TMT的實用性與實施上可能遭遇到的問題,衛生署於民國96年推動了「建構以病人為中心之電子病歷跨院資訊交換環境案」,目前已完成參與該專案各醫院的實地測試工作。
在檢視專案執行的結果後,我們發現TMT資訊系統有三項主要的缺點:一、TMT病歷標準不易閱讀,難以撰寫從醫院醫療資訊系統彙總TMT所需資料的指令集;二、製作TMT系統所須的設定檔程序過於繁複,不但時間攏長而且容易出錯;三、線上實際產生個別病患的TMT資料的時間過長,執行效能有待提昇。有鑑於此,我們設計並實作了一套適用於TMT標準的電子病歷文件產生工具,我們重新設計了規格文件檔及輔助設定檔,並提供了Schema Processor自動化工具產生這些檔案;同時,我們也改進了病歷資料彙總程序,搭配高便利性的設定檔,病歷文件產生工具在執行效能上了有明顯的改善。
在詳細、完整的規格文件檔協助下,資訊人員能更快的了解病歷標準架構及撰寫病歷資料查詢語法,以利產出正確的電子病歷文件;透過自動化工具的輔助,簡化了設定檔的製作程序,改善了耗時且容易出錯的缺點;相較於目前的TMT系統,我們的工具執行效能提昇了80%以上,產出電子病歷文件的時間只要原來的五分之一。 / The Taiwan Electronic Medical Record Template (TMT) proposed by Taiwan Association for Medical Informatics (TAMI) aims to provide a suite of standard forms that will become the common basis for developing electronic medical record (EMR) systems in Taiwan. It is specified in the XML standard for facilitating data interchange. In order to further assess the usefulness of TMT, in 2007 the Department of Health lauched the project "Building of an Information Exchange Environment for Cross-Hospital Digital Medical Record" to put the TMT to a filed test. There are in total eleven hospitals in the project and they all successfully implemented a significant subset of TMT using their hospital information systems (HIS). / However, towards the end of the project, we have identified three major shortcomings of the content aggregator for TMT provided by the TAMI: First, as the TMT Schema is rather complex, it is very difficult for hospital IT staff to prepare the required query instructions to retrieve the data stored in the HIS database. Although there is a XML data mapping tool provided to simplify the mapping process, we found that it did not ease the mapping task as the TAMI staff had expected. Second, the configuration files for preparing a patient’s EMR are too complicated, making the implementation process not only long time but also error-prone. Third, the time required to produce a single sheet of TMT is much longer than planned. There is an urgent need to improve the performance of the content aggregator. / Therefore, we propose to re-engineer the content aggregator of TMT for retrieving the required data from the HIS database. Specifically, we redesigned the specification document files and configuration files, and provided a Schema Processor tool to generate these files in a semi-automatically manner. As a result, the IT staff of hospitals can more quickly understand the structure of TMT Schema and prepare the query instructions effectively. Finally, with the powerful configuration files, our TMT document generator runs much faster than the existing one. According to our experimental results, it enhances the performance of generating a TMT sheet more than 80 percent.
|
586 |
"Ledarens roll för att hålla kugghjulet snurrande" : En studie om fenomenet policyimplementeringPaulsson, Malin, Ärlestål, Alexandra January 2016 (has links)
Syftet med studien är att öka förståelsen för ledarens roll vid implementering av en policy i en organisation. Vi vill visa på hur ledaren kan påverka policyimplementeringen för att bidra till att policyn blir ett levande dokument som genomsyrar hela organisationen.Metod: En kvalitativ forskningsstrategi, med en tvärsnittsstudie som berör fenomenet policyimplementering. Studien har genomförts med en induktiv ansats och insamling av data har skett med kvalitativa intervjuer.Slutsats: Studien visar på att ledaren besitter en viktig roll när det kommer till meningsskapande i en policyimplementering. Ledarens uppgift är att skapa mening i policyprocessen för hela organisationen och dess anställda och tydligt poängtera varför policyn finns. Ledaren ska även interagera policyfrågan i organisationens dagliga arbete och hela tiden uppdatera policyn för att skapa ett levande dokument som genomsyrar hela organisationen.
|
587 |
Samband mellan kritiska framgångsfaktorer och krav på dokumenthantering / The relation between critical success factors and requirements on document managementRinaldo, Pontus January 2016 (has links)
Detta arbete har som syfte att undersöka samband mellan kritiska framgångsfaktorer och krav rörande dokumenthantering. Detta görs för att framgångsfaktorer är något som ökar chanserna för ett lyckat resultat om de är uppfyllda och går det redan i kravfasen att uppfylla dessa kan det spara både tid och pengar. Inledningsvis kommer en litteraturstudie att samla in vilka kritiska framgångsfaktorer det kan finnas gällande dokumenthantering, därefter kommer en fallstudie att genomföras på en verksamhet inom byggbranschen med syfte att reda ut vilka krav de ställer på sin dokumenthantering och slutligen kommer en jämförelse mellan dessa två att göras av författaren till denna rapport med målet att finna och presentera samband. Resultaten visar att det finns ett tydligt samband mellan vissa krav och vissa framgångsfaktorer. Några av kraven uppfyller även flera framgångsfaktorer och vice versa. T ex: Kravet är en ”Väldefinierad teknikdisciplin” och framgångsfaktorerna som uppfylls är ”Definierad lagringsrutin”, ”Användarinvolvering”, ”Tillit till systemet” och ”Träning och utbildning av användare”. / This report is investigating the relation between critical success factors and requirements on document management. The purpose to know that is if a critical success factors are met by a requirement, the chances to succeed increase and in the long runt hat can save both time and money. Initially a literature research will be conducted to sort out the critical success factors, thereafter a case study will collect the requirements from a business in the constructing sector and finally a comparison between these results will be conducted of the author. The results show that there is a relation between certain factors and certain requirements. Some requirements even fulfil more than one factor and vice versa. For example: the requirement ”Well defined engineering discipline” fulfils ”Defined storage routine”, ”User involvement”, ”Trust to the system” and ”Training and education of users”. Keywords: Relations, Requirements, Critical success factors, Document Management.
|
588 |
Obecná chemie na středních školách v experimentálních úlohách / General chemistry at secondary schools in experimental tasksVoborská, Dana January 2010 (has links)
Title General chemistry at high schools in experimental tasks Abstract This work explores the subject of chemical equilibrium in the curriculum of general chemistry in chemical education at secondary schools. The main source is basically analysis of educational systems and mentioned curriculum topic in selected European countries (Great Britain, Switzerland, Germany, Austria, Slovakia, Slovenia) and Czech Republic. The main objective is to prepare instructional chemical equilibrium documents at the level of secondary education in the form of worksheets, that include teaching exercises and chemical experiments. The experiments are designed mainly for home laboratory work, some experiments are useful for elementary school pupils.
|
589 |
Algoritmy detekce obchodních dokumentů podle šablon / Algorithms for business document detection using templatesMichalko, Jakub January 2016 (has links)
Thesis deals with analysis and design system for automatic document recognition. The system examines the document and converts it into text data, and shall be preserved information about the initial position of the word in the original document. These data will then be reviewed and some of them will be assigned their importance. The way the data will be assigned is based on rules which may vary according to user needs. According to the data, their assignment and the importance of their position, the system finds a similar document and, if it identifies the current document examined. Powered by TCPDF (www.tcpdf.org)
|
590 |
Detekce duplicit v rozsáhlých webových bázích dat / Detection of Duplicates in Huge Web DatabasesSadloň, Vladimír January 2012 (has links)
This master thesis analyses the methods used for duplicity document detection and possibilities of their integration with a web search engine. It offers an overview of commonly used methods, from which it chooses the method of approximation of the Jaccard similarity measure in combination with shingling. The chosen method is adapted for implementation in the Egothor web search engine environment. The aim of the thesis is to present this implementation, describe its features, and find the most suitable parameters for the detection to run in real time. An important feature of the described method is also the possibility to make dynamic changes over the collection of indexed documents.
|
Page generated in 0.047 seconds