• Refine Query
  • Source
  • Publication year
  • to
  • Language
  • 92
  • 23
  • 21
  • 17
  • 16
  • 15
  • 9
  • 6
  • 2
  • 2
  • 1
  • 1
  • 1
  • 1
  • 1
  • Tagged with
  • 211
  • 211
  • 49
  • 44
  • 39
  • 38
  • 36
  • 31
  • 30
  • 30
  • 21
  • 20
  • 17
  • 16
  • 16
  • About
  • The Global ETD Search service is a free service for researchers to find electronic theses and dissertations. This service is provided by the Networked Digital Library of Theses and Dissertations.
    Our metadata is collected from universities around the world. If you manage a university/consortium/country archive and want to be added, details can be found on the NDLTD website.
171

Open data in Swedish municipalities? : Value creation and innovation in local public sector organizations / Öppna data i svenska kommuner? : Värdeskapande och innovation i lokala offentliga organisationer

TAHERIFARD, ERSHAD January 2021 (has links)
Digital transformation is highlighted as a way of solving many of the problems and challenges that the public sector faces in terms of cost developments and increased demands for better public services. Open data is a strategic resource that is necessary for this development to takeplace and the municipal sector collects information that could be published to create value inmany stages. Previous research believes that economic value is generated through new innovative services and productivity gains, but also social values such as increased civic participation and transparency. But despite previous attempts to stimulate open data, Sweden is far behind comparable countries and there is a lack of research that looks at exactly how these economic values should be captured. To investigate why this is the case and what role open datahas in value creation in the municipal sector, this study has identified several themes through qualitative interviews with an inductive approach. The study resulted in a deeper theoretical analysis of open data and its properties. By considering it as a public good, it is possible to use several explanatory models to explain its slow spread and but also understand the difficult conditions for value capture which results in incentive problems. In addition, there are structural problems linked to legislation and infrastructure that hamper the dissemination of open data and its value-creating role in the municipal sector. / Digital transformationen lyfts som ett sätt att lösa många av de problem och utmaningar som den offentliga sektorn står inför gällande kostnadsutveckling och ökade krav på bättre samhällsservice. Öppna data är en sådan strategisk resurs som är nödvändig för att dennautveckling ska ske och kommunsektorn samlar på sig information som skulle kunna publiceras för att skapa värden i många led. Dels menar tidigare forskning att ekonomiska värden kan genereras genom nya innovativa tjänster och produktivitetsökningar, men även sociala värden som ökad medborgardelaktighet och transparens. Men trots tidigare försök att stimulera öppna data, ligger Sverige långt efter jämförbara länder och det saknas forskning som tittar på exakt hur dessa ekonomiska värden ska fångas. För att undersöka varför så är fallet och vilken roll öppna data har på värdeskapande i kommunsektorn har denna studie genom kvalitativa intervjuer med en induktiv ansats identifierat flertalet teman. Studien resulterade i en djupare teoretisk analys av öppna data och dess egenskaper. Genom att betrakta det som en kollektiv vara går det att använda flera förklaringsmodeller för att förklara dess långsamma spridning och förstå de svåra förutsättningarna för värdefångst vilket resulterar i incitamentsproblem. Till det finns det strukturella problem kopplat till lagstiftning och infrastruktur som hämmarspridningen av öppna data och dess värdeskapande roll i kommunsektorn.
172

Open Waters - Digital Twins With use of Open Data and Shared Design for Swedish Water Treatment Plants / Open Waters: Digitala tvillingar med öppen data och delad design för svensk vattenrening

Nyirenda, Michael January 2020 (has links)
Digital twins (DTs) are digital copies of a physical system that incorporates the system environment, interactions, etc. to mirror the system accurately in real time. As effective decision support systems (DSS) in complex multivariate situations, DTs could be the next step in the digitalization of water management. This study is done in cooperation with the Open Waters project group at the Swedish environmental research institute (IVL). The aim of the project group is to investigate the possibility to realize DTs with the use of open data (OD), and shared design (SD), in Swedish water management while also promoting ecosystems for innovation in virtual environments. This study will aid the project group by bridging the gap between project stakeholders and water managers. A DSS developed by IVL for automatic dosage of coagulants in water treatment which is based on the same industry 4.0 technology as DTs will be evaluated as a possible starting point for DTs, OD, and SD. In depth interviews were held with representatives from water management, and experts in DTs, OD, and SD. This was to identify key opportunities and threats, and to understand water managers perception and opinion of the project. This is complimented by a brief review of Swedish water management, and the international state of DTs. There were 4 main opportunities and threats. Challenges and goals are very similar between different WTPs    Water managers are already collaborating to reach common goals    WTPs are unique in terms of treatment steps and composition/properties of raw water WTPs are objects of national security which raises questions regarding safety when digitalization is discussed. / Digitala tvillingar (DT) är digitala kopior av fysiska system som inkluderar systemets miljö, interaktioner, etc. för att noggrant spegla systemet i realtid. Som effektiva beslutsunderlag i komplexa, multivariabla situationer har DT fått uppmärksamhet inom vattensektorn och kan vara nästa steg i industrins digitalisering. Denna studie utförs i samarbete med svenska miljöinstitutets (IVLs) projektgrupp Open Waters. Syftet är att utforska möjligheten att förverkliga DT med hjälp av öppna data (OD) och delad design (SD) i den svenska vattensektorn, samt att främja innovationsekosystem i virtuella miljöer. Målet med denna studie är att överbygga klyftan mellan projektgruppen och dess målgrupp. Till hjälp kommer den IVL utvecklade DOS-modellen för automatisk dosering av fällningskemikalier för vattenrening. Denna är baserad på samma industri 4.0 teknologi som DT och ses som en startpunkt för DT, OD, och SD. Djupintervjuer hölls med representanter inom vattensektorn, såväl som experter inom DT, OD, och SD. Målet med detta var att identifiera centrala möjligheter och hot för projektet, samt för att förstå vattensektorns bild och åsikt av DT. Detta kompletteras med en övergripande genomgång av den svenska vattensektorn, och DT. 4 huvudsakliga möjligheter och hot identifierades.    Utmaningar och mål är väldigt lika mellan olika vattenverk    Det sker redan samarbeten i vattensektorn när gemensamma mål identifieras    Vattenverk är unika i förhållande till reningssteg och råvatten Vattenverk är skyddsobjekt vilket höjer frågor gällande informationssäkerhet när digitalisering diskuteras.
173

New Opportunities in Crowd-Sourced Monitoring and Non-government Data Mining for Developing Urban Air Quality Models in the US

Lu, Tianjun 15 May 2020 (has links)
Ambient air pollution is among the top 10 health risk factors in the US. With increasing concerns about adverse health effects of ambient air pollution among stakeholders including environmental scientists, health professionals, urban planners and community residents, improving air quality is a crucial goal for developing healthy communities. The US Environmental Protection Agency (EPA) aims to reduce air pollution by regulating emissions and continuously monitoring air pollution levels. Local communities also benefit from crowd-sourced monitoring to measure air pollution, particularly with the help of rapidly developed low-cost sampling technologies. The shift from relying only on government-based regulatory monitoring to crowd-sourced effort has provided new opportunities for air quality data. In addition, the fast-growing data sciences (e.g., data mining) allow for leveraging open data from different sources to improve air pollution exposure assessment. My dissertation investigates how new data sources of air quality (e.g., community-based monitoring, low-cost sensor platform) and model predictor variables (e.g., non-government open data) based on emerging modeling approaches (e.g., machine learning [ML]) could be used to improve air quality models (i.e., land use regression [LUR]) at local, regional, and national levels for refined exposure assessment. LUR models are commonly used for predicting air pollution concentrations at locations without monitoring data based on neighboring land use and geographic variables. I explore the use of crowd-sourced low-cost monitoring data, new/open dataset from government and non-government sponsored platforms, and emerging modeling techniques to develop LUR models in the US. I focus on testing whether: (1) air quality data from community-based monitoring is feasible for developing LUR models, (2) air quality data from non-government crowd-sourced low-cost sensor platforms could supplement regulatory monitors for LUR development, and (3) new/open data extracted from non-government sponsored platforms could serve as alternative datasets to traditional predictor variable sources (e.g., land use and geographic features) in LUR models. In Chapter 3, I developed LUR models using community-based sampling (n = 50) for 60 volatile organic compounds (VOC) in the city of Minneapolis, US. I assessed whether adding area source-related features improves LUR model performance and compared model performance using variables featuring area sources from government vs. non-government sponsored platforms. I developed three sets of models: (1) base-case models with land use and transportation variables, (2) base-case models adding area source variables from local business permit data (government sponsored platform), and (3) base-case models adding Google point of interest (POI) data for area sources. Models with Google POI data performed the best; for example, the total VOC (TVOC) model had better goodness-of-fit (adj-R2: 0.56; Root Mean Square Error [RMSE]: 0.32 µg/m3) as compared to the permit data model (0.42; 0.37) and the base-case model (0.26; 0.41). This work suggests that VOC LUR models can be developed using community-based samples and adding Google POI could improve model performance as compared to using local business permit data. In Chapter 4, I evaluated a national LUR model using annual average PM2.5 concentrations from low-cost sensors (i.e., PurpleAir platform) in 6 US urban areas (n = 149) and tested the feasibility of using low-cost sensor data for developing LUR models. I compared LUR models using only the PurpleAir sensors vs. hybrid LUR models (combining both the EPA regulatory monitors and the PurpleAir sensors). I found that the low-cost sensor network could serve as a promising alternative to fill the gaps of existing regulatory networks. For example, the national regulatory monitor-based LUR (i.e., CACES LUR developed as part of the Center for Air, Climate, and Energy Solutions) may fail to capture locations with high PM2.5 concentrations and the within-city spatial variability. Developing LUR models using the PurpleAir sensors was reasonable (PurpleAir sensors only: 10-fold CV R2 = 0.66, MAE = 2.01 µg/m3; PurpleAir and regulatory monitors: R2 = 0.85, MAE = 1.02 µg/m3). I also observed that incorporating PurpleAir sensor data into LUR models could help capture within-city variability and merit further investigation on areas of disagreement with the regulatory monitors. This work suggests that the use of crowd-sourced low-cost sensor networks for LUR models could potentially help exposure assessment and inform environmental and health policies, particularly for places (e.g., developing countries) where regulatory monitoring network is limited. In Chapter 5, I developed national LUR models to predict annual average concentrations of 6 criteria pollutants (NO2, PM2.5, O3, CO, SO2 and PM10) in the US to compare models using new data (Google POI, Google Street View [GSV] and Local Climate Zone [LCZ]) vs. traditional geographic variables (e.g., road lengths, area of built land) based on different modeling approaches (partial least square [PLS], stepwise regression and machine learning [ML] with and without Kriging effect). Model performance was similar for both variable scenarios (e.g., random 10-fold CV R2 of ML-kriging models for NO2, new vs. traditional: 0.89 vs. 0.91); whereas adding the new variables to the traditional LUR models didn't necessarily improve model performance. Models with kriging effect outperformed those without (e.g., CV R2 for PM2.5 using the new variables, ML-kriging vs. ML: 0.83 vs. 0.67). The importance of the new variables to LUR models highlights the potential of substituting traditional variables, thus enabling LUR models for areas with limited or no data (e.g., developing countries) and across cities. The dissertation presents the integration of new/open data from non-government sponsored platform and crowd-sourced low-cost sensor networks in LUR models based on different modeling approaches for predicting ambient air pollution. The analyses provide evidence that using new data sources of both air quality and predictor variables could serve as promising strategies to improve LUR models for tracking exposures more accurately. The results could inform environment scientists, health policy makers, as well as urban planners interested in promoting healthy communities. / Doctor of Philosophy / According to the US Centers for Disease Control and Prevention (CDC), a healthy community aims at preventing disease, reducing health gaps, and creating more accessible options for a wider population. Outdoor air pollution has been evidenced to cause a wide range of diseases (e.g., cardiovascular diseases, respiratory diseases, diabetes and adverse birth outcome), ranking as the top 10 health risks in the US. Thus, improving understanding of ambient air quality is one of the common goals among environmental scientists, urban planners, health professionals, and local residents to achieving healthy communities. To understand air pollution exposures in different areas, US Environmental Protection Agency (EPA) has regulatory monitors for outdoor air pollution measurements across the country. For locations without these regulatory monitors, land use regression (LUR) models (one type of air quality models) are commonly employed to make a prediction. Usually, information including number of people, location of bus stops, and type of roads are shared online from government websites. These datasets are often used as significant predictor variables for developing LUR models. Questions remain on whether new air quality data and alternative land use data from non-government sources could improve air quality modeling. In recent years, local communities have been actively involving in air pollution monitoring using rapidly developed low-cost sensors and sampling campaigns with the help of local residents. In the meantime, advances in data sciences make open data much easier to acquire and use, particularly from non-government sponsored platforms. My dissertation aims to explore the use of new data sources including community-based low-cost monitoring data and open dataset from non-government websites in LUR modes based on emerging modeling techniques (e.g. machine learning) to predict air pollution levels in the US. I first built LUR models for volatile organic compounds (VOC: organic chemicals with a high vapor pressure at room temperature [e.g., Benzene]) based on community-based sampling data in the City of Minneapolis, US. I added information on number of neighboring gas stations, dry cleaners, paint booths, and auto shops from both the local government and Google website into the model and compared the model performance for both data sources (Chapter 3). Then, I used PM2.5 data from a non-government website (PurpleAir low-cost sensors) for 6 US cities evaluating an existing air quality model that used air quality data from government websites. I further developed LUR models using the PurpleAir PM2.5 data to see whether this non-government source of low-cost sensor data could be as reasonable as the government data for LUR model development. I finally extracted new/open data from non-government sponsored platforms (e.g., Google products and local climate zone [LCZ: a map that describes the development patterns of land, such as high-rise vs. low-rise or trees vs. sands]) in the US to investigate if these data sources can be used to alternate the land use and geographic data often used in national LUR model development. I found that: (1) adding information (e.g., number of neighboring gas stations) from non-government sponsored sources (e.g., Google) could improve the air quality model performance for VOCs, (2) integrating non-government low-cost PM2.5 sensor data into government regulatory monitoring data to develop LUR models could improve model performance and offer more insights on the air pollution exposure, (3) new/open data from non-government sponsored platforms could be used to replace the land use and geographic data previous obtained from government websites for air quality models. These findings mean that air quality data and street-level land use characteristics could serve as alternative data sources and are capable of developing better air quality models for promoting healthy communities.
174

Applying Time-Valued Knowledge for Public Health Outbreak Response

Schlitt, James Thomas 21 June 2019 (has links)
During the early stages of any epidemic, simple interventions such as quarantine and isolation may be sufficient to halt the spread of a novel pathogen. However, should this opportunity be missed, substantially more resource-intensive, complex, and societally intrusive interventions may be required to achieve an acceptable outcome. These disparities place a differential on the value of a given unit of knowledge across the time-domains of an epidemic. Within this dissertation we explore these value-differentials via extension of the business concept of the time-value of knowledge and propose the C4 Response Model for organizing the research response to novel pathogenic outbreaks. First, we define the C4 Response Model as a progression from an initial data-hungry collect stage, iteration between open-science-centric connect stages and machine-learning centric calibrate stages, and a final visualization-centric convey stage. Secondly we analyze the trends in knowledge-building across the stages of epidemics with regard to open and closed access article publication, referencing, and citation. Thirdly, we demonstrate a Twitter message mapping application to assess the virality of tweets as a function of their source-profile category, message category, timing, urban context, tone, and use of bots. Finally, we apply an agent-based model of influenza transmission to explore the efficacy of combined antiviral, sequestration, and vaccination interventions in mitigating an outbreak of an influenza-like-illness (ILI) within a simulated military base population. We find that while closed access outbreak response articles use more recent citations and see higher mean citation counts, open access articles are published and referenced in significantly greater numbers and are growing in proportion. We observe that tweet viralities showed distinct heterogeneities across message and profile type pairing, that tweets dissipated rapidly across time and space, and that tweets published before high-tweet-volume time periods showed higher virality. Finally, we saw that while timely responses and strong pharmaceutical interventions showed the greatest impact in mitigating ILI transmission within a military base, even optimistic scenarios failed to prevent the majority of new cases. This body of work offers significant methodological contributions for the practice of computational epidemiology as well as a theoretical grounding for the further use of the C4 Response Model. / Doctor of Philosophy / During the early stages of an outbreak of disease, simple interventions such as isolating those infected may be sufficient to prevent further cases. However, should this opportunity be missed, substantially more complex interventions such as the development of novel pharmaceuticals may be required. This results in a differential value for specific knowledge across the early, middle, and late stages of epidemic. Within this dissertation we explore these differentials via extension of the business concept of the time-value of knowledge, whereby key findings may yield greater benefits during early epidemics. We propose the C4 Response Model for organizing research regarding this time-value. First, we define the C4 Response Model as a progression from an initial knowledge collection stage, iteration between knowledge connection stages and machine learning-centric calibration stages, and a final conveyance stage. Secondly we analyze the trends in knowledge-building across the stages of epidemics with regard to open and closed access scientific article publication, referencing, and citation. Thirdly, we demonstrate a Twitter application for improving public health messaging campaigns by identifying optimal combinations of source-profile categories, message categories, timing, urban origination, tone, and use of bots. Finally, we apply an agent-based model of influenza transmission to explore the efficacy of combined antiviral, isolation, and vaccination interventions in mitigating an outbreak of an influenza-like-illness (ILI) within a simulated military base population. We find that while closed access outbreak response articles use more recent citations and see higher mean citation counts, open access articles are growing in use and are published and referenced in significantly greater numbers. We observe that tweet viralities showed distinct benefits to certain message and profile type pairings, that tweets faded rapidly across time and space, and that tweets published before high-tweet-volume time periods are retweeted more. Finally, we saw that while early responses and strong pharmaceuticals showed the greatest impact in preventing influenza transmission within military base populations, even optimistic scenarios failed to prevent the majority to new cases. This body of work offers significant methodological contributions for the practice of computational epidemiology as well as a theoretical grounding for the C4 Response Model.
175

Harnessing the Value of Open Data through Business Model Adaptation : A Multiple Case Study on Data-Intelligence Service-Providers

Thalin, Simon, Svennefalk, Marcus January 2024 (has links)
Purpose - The objective of this study is to explore how Data-Intelligence Service-Providers (DISP) can adapt existing Business Model (BM) dimensions to leverage the potential value and mitigate the emerging challenges Open Data (OD) introduces. Method – By developing a multiple case study, we intend to qualitatively explore what BM practices DISPs employ when incorporating OD. Interviews are conducted in multiple phases with a total of 25 interviews and results generated using a thematic analysis. Findings – Through empirical investigation and analysis of DISPs actions and strategies, the study uncovers how these firms navigate challenges and opportunities presented by OD. By portraying the strategies across three BM dimensions—value creation, delivery, and capture—this study identifies six key practices that help DISPs competitively differentiate themselves in the OD environment. The identified practices include Use-case understanding and Data-driven Service Innovation for value creation, Enhanced Data Delivery and Collaborative Data Optimization for value delivery, and AdjustedRevenue Model and Market Expansion for value capture. Implications – In our contribution to existing literature, we present empirical evidence spanning across all dimensions of the BM, shedding light on the competitive advantages facilitated by OD. Additionally, through identifying key practices, this thesis uncovers several areas where there is a lack of understanding on ODs impact in a commercial context. Specifically, by solely focusing on the perspective of DISPs, we offer detailed insight into how these practices are practically unfolding. Furthermore, the thesis presents a framework categorizing practices based on priority and ecosystem dependency. This framework delineates certain practices that are considered fundamental when incorporating OD while also recognizing their intricate requirement of involving external parties, offering managers a visual overview of how to systematically adapt their BMs to incorporate OD into their services. In addition, we manage to address the common distortions about OD by offering a thorough theoretical foundation and defining it clearly within a commercial context, making this complex topic more accessible and better understood. Limitations and future research – As this study is limited to data-providers and DISPs, this thesis advocates for exploring end-user perspectives in future research deemed crucial for gathering a comprehensive understanding of their needs and interactions with OD solutions to solidify findings in this study. Additionally, it is encouraged that future research should investigate misalignments between data-providers and DISPs (e.g. regulatory and technical matters) which currently, are leading to massive inefficiencies in data supply chains. Understanding these issues and implementing strategies to address them can optimize OD resource utilization, thereby facilitating greater innovative potential for service-providers leveraging it.
176

Conocimiento, uso y reutilización de los datos abiertos en la ciencia española

Vidal Cabo, Christian 18 July 2022 (has links)
[ES] El Gobierno Abierto es un modo de política pública que se basa en los pilares de colaboración y participación ciudadana, transparencia y rendición de cuentas y derecho de acceso a la información pública. De la mano de las tecnologías de la información y las comunicaciones, gobiernos y administraciones llevan a cabo iniciativas de apertura de datos, movimiento conocido como Open Data (Datos Abiertos). Las plataformas digitales donde estas entidades ponen a disposición de la sociedad civil los datos son conocidas como portales de datos abiertos. Se trata de fuentes de información donde los conjuntos de datos son potencialmente reutilizables, con cualquier fin y sin ningún tipo de restricción, únicamente de referencia de autoría de los datos. La comunidad científica, personal altamente cualificado dentro de la sociedad, pueden llegar a ser reutilizadores potenciales de estas fuentes de información. El producto derivado se traduce en producción científica: artículos, usos de datos abiertos en proyectos de investigación, comunicaciones y docencia. Este estudio aborda, por una parte, el conocimiento que tienen los investigadores e investigadoras acerca de los datos abiertos. Por otra, el uso y la reutilización de los datos abiertos para generar conocimiento científico. Para llevar a cabo el estudio se ha desarrollado una metodología cuantitativa. Se ha elaborado una encuesta, distribuida en un bloque inicial de contexto con 6 preguntas y 6 bloques de carácter técnico con 24 preguntas, es decir, un cuestionario con 30 preguntas. Se obtienen un total de 783 respuestas, procedentes de 34 provincias españolas. Los investigadores e investigadoras proceden de 47 universidades españolas y 21 centros de investigación, y existe representación 19 áreas de investigación de la Agencia Estatal de Investigación. Con los datos obtenidos a través de esta metodología cuantitativa, se procesan, se normalizan y se lleva a cabo un análisis. Además, con los datos se desarrolla una plataforma para visualizar los resultados de la encuesta. / [CA] El Govern Obert és una mena de política basada en els pilars de col·laboració i participació ciutadana, transparència i rendició de comptes i dret d'accés a la informació pública. De la mà de les tecnologies de la informació i de la comunicació, els governs i les administracions duen a terme iniciatives d'apertura de dades, moviment conegut com Open Data (Dades Obertes). Les plataformes digitals, on aquestes entitats posen a disposició de la societat civil les dades, són conegudes com portals de dades obertes. Es tracta de fonts d'informació on els conjunts de dades són potencialment reutilitzables, amb qualsevol fi i sense cap mena de restricció, únicament de referència d'autoria de les dades. La comunitat científica, personal altament qualificat dins de la societat, poden arribar a ser reutilizadors potencials d'aquestes fonts d'informació. El producte derivat es tradueix en producció científica: articles, usos de dades obertes en projectes d'investigació, comunicacions i docència. Aquest estudi aborda, per una banda, el coneixement que tenen els investigadors i investigadores sobre les dades obertes; per una altra, l'ús i la reutilització de les dades obertes per a generar coneixement científic. Per a dur a terme l'estudi s'ha desenvolupat una metodologia quantitativa. S'ha elaborat una enquesta, distribuïda en un bloc inicial de context, amb 6 preguntes i 6 blocs de caràcter tècnic amb 24 preguntes, és a dir, un qüestionari amb 30 preguntes. S'obtenen un total de 783 respostes, procedents de 34 províncies espanyoles. Els investigadors i investigadores procedeixen de 47 universitats espanyoles i 21 centres de recerca, i existeix representació de 19 àrees de recerca de l'Agència Estatal de Recerca. Amb les dades obtingudes a través d'aquesta metodologia quantitativa es processen, es normalitzen i es duu a terme una anàlisi. A més, amb les dades, es desenvolupa una plataforma per a visualitzar els resultats de l'enquesta. / [EN] Open Government is a mode of public policy that is based on the pillars of collaboration and citizen participation, transparency and accountability, and right of access to public information. Hand in hand with information and communication technologies, governments and administrations carry out initiatives to open data, a movement known as Open Data. The digital platforms, where these entities make the data available to civil society, are known as Open data portals. These are sources of information where the data sets are potentially reusable, for any purpose and without any type of restriction, only for reference of authorship of the data. The scientific community, highly qualified personnel within society, can become potential re-users of these information sources. The by-product translates into scientific production: articles, uses of open data in research projects, communications and teaching. This study addresses, on the one hand, the knowledge that researchers have about open data; on the other, the use and reuse of open data to generate scientific knowledge. In order to carry out the study, a quantitative methodology has been developed. A survey has been prepared, distributed in an initial block of context with 6 questions and 6 technical blocks with 24 questions, that is, a questionnaire with 30 questions. A total of 783 responses were obtained, from 34 Spanish provinces. The researchers come from 47 Spanish universities and 21 research centers, and 19 research areas of the State Research Agency are represented. The data obtained through this quantitative methodology are processed, normalized and analyzed. In addition, a platform is developed with the data, in order to visualize the results of the survey. / Vidal Cabo, C. (2022). Conocimiento, uso y reutilización de los datos abiertos en la ciencia española [Tesis doctoral]. Universitat Politècnica de València. https://doi.org/10.4995/Thesis/10251/184664
177

Conjuntos de datos estadísticos públicos en España

Fernández López, Antonio Lázaro 03 November 2017 (has links)
This study falls within the field of information management and public sector documentation (PSI), specifically public statistical information. The study aims to analyze, evaluate and diagnose the ways of dissemination of statistical information in Spain and its degree of compliance with the main internationally recommended principles for public sector information. The work is structured in several sections: ¿ Identification of statistical information and documentation. ¿ Analysis of its organization and its stakeholders, and their importance within the set of administrations. ¿ Determination of the statistical activity, its characteristics and its quantification, with special attention to its sources of information and to the diffusion phase of its results. It pays attention to the degree of transparency of this activity in the different administrations. ¿ Identification of the media, types of documents and the current information resources and their characteristics. ¿ The degree of compliance with the principles of access and inventory in different administrations. The method used starts from a content analysis about the bases of the statistical organization and its legal foundations. This analysis determines the conceptual framework, the information sources of the activity and its informative characteristics. In a second phase the documentary typologies and characteristics of the products regarding the diffusion of results of this activity are analyzed, as well as their specific forms of dissemination. In a third phase, an empirical work is carried out in which a heuristic analysis of the web sites for each of the 18 statistical systems identified determines their degree of compliance with the access and inventory principles. This analysis is done in three hierarchical structured phases within each system, the analysis of the web site of the governments, the analysis of the web sites of the central statistical agency and the sector agency of each local administration. In a last phase of this analysis, the open data portals and a sample of the municipal portals are incorporated. In the last section, the characteristics of the forms of diffusion of some of the main document types and information products are analyzed and evaluated in a comparative way. Also new trends in forms of diffusion are identified. The work shows that there is a general difficulty in identifying this information. Dispersing and difficult access to this information is due to the lack of treatment of access and inventory principles. The govern should manage this issue in the public sector information. The proposals for solutions to the issues raised that are directly related to document management and their orientation to the reuse and use of linked data are offered. / Este estudio se enmarca dentro del campo de la gestión de la información y la documentación del sector público (ISP), específicamente de la información estadística pública. El estudio tiene como objeto analizar, evaluar y diagnosticar las formas de la difusión de la información estadística en España y su grado de cumplimiento de los principales principios recomendados internacionalmente para a información del sector público. El trabajo se estructura en diversos apartados: ¿ Identificación de la información y documentación estadística. ¿ Análisis de su organización y de sus actores y su importancia dentro del conjunto de las administraciones. ¿ Determinación de la actividad estadística, sus características y su cuantificación, con especial atención a sus fuentes de información y a la fase de difusión de sus resultados. El grado de transparencia de esta actividad en las diferentes administraciones. ¿ Identificación de los medios de difusión, las tipologías documentales y los recursos de información actuales y sus características. ¿ El grado de cumplimiento de los principios de acceso e inventario de los productos de difusión de los resultados en las distintas administraciones El método empleado parte de una fase de análisis de contenido acerca de las bases de la organización estadística y sus fundamentos legales. Este análisis determina el marco conceptual, las fuentes de información de la actividad y sus características informativas. Continúa una segunda fase en la que por este mismo método se analizan las tipologías documentales y características de los productos de difusión de resultados de esta actividad y sus formas de difusión. En una tercera fase se realiza un trabajo de campo en el que se determina por medio de un análisis heurístico de las sedes web para cada uno de los 18 sistemas estadísticos identificados su grado de cumplimiento de los principios de acceso e inventario. Este análisis se realiza en tres fases estructuradas jerárquicamente dentro de cada sistema, el análisis de la sede web de los gobiernos, el análisis de las sedes web de los órganos estadísticos centrales y el de los organismos sectoriales de cada administración. En una última fase de este análisis se incorporan los portales de datos abiertos y una muestra de los portales municipales. En el último apartado se analizan y evalúan de forma comparativa las características de las formas de difusión de algunos de los principales tipos documentales y productos de información. Y se identifican las nuevas tendencias en las formas de difusión. El trabajo demuestra que con carácter general existe dificultad en la identificación, la dispersión y la dificultad en el acceso a esta información por falta de tratamiento de los principios de acceso e inventario que deben regir la gestión de la información del sector público. Y se ofrecen propuestas de solución a las cuestiones planteadas que están directamente relacionadas con la gestión documental y su orientación a la reutilización y el uso de datos enlazados. / Aquest estudi s'emmarca dins el camp de la gestió de la informació i la documentació del sector públic (ISP), específicament de la informació estadística pública. L'estudi té com a objecte analitzar, avaluar i diagnosticar les formes de difusió de la informació estadística a Espanya i el seu grau de compliment dels principals principis recomanats internacionalment per a informació del sector públic. El treball s'estructura en diversos apartats: ¿ Identificació de la informació i documentació estadística. ¿ Anàlisi de la seva organització i dels seus actors i la seva importància dins del conjunt de les administracions. ¿ Determinació de l'activitat estadística, les característiques i la quantificació, amb especial atenció a les seves fonts d'informació i a la fase de difusió dels seus resultats; presta atenció al grau de transparència d'aquesta activitat en les diferents administracions. ¿ Identificació dels mitjans de difusió, les tipologies documentals i els recursos d'informació actuals i les seves característiques. ¿ Grau de compliment dels principis d'accés i inventari dels productes de difusió dels resultats en les diferents administracions. El mètode emprat part d'una fase d'anàlisi de contingut sobre les bases de l'organització estadística i els seus fonaments legals. Aquesta anàlisi determina el marc conceptual, que son les fonts d'informació de l'activitat i les seves característiques informatives. Continua una segona fase en la què, per aquest mateix mètode, s'analitzen les tipologies documentals i característiques dels productes de difusió de resultats d'aquesta activitat i les seves formes de difusió. En una tercera fase es realitza un treball de camp en què es determina per mitjà d'una anàlisi heurístic de les seus web per a cada un dels 18 sistemes estadístics identificats seu grau de compliment dels principis d'accés i inventari. Aquesta anàlisi es realitza en tres fases estructurades jeràrquicament dins de cada sistema: l'anàlisi de la seu web dels governs, l'anàlisi de les seus web dels òrgans estadístics centrals i el dels organismes sectorials de cada administració. En una última fase d'aquesta anàlisi s'incorporen els portals de dades obertes i una mostra dels portals municipals. En l'últim apartat s'analitzen i s'avaluen de forma comparativa les característiques de les formes de difusió d'alguns dels principals tipus documentals i productes d'informació. I s'identifiquen les noves tendències en les formes de difusió. El treball demostra que amb caràcter general existeix dificultat en la identificació, per la dispersió i difícil accés a aquesta informació per una manca de tractament dels principis d'accés i inventari que han de regir la gestió de la informació del sector públic. S'ofereixen propostes de solució a les qüestions plantejades que estan directament relacionades amb la gestió documental i la seva orientació a la reutilització i l'ús de dades enllaçats. / Fernández López, AL. (2017). Conjuntos de datos estadísticos públicos en España [Tesis doctoral]. Universitat Politècnica de València. https://doi.org/10.4995/Thesis/10251/90398
178

Wege und Mehrwerte der Vernetzung von Kulturdaten: Erfahrungen aus der Arbeit der SLUB Dresden

Munke, Martin 16 September 2024 (has links)
No description available.
179

Entwicklung einer generischen und benutzerfreundlichen Applikation zur Standortanalyse und -planung unter Berücksichtigung der Bevölkerungsverteilung in Deutschland

Garte, Lukas 21 May 2024 (has links)
Bei der Bewältigung der Folgen des demographischen Wandels stehen viele Länder vor der Herausforderung, staatliche Dienstleistungen weiterhin flächendeckend anzubieten. Hierbei geht es darum, Dienstleistungen des öffentlichen Sektors wie Schulen, Krankenhäuser, Feuerwachen, etc. möglichst gleichwertig und kosteneffizient bereitzustellen, wenn die Standorte optimal gewählt sind. Bei dieser Optimierung sind Bevölkerungsdaten eine wichtige Eingangsgröße. Das Ziel dieser Arbeit bestand darin, die sehr spezifische und aufgrund ihrer multifaktoriellen Problemstellung hochkomplexe Thematik der Standortanalyse und -planung zu generalisieren und eine benutzerfreundliche Desktop-Applikation auf Basis der von Esri Inc. bereitgestellten ArcGIS-Technologie (Location-Allocation-Analyse, Einzugsgebiet-Analyse etc.) zu entwickeln. Dabei wurde sich auf die Location-Allocation-Funktionalität fokussiert. Die beiden Hauptkomponenten dieser sind Einrichtungen und Bedarfsstellen bzw. -punkte. Einrichtungen können sowohl bestehende als auch potenzielle Standorte darstellen. Bedarfsstellen repräsentieren die Anzahl der Bürger oder Verbraucher in einem bestimmten Gebiet. Um dieses Ziel zu erreichen, wurde ein generisches Modell für Standortanalysen und -planungen entwickelt und bereitgestellt. Dies geschah nach einer Exploration von Ausgangsdatenquellen zu Einrichtungen öffentlicher Dienstleistungen und Bedarfsstellen zur Bevölkerungsverteilung in Deutschland. Die benutzerfreundliche Applikation wurde gemäß der Methodik des Software Engineerings entwickelt. Hierbei wurden eine Anforderungsanalyse und Entscheidungen des Entwurfsprozesses, konkret, berücksichtigt. Die Implementierung wurde als Add-in namens „LA-Application“ in ArcGIS Pro integriert. Ein Add-in ist eine Erweiterung des Desktop-GIS auf Basis des ArcGIS Pro SDK for .NET. Abschließend wurde ein Testdatenbestand für die Komponenten „Einrichtungen“ und „Bedarfsstellen“ erstellt, um Lösungen für die Analyse und Planung von Standorten aufzuzeigen.:Inhaltsverzeichnis Abkürzungsverzeichnis Vorwort 1. Einleitung 2. Theorie und verwandte Arbeiten 3. Exploration von Ausgangsdatenquellen 4. Bereitstellung eines generischen Modells für Standortanalysen und -planungen 5. Implementierung einer benutzerfreundlichen Applikation 6. Anwendung der Applikation und Auswertungen 7. Zusammenfassung und Ausblick Glossar Literaturverzeichnis Abbildungsverzeichnis Tabellenverzeichnis A. Geschäftsverteilungsplan der Landeshauptstadt Dresden B. Anforderungsspezifikation C. Erklärungen zur Testdatenbestandskomponente 'Network Dataset' D. Digitale Anlagen Erklärung über die eigenständige Erstellung der Arbeit
180

Automating Geospatial RDF Dataset Integration and Enrichment

Sherif, Mohamed Ahmed Mohamed 12 May 2016 (has links)
Over the last years, the Linked Open Data (LOD) has evolved from a mere 12 to more than 10,000 knowledge bases. These knowledge bases come from diverse domains including (but not limited to) publications, life sciences, social networking, government, media, linguistics. Moreover, the LOD cloud also contains a large number of crossdomain knowledge bases such as DBpedia and Yago2. These knowledge bases are commonly managed in a decentralized fashion and contain partly verlapping information. This architectural choice has led to knowledge pertaining to the same domain being published by independent entities in the LOD cloud. For example, information on drugs can be found in Diseasome as well as DBpedia and Drugbank. Furthermore, certain knowledge bases such as DBLP have been published by several bodies, which in turn has lead to duplicated content in the LOD . In addition, large amounts of geo-spatial information have been made available with the growth of heterogeneous Web of Data. The concurrent publication of knowledge bases containing related information promises to become a phenomenon of increasing importance with the growth of the number of independent data providers. Enabling the joint use of the knowledge bases published by these providers for tasks such as federated queries, cross-ontology question answering and data integration is most commonly tackled by creating links between the resources described within these knowledge bases. Within this thesis, we spur the transition from isolated knowledge bases to enriched Linked Data sets where information can be easily integrated and processed. To achieve this goal, we provide concepts, approaches and use cases that facilitate the integration and enrichment of information with other data types that are already present on the Linked Data Web with a focus on geo-spatial data. The first challenge that motivates our work is the lack of measures that use the geographic data for linking geo-spatial knowledge bases. This is partly due to the geo-spatial resources being described by the means of vector geometry. In particular, discrepancies in granularity and error measurements across knowledge bases render the selection of appropriate distance measures for geo-spatial resources difficult. We address this challenge by evaluating existing literature for point set measures that can be used to measure the similarity of vector geometries. Then, we present and evaluate the ten measures that we derived from the literature on samples of three real knowledge bases. The second challenge we address in this thesis is the lack of automatic Link Discovery (LD) approaches capable of dealing with geospatial knowledge bases with missing and erroneous data. To this end, we present Colibri, an unsupervised approach that allows discovering links between knowledge bases while improving the quality of the instance data in these knowledge bases. A Colibri iteration begins by generating links between knowledge bases. Then, the approach makes use of these links to detect resources with probably erroneous or missing information. This erroneous or missing information detected by the approach is finally corrected or added. The third challenge we address is the lack of scalable LD approaches for tackling big geo-spatial knowledge bases. Thus, we present Deterministic Particle-Swarm Optimization (DPSO), a novel load balancing technique for LD on parallel hardware based on particle-swarm optimization. We combine this approach with the Orchid algorithm for geo-spatial linking and evaluate it on real and artificial data sets. The lack of approaches for automatic updating of links of an evolving knowledge base is our fourth challenge. This challenge is addressed in this thesis by the Wombat algorithm. Wombat is a novel approach for the discovery of links between knowledge bases that relies exclusively on positive examples. Wombat is based on generalisation via an upward refinement operator to traverse the space of Link Specifications (LS). We study the theoretical characteristics of Wombat and evaluate it on different benchmark data sets. The last challenge addressed herein is the lack of automatic approaches for geo-spatial knowledge base enrichment. Thus, we propose Deer, a supervised learning approach based on a refinement operator for enriching Resource Description Framework (RDF) data sets. We show how we can use exemplary descriptions of enriched resources to generate accurate enrichment pipelines. We evaluate our approach against manually defined enrichment pipelines and show that our approach can learn accurate pipelines even when provided with a small number of training examples. Each of the proposed approaches is implemented and evaluated against state-of-the-art approaches on real and/or artificial data sets. Moreover, all approaches are peer-reviewed and published in a conference or a journal paper. Throughout this thesis, we detail the ideas, implementation and the evaluation of each of the approaches. Moreover, we discuss each approach and present lessons learned. Finally, we conclude this thesis by presenting a set of possible future extensions and use cases for each of the proposed approaches.

Page generated in 0.0675 seconds