Spelling suggestions: "subject:"datavetenskap"" "subject:"datvetenskap""
821 |
Inferring Dataset Relations using Knowledge Graph MetadataEdström, August, Isaksson, Johan January 2020 (has links)
The web site dataportalen.se aims to increase the availability of Swedish open datasets. This is achieved by collecting metadata about the open datasets provided by Swedish organizations. At the time of writing, metadata from more than two thousand datasets reside in the portal, and this number is set to increase. As the number of datasets increases, browsing for relevant information becomes increasingly difficult and time-consuming. The web site supports searching using text and then filtering the results by theme, organization, file format och license. We believe that there exists potential to connect the datasets, thus making it easier to find a dataset of interest. The idea is to find common denominators in the metadata of the datasets. Furthermore, as no user data is available, the datasets had to be connected based solely on the metadata. The datasets are annotated with metadata, such as title,description, keywords, themes. By comparing metadata from different datasets, a measure of similarity could be computed. This measure can then be used to find the most relevant datasets for a specific dataset.The achieved results suggests that it is indeed possible to find similar datasets by only analyzing the metadata. By exploring various methods, we found it to be the case that text data holds useful information that can be used to find relations between datasets. Using a related workas a benchmark, we found that our results are as good if not better. Furthermore, the approach taken in this project is quite general, and should theoretically be applicable in other scenarios where textual data is available. / Webbplatsen dataportalen.se syftar till att öka tillgängligheten av svensk öppna data. Detta görs genom att samla in metadata om de öppna datamängderna som tillhandahålls av svenska organisationer. I skrivande stund finns metadata från mer än två tusendatamängder i portalen, och detta antal kommer att öka. När antalet datamängder ökar blir genomsökandet av relevant information allt svårare och mer tidskrävande. För närvarande är det möjligt att genomsöka datamängderna med hjälp av sökning med text och sedan filtrering med tema, organisation, filformat eller licens. Vi tror att det finns mer potential att koppla samman datamängderna, vilket skulle göra det enklare att hitta datamängder av intresse. Idén är att hitta gemensamma nämnare i metadatat för datamängderna. Eftersom det inte finns någon användardata kommervi att undersöka i vilken utsträckning denna idé kan realiseras. Datamängderna kommenteras med metadata som titel, beskrivning, nyckelord, och tema. Genom att jämföra metadata från olika datamängder kan ett mått på likhet beräknas. Detta mått kan sedan användas för att hitta de mest relevanta datamängderna för en specifik datamängd. Resultaten av analysen av metadata är att liknande datamängder kan hittas. Genom att utforska olika metoder fann vi att textdata inneåaller användbar information som kan användas för att hitta relationer mellan datamängder. Genom att använda ett relaterat arbete som riktmärke fann vi att våra resultat är lika bra, om inte bättre. Resultaten visar att relaterade datamängder kan hittas med bara textdata, och vi anser att den identifierade metoden är tillräckligt generell för att ha potential att kunna användas i liknande problem där textdata är tillgänglig.
|
822 |
Automated Model Generation using Graphwalker Based On Given-When-Then SpecificationsKorhonen, Joakim January 2020 (has links)
Software testing is often a laborious and costly process, as testers need extensive domain-specific knowledge and engineering experience to manually create test cases for diverse test scenarios. These scenarios in many industrial projects are represented in requirement specification documents. Since the creation of test cases from these requirements is manual and is error-prone, researchers have proposed methods to automate the creation of tests and execution of tests. One of the most popular approaches is called model-based testing. Model-based testing uses models to manually or automatically create tests based on existing models. Since most of the effort in model-based testing lies in the creation of the model, this thesis aims at improving a model-based testing tool. This improvement is for generating a model from Natural language as this is what requirements usually are written in. Given-When-Then is a test-case writing template used to specify a system's behavior. To implement the natural language processing into a model-based testing tool, an extension for Graphwalker was created. Graphwalker is a popular open-source model-based testing tool, which can create, edit, and test the models created. The extension is using requirements as input written in natural languages and then creates a model based on the requirements provided. Graphwalker's models are based on finite state machines that have elements such as vertices and edges. The model also can change its state, change values of variables, and block access to certain elements. Graphwalker can however not generate models from natural language requirements. This thesis shows how one can transform natural language requirements into models. The extension is implemented to use requirements through both manual input and via a JSON file and it is processing the text and tags each word. These tags will then be used to interpret the sentence meaning and will either create a transition, change a value, or block access to a selected element. The results of this thesis show that this extension is an applicable method to automatically generate models for the GraphWalker tool. This extension can be used and improved by both researchers and practitioners.
|
823 |
Evaluating remote and local web rendering of real-time interactive 2D graphics using BlazorAndersson Tholin, Alexander January 2020 (has links)
With the growing popularity of the web, companies are starting to extend current development to reflect this. When extending desktop applications to the web, it can be difficult to choose what techniques and technologies to use when solving a problem, as current solutions might not be directly applicable. Rendering high-performance interactive 2D graphics on the web can be achieved in multiple ways. The rise in open standards such as the Canvas API allows the client to render natively in the browser, provided they can receive the full object state. There are some cases where this is simply not possible, where the object state is too large, or the client is lacking sufficient hardware. A possible solution is migrating the rendering of the graphic from the client to the server. However, remote rendering comes with new sets of issues as it often lacks high interaction capabilities, and would theoretically require more resources with multiple connections. This thesis will evaluate the performance differences and individual capabilities of remote and local rendering in terms of scalability and Quality of Experience using ASP.NET Core Blazor. The evaluation is done through the implementation of the four different solutions for the scenario. These implementations are based on Canvas and SVG using remote and local rendering. Different configurations of the performed tests, such as how much data should be rendered and how many clients are connected, were used to see how they affect response time and interaction latency. The results show that remote rendering performed better in all scalability tests, with remote SVG being the recommended approach. Due to implementation issues and lack of a proper testing environment, the number of concurrent clients was downsized. This caused problems when analyzing the results, and drawing concrete conclusions were difficult. In tests with increasing image size, the client solution suffered memory exceptions, preventing the local versions to be tested further. When testing interaction capabilities by measuring interaction latency, the SVG technology significantly outperformed Canvas, since SVG does not require a full re-render of the elements.
|
824 |
Privacy-preserving proof-of-location using homomorphic encryptionLee, Carmen January 2020 (has links)
Location-based software services require knowledge about a user's geographic data. Sharing these data risks compromising the user's privacy, exposes the user to targeted marketing, and enables potentially undesired behavioural profiling. Today, there exist several privacy-preserving proof-of-location solutions. However, these solutions often rely on a trusted third party, which reduces a user's control of their own data, or feature novel encryption schemes that may contain yet undiscovered security vulnerabilities. This thesis adopts a generic homomorphic encryption scheme and presents a way of generating location proofs without a user having to reveal their location.
|
825 |
A tool for extracting and visualizing source code metric trendsSjövall, Albin January 2022 (has links)
The quality of code can be measured using source code metrics. Looking at the trends of these metrics over a period of time can show a potential decrease in code quality. Low quality code will lead to technical debt resulting in higher cost of maintenance for a software project. In the context of a research project related to this topic, a tool was created to analyse source code metric trends for Git repositories. This tool had two problems, it was slow and unreliable. Both of these issues were caused by the tool requiring target repositories to be built during the metrics collection. To solve these issues the tool will be modified to collect metrics straight from the source code files, thus not requiring a build process, leading to increased performance and higher reliability.The result was a six times increase in performance as well as the tool functioning on a wider range of repositories.
|
826 |
How to incorporate DL models into a microprocessor for EDGE applicationsTóth, Zsombor January 2022 (has links)
No description available.
|
827 |
Automate memory forensics InvestigationMohamed, Azeem, Saad, Tirmizi January 2022 (has links)
The growth of digital technology spawns both positive and negative effects. Cybercrimes rise with the advancement of computer technology, necessitating a digital forensics investigation of the evolving digital world to assist in solving and tracing criminals' digital activity. We also know that every process executed in a digital system must run in memory at some point. Therefore, volatile memory forensics is at the forefront of forensic investigation and incident response. The memory analysis technique retrieves artifacts to analyze inappropriate behaviors. A bit-to-bit memory image contains significant artifacts that provide the analyst with relevant clues, such as system processes, recent activities, opened network ports, and connections. However, all this information is lost as soon as the system is shut down, which flushed the volatile memory. It also takes a long time to gather, analyze, and present data from various devices for every crime because the number of devices and the amount of data are constantly growing and adding to the backlog of devices to examine and analyze. Therefore, to eliminate human error and backlogs, we develop multiple machine learning classification models and identify the best performing model to automate the memory forensic process.
|
828 |
Latency-aware Resource Management at the EdgeToczé, Klervie January 2020 (has links)
The increasing diversity of connected devices leads to new application domains being envisioned. Some of these need ultra low latency or have privacy requirements that cannot be satisfied by the current cloud. By bringing resources closer to the end user, the recent edge computing paradigm aims to enable such applications. One critical aspect to ensure the successful deployment of the edge computing paradigm is efficient resource management. Indeed, obtaining the needed resources is crucial for the applications using the edge, but the resource picture of this paradigm is complex. First, as opposed to the nearly infinite resources provided by the cloud, the edge devices have finite resources. Moreover, different resource types are required depending on the applications and the devices supplying those resources are very heterogeneous. This thesis studies several challenges towards enabling efficient resource management for edge computing. The thesis begins by a review of the state-of-the-art research focusing on resource management in the edge computing context. A taxonomy is proposed for providing an overview of the current research and identify areas in need of further work. One of the identified challenges is studying the resource supply organization in the case where a mix of mobile and stationary devices is used to provide the edge resources. The ORCH framework is proposed as a means to orchestrate this edge device mix. The evaluation performed in a simulator shows that this combination of devices enables higher quality of service for latency-critical tasks. Another area is understanding the resource demand side. The thesis presents a study of the workload of a killer application for edge computing: mixed reality. The MR-Leo prototype is designed and used as a vehicle to understand the end-to-end latency, the throughput, and the characteristics of the workload for this type of application. A method for modeling the workload of an application is devised and applied to MR-Leo in order to obtain a synthetic workload exhibiting the same characteristics, which can be used in further studies.
|
829 |
Automatic Management of Related Data Distribution Service TopicsKazai, Gabriel January 2022 (has links)
A distributed system is a system that utilizes a network to communicate while usually being broken down into multiple software programs. Said programs generally have separate tasks, and a paradigm used in the manner above is publish-subscribe. However, to enable the publish-subscribe paradigm, DDS (Data Distribution Service) is utilized as a network middleware for said type of communication. This thesis proposes a suitable way to specify data variants to allow for code generation to automatically create topics of information and publish data based on said variants. Thus, making it easier for distributed systems developers to handle information variants. Succeeding in this task could provide developers with a tool that can be used to create data variants more cheaply and faster than doing it manually. I designed variant types to achieve the goal of the thesis, along with a specification describing said types. Lastly, along with the previously mentioned, a manual implementation of a working DDS system utilizing said variant types proved the functionality and evaluated the completed work. As the goal sought out, the thesis's result is a working method for a user to specify variants that would allow for automatic code generation, thus providing developers with a functional specification to utilize within the field.
|
830 |
En tillgängligare mobilwebb för personer med finmotoriska svårigheterWesterlund, Frida, Eklund Dafteke, Tim January 2022 (has links)
Idag behöver varken offentliga eller privata aktörer tillämpa tillgänglighetsanpassningar på sina hemsidor när de besöks från smartmobilens webbläsare. Det finns inga krav på att vara inkluderande för människor med finmotoriska svårigheter eller rörelsenedsättningar som använder sin smartmobil för att nyttja internet. Majoriteten av alla människor i Sverige har tillgång till en smartmobil och trots bristande tillgänglighet är människor med funktionsnedsättningar oftare aktiva på sociala medier än människor utan funktionsnedsättningar. För att förhindra ett digitalt utanförskap behöver vi tillgänglighetsanpassa den digitala världen så att alla kan ta del av den. Därför har vi i vår studie utforskat möjligheten att främja digital tillgänglighet på en hemsida när den besöks från smartmobilens webbläsare genom att ersätta pekskärmen och externa hjälpmedel med en digital artefakt. Vi har tagit fram frågeställningarna \say{Hur kan mobilens sensorer, som är åtkomliga genom webbläsaren, underlätta interaktionen med hemsidor för människor som har finmotoriska svårigheter?} och \say{I vilken utsträckning kan mobilens sensorer användas för att imitera funktionaliteten hos en extern artefakt, som en kontaktstyrningsenhet?}. För att besvara dem har vi genomfört en nätnografisk studie, utvecklat en prototyp baserat på den nätnografiska studien och genom en användarstudie utvärderat prototypen. Vi kunde konstatera att smartmobilens sensorer är tillräckliga för att navigera en hemsida utan pekskärmen, men vi kunde inte säkerställa om vår artefakt underlättar för personer med rörelsenedsättningar. Vad som framgår av vår studie är att mer behöver göras för att tillgängliggöra hemsidor när de besöks från smartmobilens webbläsare och att smartmobilens sensorer kan användas som ett komplement till nuvarande standarder för digital tillgänglighet.
|
Page generated in 0.0768 seconds