191 |
Сегментация сигналов электрокардиограмм в задаче неконтролируемого построения словаря волн : магистерская диссертация / Segmentation of electrocardiogram signals in the problem of unsupervised construction of a wave dictionaryЛебедев, А. П., Lebedev, A. P. January 2023 (has links)
В данной магистерской работе мы исследуем возможности построения словаря волн биомедицинских сигналов электрокардиограммы, который в дальнейшем позволит применять методы NLP для обработки временных рядов биомедицинских сигналов. В частности, мы сосредоточимся на анализе структуры пиков и интервалов электрокардиограммы здоровых и больных аритмией и другими заболеваниями людей, средствами языка python и автоматизации этого процесса для извлечения значимой информации из биомедицинских временных рядов ЭКГ. Наша конечная цель – улучшение точности и эффективности обработки и анализа биомедицинских сигналов, что имеет важное значение как для клинической диагностики, так и для научных исследований. Решение этой задачи имеет большое практическое значение для различных областей, таких как медицина, биология и фармакология, где обработка и анализ временных рядов играют важную роль. / In this master's thesis, we are exploring the possibility of constructing a dictionary of waves of biomedical electrocardiogram signals, which in the future will allow the use of NLP methods for processing time series of biomedical signals. In particular, we will focus on analyzing the structure of peaks and intervals of the electrocardiogram of healthy people and patients with arrhythmia and other diseases, using the Python language and automating this process to extract meaningful information from biomedical ECG time series. Our ultimate goal is to improve the accuracy and efficiency of biomedical signal processing and analysis, which is important for both clinical diagnostics and scientific research. The solution to this problem is of great practical importance for various fields, such as medicine, biology and pharmacology, where processing and analysis of time series play an important role.
|
192 |
Python in der Regelungs- und AutomatisierungstechnikKnoll, Carsten 28 May 2024 (has links)
No description available.
|
193 |
Entwicklung eines Systems zur Erfassung und Untersuchung von Certificate Transparency LogsMeesters, Johannes 13 July 2024 (has links)
Angesichts der zentralen Rolle der Root-Zertifizierungsstellen als Vertrauensanker der Web PKI und der in der Vergangenheit aufgetretenen Vorfälle mit unberechtigt oder inkorrekt ausgestellten Zertifikaten, ist die Transparenz und Verantwortlichkeit dieser Root CAs von großer Bedeutung. Seit der Einführung von Certificate Transparency Logs werden alle von Certificate Authorities ausgestellten Zertifikate in diese öffentlichen Logs eingetragen.
Die Arbeit stellt die Problematik der eingeschränkten Zugänglichkeit dieser Daten für die Wissenschaft dar und entwickelt ein Werkzeug, dass eine unabhängige Aufzeichnung und Auswertung von Certificate Transparency Logs ermöglicht. Das entwickelte System nutzt eine containerbasierte Architektur und Elasticsearch zur effizienten Speicherung und Analyse der Daten. Es bewältigt ein hohes Datenaufkommen von durchschnittlich 25 Millionen Log-Einträgen pro Tag und ermöglicht eine anpassbare Datenverarbeitung und -auswertung. Die Vorverarbeitung und Indexierung sowie die Auswertung der Daten erfolgte mit Python, was eine flexible Anpassung des Systems an unterschiedliche Forschungsfragen erlaubt.
Über einen Zeitraum von 42 Tagen wurden insgesamt 645 Millionen CT Log-Einträge aufgezeichnet und analysiert. Aus den Auswertungen geht hervor, wie verschiedene CAs und deren Root-Zertifikate genutzt werden und wie stark die unterschiedlichen CT Logs von CAs verwendet werden.
Die Arbeit identifiziert jedoch auch Herausforderungen, wie den hohen Speicherbedarf und notwendige Optimierungen in der Datenindexierung.:1 Einleitung
1.1 Problemstellung
1.2 Zielstellung
2 Grundlagen
2.1 X509-Zertifikate
2.1.1 Felder
2.1.2 Erweiterungen
2.2 Certificate Transparency
2.2.1 Certificate Transparency Log
2.2.2 Überprüfung durch User Agents
2.2.3 Überprüfung durch Monitors
2.2.4 Eintragung durch Certificate Authorities
3 Konzeptionierung
3.1 Abfrage der CT Logs
3.2 Verarbeitung der Zertifikate
3.3 Speicherung & Auswertung der Daten
3.4 Überwachung
3.5 Docker
4 Implementierung
4.1 Plattform
4.2 Überwachung
4.3 certstream-server
4.4 Verarbeitung
4.4.1 Pufferung (stream-to-queue-publisher)
4.4.2 Vorverarbeitung (cert-indexer)
4.5 Elasticsearch
4.5.1 Speicherverbrauch
4.5.2 Field Mappings
5 Auswertung
5.1 Logs & Log-Betreiber
5.2 Certificate Authorites
5.3 Zertifikats-Größe
5.4 Gültigkeitsdauer
6 Schluss
6.1 Fazit
6.2 Ausblick
A Beispiel X509 Leaf-Zertifikat
B Beispiel X509 Root-Zertifikat
C Beispiele Elasticsearch Abfragen
Literatur
Abbildungsverzeichnis
Tabellenverzeichnis / In view of the central role of the root certification authorities as trust anchors of the Web PKI and the incidents that have occurred in the past with unauthorised or incorrectly issued certificates, the transparency and accountability of these root CAs is of great importance. With the introduction of Certificate Transparency Logs, all certificates issued by Certificate Authorities are now entered in public logs.
The work presents the problem of the limited accessibility of this data for science and develops a tool that enables an independent recording and evaluation of Certificate Transparency Logs. The developed system uses a container-based architecture and Elasticsearch to efficiently store and analyse the data. It can handle a high volume of data, averaging 25 million log entries per day, and enables customisable data processing and analysis. Python was used to pre-process, index and analyse the data, allowing the system to be flexibly adapted to different research questions.
A total of 645 million CT log entries were recorded and analysed over a period of 42 days. The analyses show how different CAs and their root certificates are used and how much the different CT logs are used by CAs.
However, the work also identifies challenges, such as the high memory requirements and necessary optimisations in data indexing.:1 Einleitung
1.1 Problemstellung
1.2 Zielstellung
2 Grundlagen
2.1 X509-Zertifikate
2.1.1 Felder
2.1.2 Erweiterungen
2.2 Certificate Transparency
2.2.1 Certificate Transparency Log
2.2.2 Überprüfung durch User Agents
2.2.3 Überprüfung durch Monitors
2.2.4 Eintragung durch Certificate Authorities
3 Konzeptionierung
3.1 Abfrage der CT Logs
3.2 Verarbeitung der Zertifikate
3.3 Speicherung & Auswertung der Daten
3.4 Überwachung
3.5 Docker
4 Implementierung
4.1 Plattform
4.2 Überwachung
4.3 certstream-server
4.4 Verarbeitung
4.4.1 Pufferung (stream-to-queue-publisher)
4.4.2 Vorverarbeitung (cert-indexer)
4.5 Elasticsearch
4.5.1 Speicherverbrauch
4.5.2 Field Mappings
5 Auswertung
5.1 Logs & Log-Betreiber
5.2 Certificate Authorites
5.3 Zertifikats-Größe
5.4 Gültigkeitsdauer
6 Schluss
6.1 Fazit
6.2 Ausblick
A Beispiel X509 Leaf-Zertifikat
B Beispiel X509 Root-Zertifikat
C Beispiele Elasticsearch Abfragen
Literatur
Abbildungsverzeichnis
Tabellenverzeichnis
|
194 |
Automated Differentiation of Chat Application Versions and Categorisation of Changes Based on Forensic Relevance / Automatiserad Differentiering av Chattapplikationsversioner och Kategorisering av ÄndringarBaserade på Forensisk RelevansLjungsten, Ted, Makowski, Adam January 2024 (has links)
This thesis investigates the automation of forensic analysis in identifying and categorising forensically interestingchanges across different versions of chat applications on Android platforms. The focus is primarily on thedifferentiation of Android Package Kit (APK) using reverse-engineering techniques to reconstruct the originalsource code and comparing the source code from two different versions of the APK. Given the rapid evolutionof chat applications and their frequent updates, it is crucial for forensic investigators to understand thesechanges to maintain the integrity of legal investigations.The research introduces a comprehensive framework leveraging the open-source tools, Ghidra and BinDiff, toautomate the decompilation and differential analysis of APK files. This approach not only makes forensicanalysis less complicated but also ensures that investigators can keep pace with the continuous updates in chatapplications.Tests on the system are conducted on various versions of the Signal chat application. These tests aim todemonstrate the proposed tool in capturing significant changes between APK versions, such as alterations inlogging mechanisms, database interactions, and the use of encryption and cypher libraries.The results confirm that the integration of Ghidra and BinDiff provides a solution for automated forensicanalysis, facilitating the identification of changes and categorisation of methods based on their forensicrelevance. The study shows that the tool can pinpoint modifications and structural changes, which are essentialfor forensic investigations.
|
195 |
PySciDON: a Python scientific framework for development of ocean network applicationsVandenberg, Nathan 04 January 2017 (has links)
The Salish Sea is a ecologically important coastal region located on the southwest
part of British Columbia. Optical measurements were taken using a set of hyperspectral
radiometers, the SAS Solar Tracker developed by Satlantic. This sensor is
installed on the Queen of Oak Bay ferry, that runs between Nanaimo and Vancouver,
as part of the Ferry Ocean Colour Observation Systems (FOCOS) project. We developed
a computer program to process the raw sensor data and generate remote sensing
reflectance (Rrs) values. This performs similar functions to Prosoft, Satlantic’s own
software to process the data. However, we added new features such as an additional
preprocessing step to filter the data based on longitude, and new meteorological flag
testing and wind speed calculations. The system was tested using Pearson correlation
to compare our output with the output from Satlantic Prosoft. Testing helped
us identify a few issues, such as adding longitude flags to remove data at the start and
end of the trip where the sensor could produce inaccurate results if aiming at land
instead of water. Another issue was where the SAS Solar Tracker does not update
its pointing angle fast enough when the ferry makes sharp turns and could result in
inaccurate data. / Graduate
|
196 |
Towards a computer model of the historical phonology and morphology of LatinRoberts, Philip J. January 2012 (has links)
Research projects in Optimality Theory tend to take a synchronic view of a particular generalisation, and set their standards for rigour in typological terms (see for example Suzuki 1998 on dissimilation, Crosswhite 2001 on vowel reduction). The goal of this thesis is to use Stratal OT to take a diachronic view of multiple generalisations within the morpho-phonology of one language, namely Latin, with the principal empirical aim of producing an analysis that is demonstrably true to all the attested facts of the generalisations in question. To that end, I have written PyOT, a computer program implementing the OT calculus and a theory of phonological representations, which I use in this work to model the histories of Lachmann’s Law, rhotacism and the phonologically conditioned allomorphy of the -alis/aris- suffix as active generalisations within the phonological component of the grammar. Appendix A gives the results of the computer model applied to a dataset consisting of 185 attested Latin forms, which suffice to illustrate the exact conditions of the generalisations in question. I show that producing a complete analysis of the three generalisations I have chosen to model entails analysis of other generalisations that interact with them, including the treatment of the Indo-European voiced aspirates in Latin (which interacts with rhotacism), and reduplication in forming perfect stems (which interacts with Lachmann’s Law). Constraint rankings sufficient to model these interactions, and consistent with the general conditions of the interacting generalisations have been included in the model. The intention is for this work to illustrate both the utility of formal phonological theory in advancing hypotheses within historical-comparative linguistics, and the potential of PyOT as a tool for producing Optimality-Theoretic models of (eventually) a language’s entire phonology.
|
197 |
Calcul en n-dimensions sur GPUBergeron, Arnaud 04 1900 (has links)
Le code source de la libraire développée accompagne ce dépôt dans l'état où il était à ce moment. Il est possible de trouver une version plus à jour sur github (http://github.com/abergeron). / Le calcul scientifique sur processeurs graphiques (GPU) est en plein essor depuis un certain temps, en particulier dans le domaine de l'apprentissage machine.
Cette thèse présente les efforts pour établir une structure de données de table au multidimensionnel de manière efficace sur GPU.
Nous commençons par faire une revue de ce qui est actuellement similaire dans le domaine et des désavantages d'avoir une multitude d'approches.
Nous nous intéresserons particulièrement aux calculs fait à partir du langage Python.
Nous décrirons des techniques intéressantes telles que la réduction d'ordre et le calcul asynchrone automatique.
Pour terminer nous présenterons l'utilisation du module développé dans le cadre de cette thèse. / Scientific computing on GPU (graphical processing units) is on the rise, specifically in machine learning.
This thesis presents the implementation of an efficient multidimensional array on the GPU.
We will begin by a review of what currently implements similar functionality and the disadvantage of a fragmented approach.
We will focus on packages that have a Python interface.
We will explain techniques to optimize execution such as order reduction and automatic asynchronous computations.
Finally, we will present the functionality of the module developed for this thesis.
|
198 |
Hranice překladu filmového humoru: Monty Python / The humor of Monty Pyton and the Limitations of its translationSmrčková, Tereza January 2012 (has links)
This paper focuses on the translation of audiovisual humour. I have analysed Petr Palouš's translation of Monty Python's Flying Circus. The translator of audiovisual humour has to deal with the same issues and problems as the translator of literature, that is he has to find equivalents to word plays, idioms, register and so on, but at the same time he also has to comply with the restrictions of audiovisual medium. I have identified the most common translation problems when translating humour and audiovisual texts and possible strategies of their translation into the target language, and then analysed how Palouš dealt with these cruces translatorum when translating the Flying Circus.
|
199 |
Programové prostředí pro asimilační metody v radiační ochraně / Software environment for data assimilation in radiation protectionMajer, Peter January 2015 (has links)
In this work we apply data assimilation onto meteorological model WRF for local domain. We use bayesian statistics, namely Sequential Monte Carlo method combined with particle filtering. Only surface wind data are considered. An application written in Python programming language is also part of this work. This application forms interface with WRF, performs data assimilation and provides set of charts as output of data assimilation. In case of stable wind conditions, wind predictions of assimilated WRF are significantly closer to measured data than predictions of non-assimilated WRF. In this kind of conditions, this assimilated model can be used for more accurate short-term local weather predictions. Powered by TCPDF (www.tcpdf.org)
|
200 |
Programovací jazyk Python a účelnosť jeho zaradenia do výučby / Programming language Python and its suitability of his assignment to the tuitionArendáč, Tomáš January 2011 (has links)
This thesis is concerned by programming language Python and its suitability of his assignment to the tuition. The work is divided into three dominant parts. The first part describes programming language Python, its elementary characteristics and features. The purpose is to introduce its properties to the reader so that he could estimate if there is point in the deeper concern. There are elements of object-oriented programming in description, too. The second part analyses programming language Python on the basis of ten criteria which are defined considering applicability of the language in preliminary courses of programming. The purpose is to review if the Python is appropriate in these courses. The third part considers the possibilities of the Python tuition at the University of Economics in Prague. The main contribution of the thesis is to give opinion on the fundamental description of the language, to define framework and to pass judgment on the potential chance of use in preliminary courses of programming.
|
Page generated in 0.0296 seconds