• Refine Query
  • Source
  • Publication year
  • to
  • Language
  • 160
  • 14
  • 14
  • 13
  • 6
  • 1
  • 1
  • 1
  • Tagged with
  • 249
  • 249
  • 54
  • 45
  • 45
  • 42
  • 39
  • 35
  • 29
  • 28
  • 27
  • 26
  • 25
  • 24
  • 23
  • About
  • The Global ETD Search service is a free service for researchers to find electronic theses and dissertations. This service is provided by the Networked Digital Library of Theses and Dissertations.
    Our metadata is collected from universities around the world. If you manage a university/consortium/country archive and want to be added, details can be found on the NDLTD website.
201

MSE-based Linear Transceiver Designs for Multiuser MIMO Wireless Communications

Tenenbaum, Adam 11 January 2012 (has links)
This dissertation designs linear transceivers for the multiuser downlink in multiple-input multiple-output (MIMO) systems. The designs rely on an uplink/downlink duality for the mean squared error (MSE) of each individual data stream. We first consider the design of transceivers assuming channel state information (CSI) at the transmitter. We consider minimization of the sum-MSE over all users subject to a sum power constraint on each transmission. Using MSE duality, we solve a computationally simpler convex problem in a virtual uplink. The transformation back to the downlink is simplified by our demonstrating the equality of the optimal power allocations in the uplink and downlink. Our second set of designs maximize the sum throughput for all users. We establish a series of relationships linking MSE to the signal-to-interference-plus-noise ratios of individual data streams and the information theoretic channel capacity under linear minimum MSE decoding. We show that minimizing the product of MSE matrix determinants is equivalent to sum-rate maximization, but we demonstrate that this problem does not admit a computationally efficient solution. We simplify the problem by minimizing the product of mean squared errors (PMSE) and propose an iterative algorithm based on alternating optimization with near-optimal performance. The remainder of the thesis considers the more practical case of imperfections in CSI. First, we consider the impact of delay and limited-rate feedback. We propose a system which employs Kalman prediction to mitigate delay; feedback rate is limited by employing adaptive delta modulation. Next, we consider the robust design of the sum-MSE and PMSE minimizing precoders with delay-free but imperfect estimates of the CSI. We extend the MSE duality to the case of imperfect CSI, and consider a new optimization problem which jointly optimizes the energy allocations for training and data stages along with the sum-MSE/PMSE minimizing transceivers. We prove the separability of these two problems when all users have equal estimation error variances, and propose several techniques to address the more challenging case of unequal estimation errors.
202

MSE-based Linear Transceiver Designs for Multiuser MIMO Wireless Communications

Tenenbaum, Adam 11 January 2012 (has links)
This dissertation designs linear transceivers for the multiuser downlink in multiple-input multiple-output (MIMO) systems. The designs rely on an uplink/downlink duality for the mean squared error (MSE) of each individual data stream. We first consider the design of transceivers assuming channel state information (CSI) at the transmitter. We consider minimization of the sum-MSE over all users subject to a sum power constraint on each transmission. Using MSE duality, we solve a computationally simpler convex problem in a virtual uplink. The transformation back to the downlink is simplified by our demonstrating the equality of the optimal power allocations in the uplink and downlink. Our second set of designs maximize the sum throughput for all users. We establish a series of relationships linking MSE to the signal-to-interference-plus-noise ratios of individual data streams and the information theoretic channel capacity under linear minimum MSE decoding. We show that minimizing the product of MSE matrix determinants is equivalent to sum-rate maximization, but we demonstrate that this problem does not admit a computationally efficient solution. We simplify the problem by minimizing the product of mean squared errors (PMSE) and propose an iterative algorithm based on alternating optimization with near-optimal performance. The remainder of the thesis considers the more practical case of imperfections in CSI. First, we consider the impact of delay and limited-rate feedback. We propose a system which employs Kalman prediction to mitigate delay; feedback rate is limited by employing adaptive delta modulation. Next, we consider the robust design of the sum-MSE and PMSE minimizing precoders with delay-free but imperfect estimates of the CSI. We extend the MSE duality to the case of imperfect CSI, and consider a new optimization problem which jointly optimizes the energy allocations for training and data stages along with the sum-MSE/PMSE minimizing transceivers. We prove the separability of these two problems when all users have equal estimation error variances, and propose several techniques to address the more challenging case of unequal estimation errors.
203

Duality and optimality in multiobjective optimization

Bot, Radu Ioan 04 July 2003 (has links) (PDF)
The aim of this work is to make some investigations concerning duality for multiobjective optimization problems. In order to do this we study first the duality for scalar optimization problems by using the conjugacy approach. This allows us to attach three different dual problems to a primal one. We examine the relations between the optimal objective values of the duals and verify, under some appropriate assumptions, the existence of strong duality. Closely related to the strong duality we derive the optimality conditions for each of these three duals. By means of these considerations, we study the duality for two vector optimization problems, namely, a convex multiobjective problem with cone inequality constraints and a special fractional programming problem with linear inequality constraints. To each of these vector problems we associate a scalar primal and study the duality for it. The structure of both scalar duals give us an idea about how to construct a multiobjective dual. The existence of weak and strong duality is also shown. We conclude our investigations by making an analysis over different duality concepts in multiobjective optimization. To a general multiobjective problem with cone inequality constraints we introduce other six different duals for which we prove weak as well as strong duality assertions. Afterwards, we derive some inclusion results for the image sets and, respectively, for the maximal elements sets of the image sets of these problems. Moreover, we show under which conditions they become identical. A general scheme containing the relations between the six multiobjective duals and some other duals mentioned in the literature is derived. / Das Ziel dieser Arbeit ist die Durchführung einiger Untersuchungen bezüglich der Dualität für Mehrzieloptimierungsaufgaben. Zu diesem Zweck wird als erstes mit Hilfe des so genannten konjugierten Verfahrens die Dualität für skalare Optimierungsaufgaben untersucht. Das erlaubt uns zu einer primalen Aufgabe drei unterschiedliche duale Aufgaben zuzuordnen. Wir betrachten die Beziehungen zwischen den optimalen Zielfunktionswerten der drei Dualaufgaben und untersuchen die Existenz der starken Dualität unter naheliegenden Annahmen. Im Zusammenhang mit der starken Dualität leiten wir für jede dieser Dualaufgaben die Optimalitätsbedingungen her. Die obengenannten Ergebnisse werden beim Studium der Dualität für zwei Vektoroptimierungsaufgaben angewandt, und zwar für die konvexe Mehrzieloptimierungsaufgabe mit Kegel-Ungleichungen als Nebenbedingungen und für eine spezielle Quotientenoptimierungsaufgabe mit linearen Ungleichungen als Nebenbedingungen. Wir assoziieren zu jeder dieser vektoriellen Aufgaben eine skalare Aufgabe für welche die Dualität betrachtet wird. Die Formulierung der beiden skalaren Dualaufgaben führt uns zu der Konstruktion der Mehrzieloptimierungsaufgabe. Die Existenz von schwacher und starker Dualität wird bewiesen. Wir schliessen unsere Untersuchungen ab, indem wir eine Analyse von verschiedenen Dualitätskonzepten in der Mehrzieloptimierung durchführen. Zu einer allgemeinen Mehrzieloptimierungsaufgabe mit Kegel-Ungleichungen als Nebenbedingungen werden sechs verschiedene Dualaufgaben eingeführt, für die sowohl schwache als auch starke Dualitätsaussagen gezeigt werden. Danach leiten wir verschiedene Beziehungen zwischen den Bildmengen, bzw., zwischen den Mengen der maximalen Elemente dieser Bildmengen der sechs Dualaufgaben her. Dazu zeigen wir unter welchen Bedingungen werden diese Mengen identisch. Ein allgemeines Schema das die Beziehungen zwischen den sechs dualen Mehrzieloptimierungsaufgaben und andere Dualaufgaben aus der Literatur enthält, wird dargestellt.
204

Farkas - type results for convex and non - convex inequality systems

Hodrea, Ioan Bogdan 22 January 2008 (has links) (PDF)
As the title already suggests the aim of the present work is to present Farkas - type results for inequality systems involving convex and/or non - convex functions. To be able to give the desired results, we treat optimization problems which involve convex and composed convex functions or non - convex functions like DC functions or fractions. To be able to use the fruitful Fenchel - Lagrange duality approach, to the primal problem we attach an equivalent problem which is a convex optimization problem. After giving a dual problem to the problem we initially treat, we provide weak necessary conditions which secure strong duality, i.e., the case when the optimal objective value of the primal problem coincides with the optimal objective value of the dual problem and, moreover, the dual problem has an optimal solution. Further, two ideas are followed. Firstly, using the weak and strong duality between the primal problem and the dual problem, we are able to give necessary and sufficient optimality conditions for the optimal solutions of the primal problem. Secondly, provided that no duality gap lies between the primal problem and its Fenchel - Lagrange - type dual we are able to demonstrate some Farkas - type results and thus to underline once more the connections between the theorems of the alternative and the theory of duality. One statement of the above mentioned Farkas - type results is characterized using only epigraphs of functions. We conclude our investigations by providing necessary and sufficient optimality conditions for a multiobjective programming problem involving composed convex functions. Using the well-known linear scalarization to the primal multiobjective program a family of scalar optimization problems is attached. Further to each of these scalar problems the Fenchel - Lagrange dual problem is determined. Making use of the weak and strong duality between the scalarized problem and its dual the desired optimality conditions are proved. Moreover, the way the dual problem of the scalarized problem looks like gives us an idea about how to construct a vector dual problem to the initial one. Further weak and strong vector duality assertions are provided.
205

Nouvelles méthodes de calcul pour la prédiction des interactions protéine-protéine au niveau structural / Novel computational methods to predict protein-protein interactions on the structural level

Popov, Petr 28 January 2015 (has links)
Le docking moléculaire est une méthode permettant de prédire l'orientation d'une molécule donnée relativement à une autre lorsque celles-ci forment un complexe. Le premier algorithme de docking moléculaire a vu jour en 1990 afin de trouver de nouveaux candidats face à la protéase du VIH-1. Depuis, l'utilisation de protocoles de docking est devenue une pratique standard dans le domaine de la conception de nouveaux médicaments. Typiquement, un protocole de docking comporte plusieurs phases. Il requiert l'échantillonnage exhaustif du site d'interaction où les éléments impliqués sont considérées rigides. Des algorithmes de clustering sont utilisés afin de regrouper les candidats à l'appariement similaires. Des méthodes d'affinage sont appliquées pour prendre en compte la flexibilité au sein complexe moléculaire et afin d'éliminer de possibles artefacts de docking. Enfin, des algorithmes d'évaluation sont utilisés pour sélectionner les meilleurs candidats pour le docking. Cette thèse présente de nouveaux algorithmes de protocoles de docking qui facilitent la prédiction des structures de complexes protéinaires, une des cibles les plus importantes parmi les cibles visées par les méthodes de conception de médicaments. Une première contribution concerne l‘algorithme Docktrina qui permet de prédire les conformations de trimères protéinaires triangulaires. Celui-ci prend en entrée des prédictions de contacts paire-à-paire à partir d'hypothèse de corps rigides. Ensuite toutes les combinaisons possibles de paires de monomères sont évalués à l'aide d'un test de distance RMSD efficace. Cette méthode à la fois rapide et efficace améliore l'état de l'art sur les protéines trimères. Deuxièmement, nous présentons RigidRMSD une librairie C++ qui évalue en temps constant les distances RMSD entre conformations moléculaires correspondant à des transformations rigides. Cette librairie est en pratique utile lors du clustering de positions de docking, conduisant à des temps de calcul améliorés d'un facteur dix, comparé aux temps de calcul des algorithmes standards. Une troisième contribution concerne KSENIA, une fonction d'évaluation à base de connaissance pour l'étude des interactions protéine-protéine. Le problème de la reconstruction de fonction d'évaluation est alors formulé et résolu comme un problème d'optimisation convexe. Quatrièmement, CARBON, un nouvel algorithme pour l'affinage des candidats au docking basés sur des modèles corps-rigides est proposé. Le problème d'optimisation de corps-rigides est vu comme le calcul de trajectoires quasi-statiques de corps rigides influencés par la fonction énergie. CARBON fonctionne aussi bien avec un champ de force classique qu'avec une fonction d'évaluation à base de connaissance. CARBON est aussi utile pour l'affinage de complexes moléculaires qui comportent des clashes stériques modérés à importants. Finalement, une nouvelle méthode permet d'estimer les capacités de prédiction des fonctions d'évaluation. Celle-ci permet d‘évaluer de façon rigoureuse la performance de la fonction d'évaluation concernée sur des benchmarks de complexes moléculaires. La méthode manipule la distribution des scores attribués et non pas directement les scores de conformations particulières, ce qui la rend avantageuse au regard des critères standard basés sur le score le plus élevé. Les méthodes décrites au sein de la thèse sont testées et validées sur différents benchmarks protéines-protéines. Les algorithmes implémentés ont été utilisés avec succès pour la compétition CAPRI concernant la prédiction de complexes protéine-protéine. La méthodologie développée peut facilement être adaptée pour de la reconnaissance d'autres types d'interactions moléculaires impliquant par exemple des ligands, de l'ARN… Les implémentations en C++ des différents algorithmes présentés seront mises à disposition comme SAMSON Elements de la plateforme logicielle SAMSON sur http://www.samson-connect.net ou sur http://nano-d.inrialpes.fr/software. / Molecular docking is a method that predicts orientation of one molecule with respect to another one when forming a complex. The first computational method of molecular docking was applied to find new candidates against HIV-1 protease in 1990. Since then, using of docking pipelines has become a standard practice in drug discovery. Typically, a docking protocol comprises different phases. The exhaustive sampling of the binding site upon rigid-body approximation of the docking subunits is required. Clustering algorithms are used to group similar binding candidates. Refinement methods are applied to take into account flexibility of the molecular complex and to eliminate possible docking artefacts. Finally, scoring algorithms are employed to select the best binding candidates. The current thesis presents novel algorithms of docking protocols that facilitate structure prediction of protein complexes, which belong to one of the most important target classes in the structure-based drug design. First, DockTrina - a new algorithm to predict conformations of triangular protein trimers (i.e. trimers with pair-wise contacts between all three pairs of proteins) is presented. The method takes as input pair-wise contact predictions from a rigid-body docking program. It then scans and scores all possible combinations of pairs of monomers using a very fast root mean square deviation (RMSD) test. Being fast and efficient, DockTrina outperforms state-of-the-art computational methods dedicated to predict structure of protein oligomers on the collected benchmark of protein trimers. Second, RigidRMSD - a C++ library that in constant time computes RMSDs between molecular poses corresponding to rigid-body transformations is presented. The library is practically useful for clustering docking poses, resulting in ten times speed up compared to standard RMSD-based clustering algorithms. Third, KSENIA - a novel knowledge-based scoring function for protein-protein interactions is developed. The problem of scoring function reconstruction is formulated and solved as a convex optimization problem. As a result, KSENIA is a smooth function and, thus, is suitable for the gradient-base refinement of molecular structures. Remarkably, it is shown that native interfaces of protein complexes provide sufficient information to reconstruct a well-discriminative scoring function. Fourth, CARBON - a new algorithm for the rigid-body refinement of docking candidates is proposed. The rigid-body optimization problem is viewed as the calculation of quasi-static trajectories of rigid bodies influenced by the energy function. To circumvent the typical problem of incorrect stepsizes for rotation and translation movements of molecular complexes, the concept of controlled advancement is introduced. CARBON works well both in combination with a classical force-field and a knowledge-based scoring function. CARBON is also suitable for refinement of molecular complexes with moderate and large steric clashes between its subunits. Finally, a novel method to evaluate prediction capability of scoring functions is introduced. It allows to rigorously assess the performance of the scoring function of interest on benchmarks of molecular complexes. The method manipulates with the score distributions rather than with scores of particular conformations, which makes it advantageous compared to the standard hit-rate criteria. The methods described in the thesis are tested and validated on various protein-protein benchmarks. The implemented algorithms are successfully used in the CAPRI contest for structure prediction of protein-protein complexes. The developed methodology can be easily adapted to the recognition of other types of molecular interactions, involving ligands, polysaccharides, RNAs, etc. The C++ versions of the presented algorithms will be made available as SAMSON Elements for the SAMSON software platform at http://www.samson-connect.net or at http://nano-d.inrialpes.fr/software.
206

Integrality Gaps for Strong Linear Programming and Semidefinite Programming Relaxations

Georgiou, Konstantinos 17 February 2011 (has links)
The inapproximability for NP-hard combinatorial optimization problems lies in the heart of theoretical computer science. A negative result can be either conditional, where the starting point is a complexity assumption, or unconditional, where the inapproximability holds for a restricted model of computation. Algorithms based on Linear Programming (LP) and Semidefinite Programming (SDP) relaxations are among the most prominent models of computation. The related and common measure of efficiency is the integrality gap, which sets the limitations of the associated algorithmic schemes. A number of systematic procedures, known as lift-and-project systems, have been proposed to improve the integrality gap of standard relaxations. These systems build strong hierarchies of either LP relaxations, such as the Lovasz-Schrijver (LS) and the Sherali-Adams (SA) systems, or SDP relaxations, such as the Lovasz-Schrijver SDP (LS+), the Sherali-Adams SDP (SA+) and the Lasserre (La) systems. In this thesis we prove integrality gap lower bounds for the aforementioned lift-and-project systems and for a number of combinatorial optimization problems, whose inapproximability is yet unresolved. Given that lift-and-project systems produce relaxations that have given the best algorithms known for a series of combinatorial problems, the lower bounds can be read as strong evidence of the inapproximability of the corresponding optimization problems. From the results found in the thesis we highlight the following: For every epsilon>0, the level-Omega(sqrt(log n/ log log n)) LS+ relaxation of the Vertex Cover polytope has integrality gap 2-epsilon. The integrality gap of the standard SDP for Vertex Cover remains 2-o(1) even if all hypermetric inequalities are added to the relaxation. The resulting relaxations are incomparable to the SDP relaxations derived by the LS+ system. Finally, the addition of all ell1 inequalities eliminates all solutions not in the integral hull. For every epsilon>0, the level-Omega(sqrt(log n/ log log n)) SA relaxation of Vertex Cover has integrality gap 2-epsilon. The integrality gap remains tight even for superconstant-level SA+ relaxations. We prove a tight lower bound for the number of tightenings that the SA system needs in order to prove the Pigeonhole Principle. We also prove sublinear and linear rank bounds for the La and SA systems respectively for the Tseitin tautology. Linear levels of the SA+ system treat highly unsatisfiable instances of fixed predicate-P constraint satisfaction problems over q-ary alphabets as fully satisfiable, when the satisfying assignments of the predicates P can be equipped with a balanced and pairwise independent distribution. We study the performance of the Lasserre system on the cut polytope. When the input is the complete graph on 2d+1 vertices, we show that the integrality gap is at least 1+1/(4d(d+1)) for the level-d SDP relaxation.
207

Contribution à l'Étude des Techniques Hybrides de Localisation dans les Réseaux Sans-fil Hétérogènes

Laaraiedh, Mohamed 15 December 2010 (has links) (PDF)
Les avancements récents dans les technologies sans fil ont vu l'émergence de techniques de localisation qui constitue une base utile et rentable pour offrir des nouveaux services. Ces services topo-dépendants ont été de plus en plus bénéfiques pour les opérateurs et les entreprises de télécommunications. Divers services topo-dépendants peuvent être offerts 'a l'utilisateur tels que le suivi, la publicité, la sécurité, et la gestion. Les réseaux sans fil eux-mêmes peuvent bénéficier de l'information de localisation pour améliorer les performances de leurs différentes couches. Le routage, la synchronisation et l'annulation d'interférences sont quelques exemples o'u l'information de localisation peut être fructueuse. Un système de localisation doit être capable d'exécuter deux tâches principales : la mesure des paramètres topo-dépendants (RSSI, TOA, et TDOA) et l'estimation de la position en utilisant des estimateurs appropriés. L'objectif principal de cette thèse est l'étude de différentes techniques d'estimation de la position: algébriques et géométriques. Les techniques algébriques étudiées sont les moindres carrés, le maximum de vraisemblance, et la programmation semi-définie. La technique géométrique RGPA proposée est basée sur l'analyse par intervalles et la représentation géométrique des paramètres topo-dépendants. L'accent est mis sur la fusion de différents paramètres topo-dépendants et son influence sur la précision de positionnement. L'estimation et la mesure des paramètres topo-dépendants sont également étudiées en utilisant une campagne de mesures ULB afin d'avoir une compréhension complète du domaine de localisation.
208

Performance evaluation and enhancement for AF two-way relaying in the presence of channel estimation error

Wang, Chenyuan 30 April 2012 (has links)
Cooperative relaying is a promising diversity achieving technique to provide reliable transmission, high throughput and extensive coverage for wireless networks in a variety of applications. Two-way relaying is a spectrally efficient protocol, providing one solution to overcome the half-duplex loss in one-way relay channels. Moreover, incorporating the multiple-input-multiple-output (MIMO) technology can further improve the spectral efficiency and diversity gain. A lot of related work has been performed on the two-way relay network (TWRN), but most of them assume perfect channel state information (CSI). In a realistic scenario, however, the channel is estimated and the estimation error exists. So in this thesis, we explicitly take into account the CSI error, and investigate its impact on the performance of amplify-and-forward (AF) TWRN where either multiple distributed single-antenna relays or a single multiple-antenna relay station is exploited. For the distributed relay network, we consider imperfect self-interference cancellation at both sources that exchange information with the help of multiple relays, and maximal ratio combining (MRC) is then applied to improve the decision statistics under imperfect signal detection. The system performance degradation in terms of outage probability and average bit-error rate (BER) are analyzed, as well as their asymptotic trend. To further improve the spectral efficiency while maintain the spatial diversity, we utilize the maximum minimum (Max-Min) relay selection (RS), and examine the impact of imperfect CSI on this single RS scheme. To mitigate the negative effect of imperfect CSI, we resort to adaptive power allocation (PA) by minimizing either the outage probability or the average BER, which can be cast as a Geometric Programming (GP) problem. Numerical results verify the correctness of our analysis and show that the adaptive PA scheme outperforms the equal PA scheme under the aggregated effect of imperfect CSI. When employing a single MIMO relay, the problem of robust MIMO relay design has been dealt with by considering the fact that only imperfect CSI is available. We design the MIMO relay based upon the CSI estimates, where the estimation errors are included to attain the robust design under the worst-case philosophy. The optimization problem corresponding to the robust MIMO relay design is shown to be nonconvex. This motivates the pursuit of semidefinite relaxation (SDR) coupled with the randomization technique to obtain computationally efficient high-quality approximate solutions. Numerical simulations compare the proposed MIMO relay with the existing nonrobust method, and therefore validate its robustness against the channel uncertainty. / Graduate
209

Nouvelles méthodes de calcul pour la prédiction des interactions protéine-protéine au niveau structural / Novel computational methods to predict protein-protein interactions on the structural level

Popov, Petr 28 January 2015 (has links)
Le docking moléculaire est une méthode permettant de prédire l'orientation d'une molécule donnée relativement à une autre lorsque celles-ci forment un complexe. Le premier algorithme de docking moléculaire a vu jour en 1990 afin de trouver de nouveaux candidats face à la protéase du VIH-1. Depuis, l'utilisation de protocoles de docking est devenue une pratique standard dans le domaine de la conception de nouveaux médicaments. Typiquement, un protocole de docking comporte plusieurs phases. Il requiert l'échantillonnage exhaustif du site d'interaction où les éléments impliqués sont considérées rigides. Des algorithmes de clustering sont utilisés afin de regrouper les candidats à l'appariement similaires. Des méthodes d'affinage sont appliquées pour prendre en compte la flexibilité au sein complexe moléculaire et afin d'éliminer de possibles artefacts de docking. Enfin, des algorithmes d'évaluation sont utilisés pour sélectionner les meilleurs candidats pour le docking. Cette thèse présente de nouveaux algorithmes de protocoles de docking qui facilitent la prédiction des structures de complexes protéinaires, une des cibles les plus importantes parmi les cibles visées par les méthodes de conception de médicaments. Une première contribution concerne l‘algorithme Docktrina qui permet de prédire les conformations de trimères protéinaires triangulaires. Celui-ci prend en entrée des prédictions de contacts paire-à-paire à partir d'hypothèse de corps rigides. Ensuite toutes les combinaisons possibles de paires de monomères sont évalués à l'aide d'un test de distance RMSD efficace. Cette méthode à la fois rapide et efficace améliore l'état de l'art sur les protéines trimères. Deuxièmement, nous présentons RigidRMSD une librairie C++ qui évalue en temps constant les distances RMSD entre conformations moléculaires correspondant à des transformations rigides. Cette librairie est en pratique utile lors du clustering de positions de docking, conduisant à des temps de calcul améliorés d'un facteur dix, comparé aux temps de calcul des algorithmes standards. Une troisième contribution concerne KSENIA, une fonction d'évaluation à base de connaissance pour l'étude des interactions protéine-protéine. Le problème de la reconstruction de fonction d'évaluation est alors formulé et résolu comme un problème d'optimisation convexe. Quatrièmement, CARBON, un nouvel algorithme pour l'affinage des candidats au docking basés sur des modèles corps-rigides est proposé. Le problème d'optimisation de corps-rigides est vu comme le calcul de trajectoires quasi-statiques de corps rigides influencés par la fonction énergie. CARBON fonctionne aussi bien avec un champ de force classique qu'avec une fonction d'évaluation à base de connaissance. CARBON est aussi utile pour l'affinage de complexes moléculaires qui comportent des clashes stériques modérés à importants. Finalement, une nouvelle méthode permet d'estimer les capacités de prédiction des fonctions d'évaluation. Celle-ci permet d‘évaluer de façon rigoureuse la performance de la fonction d'évaluation concernée sur des benchmarks de complexes moléculaires. La méthode manipule la distribution des scores attribués et non pas directement les scores de conformations particulières, ce qui la rend avantageuse au regard des critères standard basés sur le score le plus élevé. Les méthodes décrites au sein de la thèse sont testées et validées sur différents benchmarks protéines-protéines. Les algorithmes implémentés ont été utilisés avec succès pour la compétition CAPRI concernant la prédiction de complexes protéine-protéine. La méthodologie développée peut facilement être adaptée pour de la reconnaissance d'autres types d'interactions moléculaires impliquant par exemple des ligands, de l'ARN… Les implémentations en C++ des différents algorithmes présentés seront mises à disposition comme SAMSON Elements de la plateforme logicielle SAMSON sur http://www.samson-connect.net ou sur http://nano-d.inrialpes.fr/software. / Molecular docking is a method that predicts orientation of one molecule with respect to another one when forming a complex. The first computational method of molecular docking was applied to find new candidates against HIV-1 protease in 1990. Since then, using of docking pipelines has become a standard practice in drug discovery. Typically, a docking protocol comprises different phases. The exhaustive sampling of the binding site upon rigid-body approximation of the docking subunits is required. Clustering algorithms are used to group similar binding candidates. Refinement methods are applied to take into account flexibility of the molecular complex and to eliminate possible docking artefacts. Finally, scoring algorithms are employed to select the best binding candidates. The current thesis presents novel algorithms of docking protocols that facilitate structure prediction of protein complexes, which belong to one of the most important target classes in the structure-based drug design. First, DockTrina - a new algorithm to predict conformations of triangular protein trimers (i.e. trimers with pair-wise contacts between all three pairs of proteins) is presented. The method takes as input pair-wise contact predictions from a rigid-body docking program. It then scans and scores all possible combinations of pairs of monomers using a very fast root mean square deviation (RMSD) test. Being fast and efficient, DockTrina outperforms state-of-the-art computational methods dedicated to predict structure of protein oligomers on the collected benchmark of protein trimers. Second, RigidRMSD - a C++ library that in constant time computes RMSDs between molecular poses corresponding to rigid-body transformations is presented. The library is practically useful for clustering docking poses, resulting in ten times speed up compared to standard RMSD-based clustering algorithms. Third, KSENIA - a novel knowledge-based scoring function for protein-protein interactions is developed. The problem of scoring function reconstruction is formulated and solved as a convex optimization problem. As a result, KSENIA is a smooth function and, thus, is suitable for the gradient-base refinement of molecular structures. Remarkably, it is shown that native interfaces of protein complexes provide sufficient information to reconstruct a well-discriminative scoring function. Fourth, CARBON - a new algorithm for the rigid-body refinement of docking candidates is proposed. The rigid-body optimization problem is viewed as the calculation of quasi-static trajectories of rigid bodies influenced by the energy function. To circumvent the typical problem of incorrect stepsizes for rotation and translation movements of molecular complexes, the concept of controlled advancement is introduced. CARBON works well both in combination with a classical force-field and a knowledge-based scoring function. CARBON is also suitable for refinement of molecular complexes with moderate and large steric clashes between its subunits. Finally, a novel method to evaluate prediction capability of scoring functions is introduced. It allows to rigorously assess the performance of the scoring function of interest on benchmarks of molecular complexes. The method manipulates with the score distributions rather than with scores of particular conformations, which makes it advantageous compared to the standard hit-rate criteria. The methods described in the thesis are tested and validated on various protein-protein benchmarks. The implemented algorithms are successfully used in the CAPRI contest for structure prediction of protein-protein complexes. The developed methodology can be easily adapted to the recognition of other types of molecular interactions, involving ligands, polysaccharides, RNAs, etc. The C++ versions of the presented algorithms will be made available as SAMSON Elements for the SAMSON software platform at http://www.samson-connect.net or at http://nano-d.inrialpes.fr/software.
210

[en] A NEURAL NETWORK FOR ONLINE PORTFOLIO SELECTION WITH SIDE INFORMATION / [pt] UMA REDE NEURAL PARA O PROBLEMA DE SELEÇÃO ONLINE DE PORTFÓLIO COM INFORMAÇÃO LATERAL

GUILHERME AUGUSTO SCHUTZ 15 January 2019 (has links)
[pt] O mercado financeiro é essencial na economia, trazendo estabilidade, acesso a novos tipos de investimentos, e aumentando a capacidade das empresas no acesso ao crédito. A constante busca por reduzir o papel de especialistas humanos na tomada de decisão, visa reduzir o risco inerente as emoções intrínsecas do ser humano, do qual a máquina não compartilha. Como consequência, reduzindo efeitos especulativos no mercado, e aumentando a precisão nas decisões tomadas. Neste trabalho é discutido o problema de seleção de portfólios online, onde um vetor de alocações de ativos é requerido em cada passo. O algoritmo proposto é o multilayer perceptron with side information - MLPi. Este algoritmo utiliza redes neurais para a solução do problema quando o investidor tem acesso a informações futuras sobre o preço dos ativos. Para avaliar o uso da informação lateral na seleção de portfolio, testamos empiricamente o MLPi em contraste com dois algoritmos, um baseline e o estado-da-arte. Como baseline é utilizado o buy-and-hold. O estado-da-arte é o algoritmo online moving average mean reversion proposto por Li e Hoi (2012). Para avaliar a utilização de informação lateral no algoritmo MLPi é definido um benchmark baseado numa solução ótima simples utilizando a informação lateral, mas sem considerar a acurácia da informação futura. Para os experimentos, utilizamos informações a nível de minuto do mercado de ações brasileiro, operados na bolsa de valores B3. É simulado um preditor de preço com 7 níveis de acurácia diferentes para 200 portfólios. Os resultados apontam que tanto o benchmark quanto o MLPi superam os dois algoritmos selecionados, para níveis de acurácia de um ativo maiores que 50 por cento, e na média, o MLPi supera o benchmark em todos os níveis de acurácia simulados. / [en] The financial market is essential in the economy, bringing stability, access to new types of investments, and increasing the ability of companies to access credit. The constant search for reducing the role of human specialists in decision making aims to reduce the risk inherent in the intrinsic emotions of the human being, which the machine does not share. As a consequence, reducing speculative effects in the market, and increasing the precision in the decisions taken. In this paper, we discuss the problem of selecting portfolios online, where a vector of asset allocations is required in each step. The proposed algorithm is the multilayer perceptron with side information - MLPi. This algorithm uses neural networks to solve the problem when the investor has access to future information on the price of the assets. To evaluate the use of side information in portfolio selection, we empirically tested MLPi in contrast to two algorithms, a baseline and the state-of-the-art. As a baseline, buy-andhold is used. The state-of-the-art is the online moving average mean reversion algorithm proposed by Li and Hoi (2012). To evaluate the use of side information in the algorithm MLPi a benchmark based on a simple optimal solution using the side information is defined, but without considering the accuracy of the future information. For the experiments, we use minute-level information from the Brazilian stock market, traded on the B3 stock exchange. A price predictor is simulated with 7 different accuracy levels for 200 portfolios. The results show that both the benchmark and MLPi outperform the two algorithms selected, for asset accuracy levels greater than 50 percent, and on average, MLPi outperforms the benchmark at all levels of simulated accuracy.

Page generated in 0.1272 seconds