Return to search

Multivariate Information Measures

<div>Many important scientific, engineering, and societal challenges involve large systems of individual agents or components interacting in complex ways. For example, to understand the emergence of consciousness, we study the dendritic integration in neurons; to prevent disease and rumor outbreaks, we trace the dynamics of social networks; to perform complicated scientific experiments, we separate and control the independent variables. Collectively, the interactions between individual neurons/agents/variables are often non-linear, i.e., a subset of the agents jointly behave in a manner unlike the marginal behaviors of the individuals.</div><div><br></div><div>The goal of this thesis is to construct a theoretical framework for measuring, comparing, and representing complex interactions in stochastic systems. Specifically, tools from information theory, differential geometry, lattice theory, and linear algebra are used to identify and characterize higher-order interactions among random variables.</div><div><br></div><div>We first propose measures of unique, redundant, and synergistic interactions for small stochastic systems using information projections for the exponential family. Their magnitudes are endowed with information theoretical meanings naturally, since they are measured by the Kullback-Leibler divergence. We prove that these quantities satisfy various desired properties.</div><div><br></div><div>We next apply these measures to hypothesis testing and network communication. We interpret the unique information as the two types of error components in a hypothesis testing problem. We analytically show that there is a duality between the synergistic and redundant information in Gaussian Multiple Access Channels (MAC) and Broadcast Channels (BC). We establish a novel duality between the partial information decomposition components for MAC and BC in the general case.</div><div><br></div><div>We lastly propose a new concept of representing the partial information decomposition framework with random variables. We give necessary and sufficient conditions for the representation under the assumption of Gaussianity and develop a construction method.</div><div><br></div><div>This research has the potential to advance the fields of information theory, statistics, and machine learning by contributing novel ideas, implementing these ideas with innovative tools, and constructing new simulation methods.</div>

  1. 10.25394/pgs.17263361.v1
Identiferoai:union.ndltd.org:purdue.edu/oai:figshare.com:article/17263361
Date18 December 2021
CreatorsXueyan Niu (11850761)
Source SetsPurdue University
Detected LanguageEnglish
TypeText, Thesis
RightsIn Copyright
Relationhttps://figshare.com/articles/thesis/Multivariate_Information_Measures/17263361

Page generated in 0.0019 seconds