TY - GEN
T1 - Heterogeneous Transfer Learning from a Partial Information Decomposition Perspective
AU - Gianini, Gabriele
AU - Barsotti, Annalisa
AU - Mio, Corrado
AU - Lin, Jianyi
PY - 2024
Y1 - 2024
N2 - Transfer Learning (TL) encompasses a number of Machine Learning Techniques that take a pre-trained model aimed at solving a task in a Source Domain and try to reuse it to improve the performance of a related task in a Target Domain An important issue in TL is that the effectiveness of those techniques is strongly dataset-dependent. In this work, we investigate the possible structural causes of the varying performance of Heterogeneous Transfer Learning (HTL) across domains characterized by different, but overlapping feature sets (this naturally determine a partition of the features into Source Domain specific subset, Target Domain specific subset, and shared subset). To this purpose, we use the Partial Information Decomposition (PID) framework, which breaks down the multivariate information that input variables hold about an output variable into three kinds of components: Unique, Synergistic, and Redundant. We consider that each domain can hold the PID components in implicit form: this restricts the information directly accessible to each domain. Based on the relative PID structure of the above mentioned feature subsets, the framework is able to tell, in principle: 1) which kind of information components are lost in passing from one domain to the other, 2) which kind of information components are at least implicitly available to a domain, and 3) what kind information components could be recovered through the bridge of the shared features. We show an example of a bridging scenario based on synthetic data.
AB - Transfer Learning (TL) encompasses a number of Machine Learning Techniques that take a pre-trained model aimed at solving a task in a Source Domain and try to reuse it to improve the performance of a related task in a Target Domain An important issue in TL is that the effectiveness of those techniques is strongly dataset-dependent. In this work, we investigate the possible structural causes of the varying performance of Heterogeneous Transfer Learning (HTL) across domains characterized by different, but overlapping feature sets (this naturally determine a partition of the features into Source Domain specific subset, Target Domain specific subset, and shared subset). To this purpose, we use the Partial Information Decomposition (PID) framework, which breaks down the multivariate information that input variables hold about an output variable into three kinds of components: Unique, Synergistic, and Redundant. We consider that each domain can hold the PID components in implicit form: this restricts the information directly accessible to each domain. Based on the relative PID structure of the above mentioned feature subsets, the framework is able to tell, in principle: 1) which kind of information components are lost in passing from one domain to the other, 2) which kind of information components are at least implicitly available to a domain, and 3) what kind information components could be recovered through the bridge of the shared features. We show an example of a bridging scenario based on synthetic data.
KW - Heterogeneous Transfer Learning
KW - Transferable Information Components
KW - Partial Information Decomposition
KW - Heterogeneous Transfer Learning
KW - Transferable Information Components
KW - Partial Information Decomposition
UR - http://hdl.handle.net/10807/279116
U2 - 10.1007/978-3-031-51643-6_10
DO - 10.1007/978-3-031-51643-6_10
M3 - Conference contribution
VL - 2022
T3 - COMMUNICATIONS IN COMPUTER AND INFORMATION SCIENCE
SP - 133
EP - 146
BT - Communications in Computer and Information Science
T2 - 15th International Conference on Management of Digital, MEDES 2023
Y2 - 5 May 2023 through 7 May 2023
ER -