Multi-view multi-instance multi-label learning (M3L) can model complex objects (bags) that are composed of multiple instances, represented with heterogeneous feature views and annotated with multiple related semantic labels. Although significant progress has been made toward M3L tasks, the current solutions still focus on a single-type of complex objects, and cannot effectively mine the widely-witnessed interconnected objects of multi-types. To bridge this gap, we propose a Deep Multi-type objects Multi-view Multi-instance Multi-label Learning solution (DeepM4L) based on heterogeneous network embedding. DeepM4L first encodes the inter- and intra-relations among multi-type objects using a heterogeneous network, and performs instance neighbor embedding to learn the representation vectors of instances. Next, it obtains the instance-label score tensor for each view and uses a max pooling operation to induce the bag-label score tensor for each bag. After that, it combines bag-label scores by multi-view learning to guarantee the semantic consistency between bags of different views. Our empirical study on benchmark datasets shows that DeepM4L is significantly superior to the recent advanced baselines.
|Original language||English (US)|
|Title of host publication||Proceedings of the 2021 SIAM International Conference on Data Mining (SDM)|
|Publisher||Society for Industrial and Applied Mathematics|
|Number of pages||9|
|State||Published - Apr 26 2021|