Learning prototypes for multiple instance learning

Learning prototypes for multiple instance learning

Multiple instance learning (MIL) is a weakly supervised learning method that works on the labeled bag of instances data. A prototypical network is a popular embedding approach in MIL. They overcome the common problems that other MIL approaches may have to deal with including dimensionality, loss of instance-level information, and complexity. They demonstrate competitive performance in classification. This work proposes a simple model that provides a permutation invariant prototype generator from a given MIL data set. We aim to find out prototypes in the feature space to map the collection of instances (i.e. bags) to a distance feature space and simultaneously learn a linear classifier for MIL. Another advantage of prototypical networks is that they are commonly used in the machine learning domain to facilitate interpretability. Our experiments on classical MIL benchmark data sets demonstrate that the proposed framework is an accurate and efficient classifier compared to the existing approaches.

___

  • [1] Al-Hussaini I, Xiao C, Westover MB, Sun J. SLEEPER: interpretable Sleep staging via Prototypes from Expert Rules. 2019.
  • [2] Simonyan K, Vedaldi A, Zisserman A. Deep Inside Convolutional Networks: Visualising Image Classification Models and Saliency Maps. Workshop at International Conference on Learning Representations 2014.
  • [3] Ilse M, Tomczak JM, Welling M. Attention-based Deep Multiple Instance Learning. Proceedings of the 35th International Conference on Machine Learning 2018; 80: 2127-2136.
  • [4] Montavon G, Samek W, Müller KR. Methods for Interpreting and Understanding Deep Neural Networks. Digital Signal Processing 2018; 73:1-15. doi: 10.1016/j.dsp.2017.10.011
  • [5] Viola P, Platt J, Zhang C. Multiple Instance Boosting for Object Detection. Advances in neural information processing systems 18 2005: 1417-1424.
  • [6] Alpaydin E, Cheplygina V, Loog M, Tax D. Single- vs. Multiple-Instance Classification. Pattern Recognition 2015; 48 (9). doi: 10.1016/j.patcog.2015.04.006
  • [7] Raykar VC, Krishnapuram B, Bi J, Dundar M, Rao RB. Bayesian multiple instance learning: automatic feature selection and inductive transfer. In Proceedings of the 25th international conference on Machine learning (ICML ’08) 2008. Association for Computing Machinery, New York, NY, USA, 808–815. doi: 10.1145/1390156.1390258
  • [8] Weidmann N, Frank E, Pfahringer B. A Two-level Learning Method for Generalized Multi-instance Problems. Machine Learning: ECML 2003: 468-479. doi: 10.1007/978-3-540-39857-8_42
  • [9] Li Y, Tax D, Duin R, Loog M. Multiple-instance learning as a classifier combining problem. Pattern Recognition 2013; 46 (3): 865–874. doi: 10.1016/j.patcog.2012.08.018
  • [10] Wang J, Zucker JD. Solving the multiple-instance problem: A lazy learning approach.Proceedings of the Seventeenth International Conference on Machine Learning (ICML 2000) 2000: 1119-1126.
  • [11] Dietterich T, Lathrop R, Lozano-Pérez T. Solving the Multiple Instance Problem with Axis-Parallel Rectangles. Artificial Intelligence 2001; 89 (1-2): 31-71. doi: 10.1016/S0004-3702(96)00034-3
  • [12] Cheplygina V, Tax D, Loog M. Multiple Instance Learning with Bag Dissimilarities. Pattern Recognition 2015; 48 (1): 264-275. 10.1016/j.patcog.2014.07.022.
  • [13] Tax D, Loog M, Duin R, Cheplygina V, Lee WJ. Bag Dissimilarities for Multiple Instance Learning. 7005. 222-234. doi: 10.1007/978-3-642-24471-1_16
  • [14] Pekalska E, Duin RPW. The Dissimilarity Representation for Pattern Recognition: Foundations and Applications. Singapore: World Scientific 2005.
  • [15] Chen Y, Bi J, Wang J. MILES: Multiple-Instance Learning via Embedded instance Selection. IEEE transactions on pattern analysis and machine intelligence 2007; 28: 1931-1947. doi:10.1109/TPAMI.2006.248
  • [16] Cheplygina V, Tax D, Loog M. Combining Instance Information to Classify Bags. Multiple Classifier Systems, submitted 2013. doi: 10.1007/978-3-642-38067-9_2.
  • [17] Cheplygina V, Tax DMJ, Loog M. Dissimilarity-Based Ensembles for Multiple Instance Learning. IEEE Transactions on Neural Networks and Learning Systems 2016; 27 (6): 1379-1391. doi: 10.1109/TNNLS.2015.2424254
  • [18] Yang HM, Zhang XY, Yin F, Liu CL. Robust Classification with Convolutional Prototype Learning. 2018. 3474- 3482. doi: 10.1109/CVPR.2018.00366.
  • [19] Snell J, Swersky K, Zemel R. Prototypical Networks for Few-shot Learning. Advances in Neural Information Processing Systems 2017; 30: 4077-4087.
  • [20] Wang X, Yan Y, Tang P, Bai X, and Liu W. Revisiting multiple instance neural networks: Pattern Recognition 2018; 74: 15–24. doi:10.1016/j.patcog.2017.08.026
  • [21] Wu J, Yu Y, Huang C, Yu K. Deep multiple instance learning for image classification and auto-annotation. 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR) 2015: 3460-3469. doi: 10.1109/CVPR.2015.7298968.
  • [22] Tang P, Wang X, Bai S, Shen W, Bai X et al. PCL: Proposal Cluster Learning for Weakly Supervised Object Detection. IEEE Transactions on Pattern Analysis and Machine Intelligence 2020; 42 (1): 176–191. doi:10.1109/TPAMI.2018.2876304
  • [23] Fang W, Chang L, Wei K, Xiangyang J, Jianbin J et al. C-MIL: Continuation Multiple Instance Learning for Weakly Supervised Object Detection 2019. IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR). doi: 10.1109/CVPR.2019.00230.
  • [24] Dennis DK, Pabbaraju C, Simhadri HV, Jain P. Multiple instance learning for efficient sequential data classification on resource-constrained devices. In Proceedings of the 32nd International Conference on Neural Information Processing Systems (NIPS’18) 2018. Curran Associates Inc., Red Hook, NY, USA, 10976–10987.
  • [25] Angelidis S, Lapata M. Multiple Instance Learning Networks for Fine-Grained Sentiment Analysis. Transactions of the Association for Computational Linguistics 2018; 6: 17-31. doi:10.1162/tacl_a_00002
  • [26] McFee B, Salamon J, Bello JP. Adaptive Pooling Operators for Weakly Labeled Sound Event Detection. IEEE/ACM Transactions on Audio, Speech and Language Processing 2018; 26 (11): 2180–2193.
  • [27] Ba JJ, Kiros RJ, Hinton GE. Layer Normalization. CoRR 2016. arXiv preprint arXiv:1607.06450
  • [28] Ioffe S, Szegedy C. Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift. Proceedings of the 32Nd International Conference on International Conference on Machine Learning 2015;
  • 37: 448-456. [29] Kingma DP, Ba J. Adam: A Method for Stochastic Optimization. 3rd International Conference on Learning Representations, ICLR 2015. https://dblp.org/rec/bib/journals/corr/KingmaB14
  • [30] Kucukasci E, Baydogan MG. Multiple Instance Learning Repository. 2018. http://www.multipleinstancelearning.com/
  • [31] Zhou ZH, Zhang ML. Solving multi-instance problems with classifier ensemble based on constructive clustering. Knowledge and Information Systems 2007; 11(2): 155-170. doi: 10.1007/s10115-006-0029-3
  • [32] Wei XS, Wu J, Zhou ZH. Scalable algorithms for multi-instance learning. IEEE transactions on neural networks and learning systems 2017; 28 (4): 975-987. doi: 10.1109/ICDM.2014.16
  • [33] Paszke A, Gross S, Chintala S, Chanan G, Yang A et. al. Automatic differentiation in PyTorch 2017.
  • [34] The GPyOpt authors. GPyOpt: A Bayesian Optimization framework in python. 2016. http://github.com/SheffieldML/GPyOpt/
  • [35] Friedman M. A Comparison of Alternative Tests of Significance for the Problem of m Rankings. The Annals of Mathematical Statistics 1940; 11 (1): 86-92.
  • [36] Nemenyi P. Distribution-free multiple comparisons. 1963.
Turkish Journal of Electrical Engineering and Computer Sciences-Cover
  • ISSN: 1300-0632
  • Yayın Aralığı: 6
  • Yayıncı: TÜBİTAK
Sayıdaki Diğer Makaleler

Robust and efficient EBG-backed wearable antenna for ISM applications

Ayesha Saeed, Humayun Shahid, Yasar Amin, Asma Ejaz, Hannu Tenhunen

Opinion dynamics of stubborn agents under the presence of a troll as differential game

Aykut YILDIZ, Arif Bülent ÖZGÜLER

Advanced single-loop discrete-time control for T-type voltage source inverter with minimum capacitor voltage ripple modulation

Manh Linh NGUYEN, Phuong VU

Scale-invariant histogram of oriented gradients: novel approach for pedestrian detection in multiresolution image dataset

Sweta PANIGRAHI, Surya Narayana Raju UNDI

Presentation attack detection for face recognition using remote photoplethysmography and cascaded fusion

Mehmet Fatih GÜNDOĞAR, Çiğdem EROĞLU ERDEM

HC-FFT: highly configurable and efficient FFT implementation on FPGA

Pakize ERGÜL, H. Fatih UĞURDAĞ, Doğancan DAVUTOĞLU

Design and analysis of a truncated elliptical-shaped chipless RFID tag

Ameer Taimour KHAN, Yassin ABDULLAH, Sidra FARHAT, Wasim NAWAZ, Usman RAUF

A hybrid approach based on transfer and ensemble learning for improving performances of deep learning models on small datasets

Aybars Uğur, Tunç Gültekin

Gene expression data classification using genetic algorithm-based feature selection

Tolga ENSARİ, Mustafa DAĞTEKİN, Öznur Sinem SÖNMEZ

Joint carrier frequency and phase offset estimation algorithm for CPM-DSSS based secure point-to-point communication

Saima SHEHZADI, Farzana KULSOOM, Muhammad ZEESHAN, Qasim Umar KHAN, Shahzad Amin SHEIKH