Loading...
MetaModelNet Based Dynamic Transfer Learning Framework for Long Tail Recognition


Journal of Computer and Communication Networks

Received On : 02 February 2025

Revised On : 23 March 2025

Accepted On : 18 April 2025

Published On : 02 May 2025

Volume 01, 2025

Pages : 096-107


Abstract

Long-tail recognition is a situation where the classes or categories in a given dataset are distributed in the same way as the long-tail. This implies that there are a few programs that have many instances and many classes with few instances (i.e. the head and tail classes respectively). In this article, we present MetaModelNet, a new framework that is designed to tackle the problems of long-tail identification in the computer vision field. MetaModelNet also incorporates advanced deep learning techniques in data acquisition, testing, and statistical analysis for the dynamic transfer learning of head classes with many samples to tail classes with fewer samples. MetaModelNet guarantees high accuracy on different datasets such as SUN-397, Places, and ImageNet due to proper design of the architecture. This involves adjusting the hyperparameters as well as assessing the measures for long-tail identification. MetaModelNet has been validated by comprehensive ablation testing and comparative analysis of the dynamic transfer learning efficacy and versatility. Therefore, from these tests, it has been observed that MetaModelNet has a higher recognition accuracy compared to the current methods. Furthermore, the detailed analysis of the relationships within the given model will contribute to the identification of new perspectives on the investigation of the regularization procedures and adaptive learning algorithms. This has the potential for the development of recognition systems that are capable of dealing with imbalanced data in various real-life applications in a stable and efficient manner.

Keywords

Long-Tail Recognition, Long-Tail Patterns, MetaModelNet, Dynamic Transfer Learning Techniques, Adaptive Learning Algorithms.

  1. R. Chellappa, S. Theodoridis, and A. van Schaik, “Advances in Machine Learning and Deep Neural Networks,” Proceedings of the IEEE, vol. 109, no. 5, pp. 607–611, May 2021, doi: 10.1109/jproc.2021.3072172.
  2. V. Sevetlidis, G. Pavlidis, S. Mouroutsos, and A. Gasteratos, “Tackling Dataset Bias With an Automated Collection of Real-World Samples,” IEEE Access, vol. 10, pp. 126832–126844, 2022, doi: 10.1109/access.2022.3226517.
  3. M. Saul and S. Rostami, “Assessing performance of artificial neural networks and re-sampling techniques for healthcare datasets,” Health Informatics Journal, vol. 28, no. 1, Jan. 2022, doi: 10.1177/14604582221087109.
  4. J. Chen and B. Su, “Instance-Specific Semantic Augmentation for Long-Tailed Image Classification,” IEEE Transactions on Image Processing, vol. 33, pp. 2544–2557, 2024, doi: 10.1109/tip.2024.3379929.
  5. P. Franco, J. M. Martinez, Y.-C. Kim, and M. A. Ahmed, “IoT Based Approach for Load Monitoring and Activity Recognition in Smart Homes,” IEEE Access, vol. 9, pp. 45325–45339, 2021, doi: 10.1109/access.2021.3067029.
  6. Y. Fu et al., “Long-tailed visual recognition with deep models: A methodological survey and evaluation,” Neurocomputing, vol. 509, pp. 290–309, Oct. 2022, doi: 10.1016/j.neucom.2022.08.031.
  7. K. T. Chitty-Venkata, M. Emani, V. Vishwanath, and A. K. Somani, “Neural Architecture Search Benchmarks: Insights and Survey,” IEEE Access, vol. 11, pp. 25217–25236, 2023, doi: 10.1109/access.2023.3253818.
  8. J. M. Johnson and T. M. Khoshgoftaar, “Survey on deep learning with class imbalance,” Journal of Big Data, vol. 6, no. 1, Mar. 2019, doi: 10.1186/s40537-019-0192-5.
  9. M. Buda, A. Maki, and M. A. Mazurowski, “A systematic study of the class imbalance problem in convolutional neural networks,” Neural Networks, vol. 106, pp. 249–259, Oct. 2018, doi: 10.1016/j.neunet.2018.07.011.
  10. S. S. Patil and S. P. Sonavane, “Improved classification of large imbalanced data sets using rationalized technique: Updated Class Purity Maximization Over_Sampling Technique (UCPMOT),” Journal of Big Data, vol. 4, no. 1, Dec. 2017, doi: 10.1186/s40537-017-0108-1.
  11. R. Miotto, F. Wang, S. Wang, X. Jiang, and J. T. Dudley, “Deep learning for healthcare: review, opportunities and challenges,” Briefings in Bioinformatics, vol. 19, no. 6, pp. 1236–1246, May 2017, doi: 10.1093/bib/bbx044.
  12. A. O’Mara-Eves, J. Thomas, J. McNaught, M. Miwa, and S. Ananiadou, “Using text mining for study identification in systematic reviews: a systematic review of current approaches,” Systematic Reviews, vol. 4, no. 1, Jan. 2015, doi: 10.1186/2046-4053-4-5.
  13. K. J. Chen, S. S. Gao, D. Duangthip, E. C. M. Lo, and C. H. Chu, “Early childhood caries and oral health care of Hong Kong preschool children,” Clinical, Cosmetic and Investigational Dentistry, vol. Volume 11, pp. 27–35, Jan. 2019, doi: 10.2147/ccide.s190993.
  14. M. A. Bolt, L. N. Killough, and H. C. Koh, “Testing the Interaction Effects of Task Complexity in Computer Training Using the Social Cognitive Model,” Decision Sciences, vol. 32, no. 1, pp. 1–20, Mar. 2001, doi: 10.1111/j.1540-5915.2001.tb00951.x.
  15. H. Zhu, L. Jiao, W. Ma, F. Liu, and W. Zhao, “A Novel Neural Network for Remote Sensing Image Matching,” IEEE Transactions on Neural Networks and Learning Systems, vol. 30, no. 9, pp. 2853–2865, Sep. 2019, doi: 10.1109/tnnls.2018.2888757.
  16. L. F. Hoogerheide, J. F. Kaashoek, and H. K. van Dijk, “On the shape of posterior densities and credible sets in instrumental variable regression models with reduced rank: An application of flexible sampling methods using neural networks,” Journal of Econometrics, vol. 139, no. 1, pp. 154–180, Jul. 2007, doi: 10.1016/j.jeconom.2006.06.009.
  17. M. Temraz and M. T. Keane, “Solving the class imbalance problem using a counterfactual method for data augmentation,” Machine Learning with Applications, vol. 9, p. 100375, Sep. 2022, doi: 10.1016/j.mlwa.2022.100375.
  18. Y. Zhang, B. Kang, B. Hooi, S. Yan, and J. Feng, “Deep Long-Tailed Learning: A Survey,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 45, no. 9, pp. 10795–10816, Sep. 2023, doi: 10.1109/tpami.2023.3268118.
  19. A. Mahanti, N. Carlsson, A. Mahanti, M. Arlitt, and C. Williamson, “A tale of the tails: Power-laws in internet measurements,” IEEE Network, vol. 27, no. 1, pp. 59–64, Jan. 2013, doi: 10.1109/mnet.2013.6423193.
  20. Z. Xiang, Z. Schwartz, J. H. Gerdes, and M. Uysal, “What can big data and text analytics tell us about hotel guest experience and satisfaction?,” International Journal of Hospitality Management, vol. 44, pp. 120–130, Jan. 2015, doi: 10.1016/j.ijhm.2014.10.013.
  21. M. Olmedilla, M. R. Martínez-Torres, and S. L. Toral, “The superhit effect and long tail phenomenon in the context of electronic word of mouth,” Decision Support Systems, vol. 125, p. 113120, Oct. 2019, doi: 10.1016/j.dss.2019.113120.
  22. T. Blaschke, “Object based image analysis for remote sensing,” ISPRS Journal of Photogrammetry and Remote Sensing, vol. 65, no. 1, pp. 2–16, Jan. 2010, doi: 10.1016/j.isprsjprs.2009.06.004.
  23. M. Huisman, J. N. van Rijn, and A. Plaat, “A survey of deep meta-learning,” Artificial Intelligence Review, vol. 54, no. 6, pp. 4483–4541, Apr. 2021, doi: 10.1007/s10462-021-10004-4.
  24. Z. Ji, Z. Hou, X. Liu, Y. Pang, and X. Li, “Memorizing Complementation Network for Few-Shot Class-Incremental Learning,” IEEE Transactions on Image Processing, vol. 32, pp. 937–948, 2023, doi: 10.1109/tip.2023.3236160.
  25. M. Davies, “Concept mapping, mind mapping and argument mapping: what are the differences and do they matter?,” Higher Education, vol. 62, no. 3, pp. 279–301, Nov. 2010, doi: 10.1007/s10734-010-9387-6.
  26. M. Haggag, S. Abdelhay, A. Mecheter, S. Gowid, F. Musharavati, and S. Ghani, “An Intelligent Hybrid Experimental-Based Deep Learning Algorithm for Tomato-Sorting Controllers,” IEEE Access, vol. 7, pp. 106890–106898, 2019, doi: 10.1109/access.2019.2932730.
  27. M. A. Mazurowski, P. A. Habas, J. M. Zurada, J. Y. Lo, J. A. Baker, and G. D. Tourassi, “Training neural network classifiers for medical decision making: The effects of imbalanced datasets on classification performance,” Neural Networks, vol. 21, no. 2–3, pp. 427–436, Mar. 2008, doi: 10.1016/j.neunet.2007.12.031.
  28. S. Soni, S. S. Chouhan, and S. S. Rathore, “TextConvoNet: a convolutional neural network based architecture for text classification,” Applied Intelligence, vol. 53, no. 11, pp. 14249–14268, Oct. 2022, doi: 10.1007/s10489-022-04221-9.
  29. J. P. Monteiro, D. Ramos, D. Carneiro, F. Duarte, J. M. Fernandes, and P. Novais, “Meta‐learning and the new challenges of machine learning,” International Journal of Intelligent Systems, vol. 36, no. 11, pp. 6240–6272, Jun. 2021, doi: 10.1002/int.22549.
  30. M. Hall, E. Frank, G. Holmes, B. Pfahringer, P. Reutemann, and I. H. Witten, “The WEKA data mining software,” ACM SIGKDD Explorations Newsletter, vol. 11, no. 1, pp. 10–18, Nov. 2009, doi: 10.1145/1656274.1656278.
  31. A. P. Dempster, N. M. Laird, and D. B. Rubin, “Maximum Likelihood from Incomplete Data Via the EM Algorithm,” Journal of the Royal Statistical Society Series B: Statistical Methodology, vol. 39, no. 1, pp. 1–22, Sep. 1977, doi: 10.1111/j.2517-6161.1977.tb01600.x.
  32. W. Wang, Z. Zhao, P. Wang, F. Su, and H. Meng, “Attentive feature augmentation for Long-Tailed visual recognition,” IEEE Transactions on Circuits and Systems for Video Technology, vol. 32, no. 9, pp. 5803–5816, Sep. 2022, doi: 10.1109/tcsvt.2022.3161427.
  33. M. Cumpston et al., “Updated guidance for trusted systematic reviews: a new edition of the Cochrane Handbook for Systematic Reviews of Interventions,” Cochrane Database of Systematic Reviews, Oct. 2019, doi: 10.1002/14651858.ed000142.
  34. M. Gevrey, I. Dimopoulos, and S. Lek, “Review and comparison of methods to study the contribution of variables in artificial neural network models,” Ecological Modelling, vol. 160, no. 3, pp. 249–264, Feb. 2003, doi: 10.1016/s0304-3800(02)00257-0.
  35. Z. Li and D. Hoiem, “Learning without Forgetting,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 40, no. 12, pp. 2935–2947, Dec. 2018, doi: 10.1109/tpami.2017.2773081.
  36. G. D. Bader and C. W. Hogue, “An automated method for finding molecular complexes in large protein interaction networks,” BMC Bioinformatics, vol. 4, no. 1, Jan. 2003, doi: 10.1186/1471-2105-4-2.
  37. V. M. Gonzalez, J. Goeppinger, and K. Lorig, “Four psychosocial theories and their application to patient education and clinical practice,” Arthritis & Rheumatism, vol. 3, no. 3, pp. 132–143, Sep. 1990, doi: 10.1002/art.1790030305.
  38. Y. S. Boutalis, T. Kottas, and M. A. Christodoulou, “Adaptive estimation of fuzzy cognitive maps with proven stability and parameter convergence,” IEEE Transactions on Fuzzy Systems, vol. 17, no. 4, pp. 874–889, Aug. 2009, doi: 10.1109/tfuzz.2009.2017519.
  39. T. Mutanda, D. Ramesh, S. Karthikeyan, S. Kumari, A. Anandraj, and F. Bux, “Bioprospecting for hyper-lipid producing microalgal strains for sustainable biofuel production,” Bioresource Technology, vol. 102, no. 1, pp. 57–70, Jan. 2011, doi: 10.1016/j.biortech.2010.06.077.
  40. A. Phaphuangwittayakul, Y. Guo, and F. Ying, “Fast Adaptive Meta-Learning for Few-Shot Image Generation,” IEEE Transactions on Multimedia, vol. 24, pp. 2205–2217, 2022, doi: 10.1109/tmm.2021.3077729.
  41. J. Guo and J. Guan, “Classification of Marine Plankton Based on Few-shot Learning,” Arabian Journal for Science and Engineering, vol. 46, no. 9, pp. 9253–9262, Jun. 2021, doi: 10.1007/s13369-021-05786-2.
  42. D. Saunders, “Domain Adaptation and Multi-Domain Adaptation for Neural Machine Translation: A Survey,” Journal of Artificial Intelligence Research, vol. 75, pp. 351–424, Sep. 2022, doi: 10.1613/jair.1.13566.
  43. Y. Zhang, X. Shi, S. Zhang, and A. Abraham, “A XGBoost-Based Lane Change Prediction on Time Series Data Using Feature Engineering for Autopilot Vehicles,” IEEE Transactions on Intelligent Transportation Systems, vol. 23, no. 10, pp. 19187–19200, Oct. 2022, doi: 10.1109/tits.2022.3170628.
  44. K. C. CHAN, G. A. KAROLYI, F. A. LONGSTAFF, and A. B. SANDERS, “An Empirical Comparison of Alternative Models of the Short‐Term Interest Rate,” The Journal of Finance, vol. 47, no. 3, pp. 1209–1227, Jul. 1992, doi: 10.1111/j.1540-6261.1992.tb04011.x.
  45. S. Ji, S. Pan, E. Cambria, P. Marttinen, and P. S. Yu, “A Survey on Knowledge Graphs: Representation, Acquisition, and Applications,” IEEE Transactions on Neural Networks and Learning Systems, vol. 33, no. 2, pp. 494–514, Feb. 2022, doi: 10.1109/tnnls.2021.3070843.
  46. D. Chen, Y. Lu, and C.-Y. Hsu, “Measurement Invariance Investigation for Performance of Deep Learning Architectures,” IEEE Access, vol. 10, pp. 78070–78087, 2022, doi: 10.1109/access.2022.3192468.
  47. R. Chambers, B. C. Y. Lo, and N. B. Allen, “The impact of intensive mindfulness training on attentional control, cognitive style, and affect,” Cognitive Therapy and Research, vol. 32, no. 3, pp. 303–322, Feb. 2007, doi: 10.1007/s10608-007-9119-0.
  48. T. H. Fischer and J. Almlof, “General methods for geometry and wave function optimization,” The Journal of Physical Chemistry, vol. 96, no. 24, pp. 9768–9774, Nov. 1992, doi: 10.1021/j100203a036.
  49. J. Kivinen and M. K. Warmuth, “Exponentiated Gradient versus Gradient Descent for Linear Predictors,” Information and Computation, vol. 132, no. 1, pp. 1–63, Jan. 1997, doi: 10.1006/inco.1996.2612.
  50. G. P. Cimellaro, A. M. Reinhorn, and M. Bruneau, “Framework for analytical quantification of disaster resilience,” Engineering Structures, vol. 32, no. 11, pp. 3639–3649, Nov. 2010, doi: 10.1016/j.engstruct.2010.08.008.
  51. T. D. A. Widhianingsih and D.-K. Kang, “Augmented domain agreement for adaptable Meta-Learner on Few-Shot classification,” Applied Intelligence, vol. 52, no. 7, pp. 7037–7053, Sep. 2021, doi: 10.1007/s10489-021-02744-1.
  52. D. Macêdo, C. Zanchettin, A. L. I. Oliveira, and T. Ludermir, “Enhancing batch normalized convolutional networks using displaced rectifier linear units: A systematic comparative study,” Expert Systems with Applications, vol. 124, pp. 271–281, Jun. 2019, doi: 10.1016/j.eswa.2019.01.066.
  53. “Visual Tracking: An Experimental Survey,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 36, no. 7, pp. 1442–1468, Jul. 2014, doi: 10.1109/tpami.2013.230.
  54. X. Xu, W. Li, and Q. Duan, “Transfer learning and SE-ResNet152 networks-based for small-scale unbalanced fish species identification,” Computers and Electronics in Agriculture, vol. 180, p. 105878, Jan. 2021, doi: 10.1016/j.compag.2020.105878.
  55. H. Liu, J. Feng, M. Qi, J. Jiang, and S. Yan, “End-to-End Comparative Attention Networks for Person Re-Identification,” IEEE Transactions on Image Processing, vol. 26, no. 7, pp. 3492–3506, Jul. 2017, doi: 10.1109/tip.2017.2700762.
  56. G. Patterson, C. Xu, H. Su, and J. Hays, “The SUN Attribute Database: Beyond Categories for Deeper Scene Understanding,” International Journal of Computer Vision, vol. 108, no. 1–2, pp. 59–81, Jan. 2014, doi: 10.1007/s11263-013-0695-z.
  57. K. G. Provan and H. B. Milward, “A Preliminary Theory of Interorganizational Network Effectiveness: A Comparative Study of Four Community Mental Health Systems,” Administrative Science Quarterly, vol. 40, no. 1, p. 1, Mar. 1995, doi: 10.2307/2393698.
  58. Jia Deng, Wei Dong, R. Socher, Li-Jia Li, Kai Li, and Li Fei-Fei, “ImageNet: A large-scale hierarchical image database,” 2009 IEEE Conference on Computer Vision and Pattern Recognition, Jun. 2009, doi: 10.1109/cvprw.2009.5206848.
  59. J. Kim, H. Yoon, and M.-S. Kim, “Tweaking Deep Neural Networks,” IEEE Transactions on Pattern Analysis and Machine Intelligence, pp. 1–1, 2021, doi: 10.1109/tpami.2021.3079511.
  60. Y. Xu and R. Goodacre, “On Splitting Training and Validation Set: A Comparative Study of Cross-Validation, Bootstrap and Systematic Sampling for Estimating the Generalization Performance of Supervised Learning,” Journal of Analysis and Testing, vol. 2, no. 3, pp. 249–262, Jul. 2018, doi: 10.1007/s41664-018-0068-2.
  61. E. S. Lander et al., “Initial sequencing and analysis of the human genome,” Nature, vol. 409, no. 6822, pp. 860–921, Feb. 2001, doi: 10.1038/35057062.
  62. R. T. Collins, A. J. Lipton, and T. Kanade, “Introduction to the special section on video surveillance,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 22, no. 8, pp. 745–746, Aug. 2000, doi: 10.1109/tpami.2000.868676.
  63. X. Yang, Q. Kuang, W. Zhang, and G. Zhang, “AMDO: An Over-Sampling Technique for Multi-Class Imbalanced Problems,” IEEE Transactions on Knowledge and Data Engineering, vol. 30, no. 9, pp. 1672–1685, Sep. 2018, doi: 10.1109/tkde.2017.2761347.
  64. S.-J. Yen and Y.-S. Lee, “Cluster-based under-sampling approaches for imbalanced data distributions,” Expert Systems with Applications, vol. 36, no. 3, pp. 5718–5727, Apr. 2009, doi: 10.1016/j.eswa.2008.06.108.
  65. Y. Sun, M. S. Kamel, A. K. C. Wong, and Y. Wang, “Cost-sensitive boosting for classification of imbalanced data,” Pattern Recognition, vol. 40, no. 12, pp. 3358–3378, Dec. 2007, doi: 10.1016/j.patcog.2007.04.009.
  66. B. Liu, H. Li, H. Kang, G. Hua, and N. Vasconcelos, “GistNet: a Geometric Structure Transfer Network for Long-Tailed Recognition,” 2021 IEEE/CVF International Conference on Computer Vision (ICCV), pp. 8189–8198, Oct. 2021, doi: 10.1109/iccv48922.2021.00810.
  67. M. K. Panda, B. N. Subudhi, T. Veerakumar, and V. Jakhetiya, “Modified ResNet-152 Network With Hybrid Pyramidal Pooling for Local Change Detection,” IEEE Transactions on Artificial Intelligence, vol. 5, no. 4, pp. 1599–1612, Apr. 2024, doi: 10.1109/tai.2023.3299903.
  68. C. Lu and P. Koniusz, “Few-shot Keypoint Detection with Uncertainty Learning for Unseen Species,” 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 19394–19404, Jun. 2022, doi: 10.1109/cvpr52688.2022.01881.
  69. M. B. Ferraro, R. Coppi, G. González Rodríguez, and A. Colubi, “A linear regression model for imprecise response,” International Journal of Approximate Reasoning, vol. 51, no. 7, pp. 759–770, Sep. 2010, doi: 10.1016/j.ijar.2010.04.003.
  70. A. Furutanpey, P. Raith, and S. Dustdar, “FrankenSplit: Efficient Neural Feature Compression With Shallow Variational Bottleneck Injection for Mobile Edge Computing,” IEEE Transactions on Mobile Computing, vol. 23, no. 12, pp. 10770–10786, Dec. 2024, doi: 10.1109/tmc.2024.3381952.
CRediT Author Statement

The author reviewed the paper and approved the final version of the manuscript.

Acknowledgements

Author(s) thanks to University of Essex for research lab and equipment support.

Funding

No funding was received to assist with the preparation of this manuscript.

Ethics Declarations

Conflict of interest

The authors have no conflicts of interest to declare that are relevant to the content of this article.

Availability of Data and Materials

Data sharing is not applicable to this article as no new data were created or analysed in this study.

Author Information

Contributions

All authors have equal contribution in the paper and all authors have read and agreed to the published version of the manuscript.

Corresponding Author



Rights and permissions

Open Access This article is licensed under a Creative Commons Attribution NoDerivs is a more restrictive license. It allows you to redistribute the material commercially or non-commercially but the user cannot make any changes whatsoever to the original, i.e. no derivatives of the original work. To view a copy of this license, visit: https://creativecommons.org/licenses/by-nc-nd/4.0/

Cite this Article

Hossein Anisi, “MetaModelNet Based Dynamic Transfer Learning Framework for Long Tail Recognition”, Journal of Computer and Communication Networks, pp. 096-107, 2025, doi: 10.64026/JCCN/2025010.

Copyright

© 2025 Hossein Anisi. This is an open access article distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.