Advancing AI: A Comprehensive Study of Novel Machine Learning Architectures

  • Authors

    • Ridwan Boya Marqas Computer Engineering
    • Saman M. Almufty Reasearcher at Computer Science dept., Knowledge University, Erbil, Iraq
    • Renas R. Asaad Reasearcher at Computer Science dept., Knowledge University, Erbil, Iraq
    • Dr. Tamara Saad mohamed Computer Technics Department,Kut university college, Kut, Iraq

    Received date: January 1, 2025

    Accepted date: February 9, 2025

    Published date: February 20, 2025

    https://doi.org/10.14419/kwb24564
  • Artificial Intelligence, Machine Learning Architectures, Transformers, Graph Neural Networks, Capsule Networks, Spiking Neural Networks, Explainable AI, Edge Computing, Quantum Machine Learning, Few-Shot Learning, Scalability.
  • Abstract

    The rapid evolution of machine learning (ML) and artificial intelligence (AI) has led to groundbreaking advancements in computational models, empowering applications across diverse domains. This paper provides an in-depth exploration of advanced ML architectures, including transformers, Graph Neural Networks (GNNs), capsule networks, spiking neural networks (SNNs), and hybrid models. These architectures address the limitations of traditional models like convolutional and recurrent neural networks, offering superior accuracy, scalability, and efficiency for complex data. Key applications are discussed, ranging from healthcare diagnostics and drug discovery to financial fraud detection, autonomous systems, and logistics optimization. Despite their potential, these architectures face challenges such as computational overhead, scalability, and interpretability, necessitating interdisciplinary solutions. The paper also outlines future directions in edge computing, explainable AI, quantum machine learning, and few-shot learning, emphasizing the transformative role of advanced ML architectures in reshaping AI’s future.

  • References

    1. Arrieta, A. B., et al. (2020). Explainable Artificial Intelligence (XAI): Concepts, taxonomies, opportunities, and challenges. Infor-mation Fusion, 58, 82–115. https://doi.org/10.1016/j.inffus.2019.12.012
    2. Brown, T., et al. (2020). Language models are few-shot learners. NeurIPS, 33, 1877–1901.
    3. Devlin, J., et al. (2018). BERT: Pre-training of deep bidirectional transformers for language understanding. NAACL-HLT. https://arxiv.org/abs/1810.04805
    4. Dosovitskiy, A., et al. (2021). An image is worth 16x16 words: Transformers for image recognition at scale. ICLR. https://arxiv.org/abs/2010.11929
    5. Fan, X., et al. (2019). Graph-based supply chain optimization: A review and future directions. OR Spectrum, 41(3), 543–576. https://doi.org/10.1007/s00291-019-00553-w
    6. Finn, C., et al. (2017). Model-agnostic meta-learning for fast adaptation of deep networks. ICML. https://arxiv.org/abs/1703.03400
    7. Gilmer, J., et al. (2017). Neural message passing for quantum chemistry. ICML. https://arxiv.org/abs/1704.01212
    8. Jumper, J., et al. (2021). Highly accurate protein structure prediction with AlphaFold. Nature, 596(7873), 583–589. https://doi.org/10.1038/s41586-021-03819-2
    9. Kipf, T. N., & Welling, M. (2017). Semi-supervised classification with graph convolutional networks. ICLR. https://arxiv.org/abs/1609.02907
    10. Ruff, L., et al. (2019). Deep one-class classification. ICML. https://arxiv.org/abs/1801.05365
    11. Sabour, S., Frosst, N., & Hinton, G. E. (2017). Dynamic routing between capsules. NeurIPS, 30, 3856–3866. https://arxiv.org/abs/1710.09829
    12. Scarselli, F., et al. (2009). The graph neural network model. IEEE Transactions on Neural Networks, 20(1), 61–80. https://doi.org/10.1109/TNN.2008.2005605
    13. Strubell, E., Ganesh, A., & McCallum, A. (2019). Energy and policy considerations for deep learning in NLP. ACL. https://arxiv.org/abs/1906.02243
    14. Tavanaei, A., et al. (2019). Deep learning in spiking neural networks. Neural Networks, 111, 47–63. https://doi.org/10.1016/j.neunet.2018.12.002
    15. Vaswani, A., et al. (2017). Attention is all you need. NeurIPS, 30, 5998–6008. https://arxiv.org/abs/1706.03762
    16. Velickovic, P., et al. (2018). Graph attention networks. ICLR. https://arxiv.org/abs/1710.10903
    17. Yang, K., et al. (2020). Analyzing molecular targets using GNNs. Chemical Science, 11(5), 1231–1240. https://doi.org/10.1039/C9SC04032F
    18. Özyurt, F., Marqas, R., & Tuncer, S. A. (2024). Employing LSTM and Random Forest techniques for precision identification of misinformation. In New Trends in Computer Sciences (Vol. 1, pp. 46–65). All Sciences Academy. https://doi.org/10.5281/zenodo.10890229
    19. Awadh Al-Saiari, et al. (2024). Quantum-enhanced machine learning architectures. Quantum AI Journal, 4(3), 45–60. https://doi.org/10.48161/qaj.v4n3a760
    20. Zoph, B., et al. (2018). Learning transferable architectures for scalable image recognition. CVPR, 8697–8710. https://doi.org/10.1109/CVPR.2018.00907
    21. S. M. Abdulrahman, R. R. Asaad, H. B. Ahmad, A. Alaa Hani, S. R. M. Zeebaree, and A. B. Sallow, “Machine Learning in Nonline-ar Material Physics,” Journal of Soft Computing and Data Mining, vol. 5, no. 1, Jun. 2024, doi: 10.30880/jscdm.2024.05.01.010.
    22. A. B. Sallow, R. R. Asaad, H. B. Ahmad, S. Mohammed Abdulrahman, A. A. Hani, and S. R. M. Zeebaree, “Machine Learning Skills To K–12,” Journal of Soft Computing and Data Mining, vol. 5, no. 1, Jun. 2024, doi: 10.30880/jscdm.2024.05.01.011.
    23. S. M. Almufti et al., “INTELLIGENT HOME IOT DEVICES: AN EXPLORATION OF MACHINE LEARNING-BASED NET-WORKED TRAFFIC INVESTIGATION,” Jurnal Ilmiah Ilmu Terapan Universitas Jambi, vol. 8, no. 1, pp. 1–10, May 2024, doi: 10.22437/jiituj.v8i1.32767.
    24. S. M. Almufti and S. R. M. Zeebaree, “Leveraging Distributed Systems for Fault-Tolerant Cloud Computing: A Review of Strategies and Frameworks,” Academic Journal of Nawroz University, vol. 13, no. 2, pp. 9–29, May 2024, doi: 10.25007/ajnu.v13n2a2012.
    25. H. B. Ahmad, R. R. Asaad, S. M. Almufti, A. A. Hani, A. B. Sallow, and S. R. M. Zeebaree, “SMART HOME ENERGY SAVING WITH BIG DATA AND MACHINE LEARNING,” Jurnal Ilmiah Ilmu Terapan Universitas Jambi, vol. 8, no. 1, pp. 11–20, May 2024, doi: 10.22437/jiituj.v8i1.32598.
    26. T. Thirugnanam et al., “PIRAP: Medical Cancer Rehabilitation Healthcare Center Data Maintenance Based on IoT-Based Deep Fed-erated Collaborative Learning,” Int J Coop Inf Syst, vol. 33, no. 01, Mar. 2024, doi: 10.1142/S0218843023500053.
    27. R. Boya Marqas, S. M. Almufti, and R. Rajab Asaad, “FIREBASE EFFICIENCY IN CSV DATA EXCHANGE THROUGH PHP-BASED WEBSITES,” Academic Journal of Nawroz University, vol. 11, no. 3, pp. 410–414, Aug. 2022, doi: 10.25007/ajnu.v11n3a1480.
    28. S. M. Almufti, R. B. Marqas, Z. A. Nayef, and T. S. Mohamed, “Real Time Face-mask Detection with Arduino to Prevent COVID-19 Spreading,” Qubahan Academic Journal, vol. 1, no. 2, pp. 39–46, Apr. 2021, doi: 10.48161/qaj.v1n2a47.
    29. Mohamed, T. S., & Khalifah, S. M. (2022, December). Breast Cancer Prediction: The Classification of Non-Recurrence-Events and Recurrence-Events Using Functions Classifiers. In 2022 3rd Information Technology To Enhance e-learning and Other Application (IT-ELA) (pp. 55-60). IEEE.
  • Downloads

  • How to Cite

    Boya Marqas , R., Saman M. Almufty, Renas R. Asaad, & Tamara Saad mohamed, D. (2025). Advancing AI: A Comprehensive Study of Novel Machine Learning Architectures. International Journal of Scientific World, 11(1), 48-85. https://doi.org/10.14419/kwb24564

    Received date: January 1, 2025

    Accepted date: February 9, 2025

    Published date: February 20, 2025