Mesopotamian Journal of Big Data
https://journals.mesopotamian.press/index.php/bigdata
<p style="text-align: justify;">Attention scholars and researchers in the Big Data realm! The Mesopotamian Journal of Big Data, already with three published issues, invites your cutting-edge contributions to shape the future of this field. Our platform aims to disseminate groundbreaking discoveries and transformative applications in Big Data, emphasizing data analytics, machine learning, and related areas. We encourage interdisciplinary collaboration to drive advancements in this rapidly evolving domain. Your expertise is crucial—join us in this impactful journey.Submit your research to the Mesopotamian Journal of Big Data and be a part of the vanguard shaping knowledge in this transformative field.</p>Mesopotamian Academic Pressen-USMesopotamian Journal of Big Data2958-6453Assessing the Transformative Influence of ChatGPT on Research Practices among Scholars in Pakistan
https://journals.mesopotamian.press/index.php/bigdata/article/view/234
<p>This article investigates the transformative impact of ChatGPT on research practices within the scholarly community in Pakistan. ChatGPT, a powerful AI language model, has added significant consideration for its possibility of improving academic research. Survey data was gathered via a structured questionnaire distributed to researchers in Pakistan. A total of 278 questionnaires were distributed for the randomly chosen sample, of which 223 were returned. For calculating descriptive statistics, SPSS was utilized. Results of the study indicated that 90% of scholars are familiar with the practice of ChatGPT in research activities. 86% of scholars used 3.5 (Basic version) of ChatGPT for their research and only 14% used 4 (Plus version) of ChatGPT for their research work. The overall satisfaction level was 46% response satisfied with the usage of ChatGPT in research activities. The article discusses how ChatGPT's natural language processing capabilities have advanced literature reviews, data analysis, and content generation, thereby saving time and fostering greater productivity. Moreover, it examines how the tool’s accessibility and affordability have democratized research, making it more inclusive and open to a broader range of scholars. By shedding light on these critical aspects, this article provides valuable insights into the evolving landscape of research practices in Pakistan and highlights the potential for ChatGPT to revolutionize academic scholarship in the digital age.</p>Nayab Arshad Mehran Ullah BaberAdnan Ullah
Copyright (c) 2024 Adnan Ullah, Mehran Ullah Baber, Nayab Arshad
https://creativecommons.org/licenses/by/4.0
2024-01-102024-01-10202411010.58496/MJBD/2024/001Leveraging AI and Big Data in Low-Resource Healthcare Settings
https://journals.mesopotamian.press/index.php/bigdata/article/view/337
<p>Big data and artificial intelligence are game-changing technologies for the underdeveloped healthcare industry because they help optimize the entire supply chain and deliver more exact patient outcome information. Machine learning approaches that have recently seen more growing popularity include deep learning models that have brought revolution within the healthcare system in the previous years due to more complicated data compared to previous years . Machine learning is an essential data analysis procedure to describe efficient and effective methods to extract hidden information from large amounts of data that it would take logical analytics too long to manage. Recent years have seen an expansion and growth of advanced intelligent systems that have been able to learn more about clinical treatments and glean untapped medical information emanating from vast quantities of data when it comes to drug discovery and chemistry. The aim of this chapter is, therefore, to assess which big data and artificial intelligence approaches are prevalent in healthcare systems by investigating the most advanced big data structures, applications, and industry trends today available. First and foremost, the purpose is to provide a comprehensive overview of how the artificial intelligence and big data models can allocation in healthcare solutions fill the gap between machine learning approaches’ lack of human coverage and the healthcare data’s complexity. Moreover, current artificial intelligence technologies, including generative models, Bayesian deep learning, reinforcement learning, and self-driving laboratories, are also increasingly being used for drug discovery and chemistry . Finally, the work presents the existing open challenges and the future directions in the drug formulation development field. To this end, the review will cover on published algorithms/automation tools for artificial intelligence applied to large scale-data in the case of healthcare .</p>Ahmed Hussein Ali Saad Ahmed Dheyab Abdullah Hussein Alamoodi Aws Abed Al Raheem Magableh Yuantong Gu
Copyright (c) 2024 Ahmed Hussein Ali , Saad Ahmed Dheyab , Luís Martínez , , Iman Mohamad Sharaf , Abdullah Hussein Alamoodi , Aws Abed Al Raheem Magableh , Witold Pedrycz , Yuantong Gu
https://creativecommons.org/licenses/by/4.0
2024-02-142024-02-142024112210.58496/MJBD/2024/002Enhancing XML-based Compiler Construction with Large Language Models: A Novel Approach
https://journals.mesopotamian.press/index.php/bigdata/article/view/343
<p>Considering the prevailing rule of Large Language Models (LLMs) applications and the benefits of XML in a compiler context. This manuscript explores the synergistic integration of Large Language Models with XML-based compiler tools and advanced computing technologies. Marking a significant stride toward redefining compiler construction and data representation paradigms. As computing power and internet proliferation advance, XML emerges as a pivotal technology for representing, exchanging, and transforming documents and data. This study builds on the foundational work of Chomsky's Context-Free Grammar (CFG). Recognized for their critical role in compiler construction, to address and mitigate the speed penalties associated with traditional compiler systems and parser generators through the development of an efficient XML parser generator employing compiler techniques. Our research employs a methodical approach to harness the sophisticated capabilities of LLMs, alongside XML technologies. The key is to automate grammar optimization, facilitate natural language processing capabilities, and pioneer advanced parsing algorithms. To demonstrate their effectiveness, we thoroughly run experiments and compare them to other techniques. This way, we call attention to the efficiency, adaptability, and user-friendliness of the XML-based compiler tools with the help of these integrations. And the target will be the elimination of left-recursive grammars and the development of a global schema for LL(1) grammars, the latter taking advantage of the XML technology, to support the LL(1) grammars construction. The findings in this research not only underscore the signification of these innovations in the field of compilation construction but also indicate a paradigm move towards the use of AI technologies and XML in the context of the resolution of programming traditional issues. The outlined methodology serves as a roadmap for future research and development in compiler technology, which paves the way for open-source software to sweep across all fields. Gradually ushering in a new era of compiler technology featuring better efficiency, adaptability, and all CFGs processed through existing XML utilities on a global basis.</p>Idrees A. ZahidShahad Sabbar Joudar
Copyright (c) 2024 Idrees A. Zahid, Shahad Sabbar Joudar
https://creativecommons.org/licenses/by/4.0
2024-03-202024-03-202024233910.58496/MJBD/2024/003Agent-Interacted Big Data-Driven Dynamic Cartoon Video Generator
https://journals.mesopotamian.press/index.php/bigdata/article/view/366
<p>This study presents a novel method for animating videos using three Kaggle cartoon faces data sets. Dynamic interactions between cartoon agents and random backgrounds, as well as Gaussian blur, rotation, and noise addition, make cartoon visuals look better. This approach also evaluates video quality and animation design by calculating the backdrop colour's average and standard deviation, ensuring visually appealing material. This technology uses massive datasets to generate attractive animated videos for entertainment, teaching, and marketing.</p>Yasmin Makki MohialdenAbbas Akram khorsheedNadia Mahmood Hussien
Copyright (c) 2024 Yasmin Makki Mohialden, Abbas Akram khorsheed, Nadia Mahmood Hussien
https://creativecommons.org/licenses/by/4.0
2024-04-172024-04-172024404710.58496/MJBD/2024/004MLP and RBF Algorithms in Finance: Predicting and Classifying Stock Prices amidst Economic Policy Uncertainty
https://journals.mesopotamian.press/index.php/bigdata/article/view/386
<p>In the realm of stock market prediction and classification, the use of machine learning algorithms has gained significant attention. In this study, we explore the application of Multilayer Perceptron (MLP) and Radial Basis Function (RBF) algorithms in predicting and classifying stock prices, specifically amidst economic policy uncertainty. Stock market fluctuations are greatly influenced by economic policies implemented by governments and central banks. These policies can create uncertainty and volatility, which in turn makes accurate predictions and classifications of stock prices more challenging. By leveraging MLP and RBF algorithms, we aim to develop models that can effectively navigate these uncertainties and provide valuable insights to investors and financial analysts. The MLP algorithm, based on artificial neural networks, is able to learn complex patterns and relationships within financial data. The RBF algorithm, on the other hand, utilizes radial basis functions to capture non-linear relationships and identify hidden patterns within the data. By combining these algorithms, we aim to enhance the accuracy of stock price prediction and classification models. The results showed that both MLB and RBF predicted stock prices well for a group of countries using an index reflecting the impact of news on economic policy and expectations, where the MLB algorithm proved its ability to predict chain data. Countries were also classified according to stock price data and uncertainty in economic policy, allowing us to determine the best country to invest in according to the data. The uncertainty surrounding economic policy is what makes stock price forecasting so crucial. Investors must consider the degree of economic policy uncertainty and how it affects asset prices when deciding how to allocate their assets.</p>Bushra AliKhder AlakkariMostafa AbotalebMaad M MijwilKlodian Dhoska
Copyright (c) 2024 Bushra Ali, Khder Alakkari, Mostafa Abotaleb, Maad M Mijwil, Klodian Dhoska
https://creativecommons.org/licenses/by/4.0
2024-05-112024-05-112024486710.58496/MJBD/2024/005Generalized Time Domain Prediction Model for Motor Imagery-based Wheelchair Movement Control
https://journals.mesopotamian.press/index.php/bigdata/article/view/429
<p>Brain-computer interface (BCI-MI)-based wheelchair control is, in principle, an appropriate method for completely paralyzed people with a healthy brain. In a BCI-based wheelchair control system, pattern recognition in terms of preprocessing, feature extraction, and classification plays a significant role in avoiding recognition errors, which can lead to the initiation of the wrong command that will put the user in unsafe condition. Therefore, this research's goal is to create a time-domain generic pattern recognition model (GPRM) of two-class EEG-MI signals for use in a wheelchair control system.<br />This GPRM has the advantage of having a model that is applicable to unknown subjects, not just one. This GPRM has been developed, evaluated, and validated by utilizing two datasets, namely, the BCI Competition IV and the Emotive EPOC datasets. Initially, fifteen-time windows were investigated with seven machine learning methods to determine the optimal time window as well as the best classification method with strong generalizability. Evidently, the experimental results of this study revealed that the duration of the EEG-MI signal in the range of 4–6 seconds (4–6 s) has a high impact on the classification accuracy while extracting the signal features using five statistical methods. Additionally, the results demonstrate a one-second latency after each command cue when using the eight-second EEG-MI signal that the Graz protocol used in this study. This one-second latency is inevitable because it is practically impossible for the subjects to imagine their MI hand movement instantly. Therefore, at least one second is required for subjects to prepare to initiate their motor imagery hand movement. Practically, the five statistical methods are efficient and viable for decoding the EEG-MI signal in the time domain. Evidently, the GPRM model based on the LR classifier showed its ability to achieve an impressive classification accuracy of 90%, which was validated on the Emotive EPOC dataset. The GPRM developed in this study is highly adaptable and recommended for deployment in real-time EEG-MI-based wheelchair control systems.</p>Z.T. Al-Qaysi M. S Suzani Nazre bin Abdul Rashid Reem D. Ismail M.A. Ahmed Rasha A. AljanabiVeronica Gil-Costa
Copyright (c) 2024 Z.T. Al-Qaysi , M. S Suzani , Nazre bin Abdul Rashid , Reem D. Ismail , M.A. Ahmed , Rasha A. Aljanabi, Veronica Gil-Costa
https://creativecommons.org/licenses/by/4.0
2024-06-152024-06-152024688110.58496/MJBD/2024/006Face Morphing Attacks Detection Approaches: A Review
https://journals.mesopotamian.press/index.php/bigdata/article/view/475
<p>Face recognition systems (FRSs) that are applied by real-time applications such as border control are vulnerable to attacks such as face morphing, which blends two or more facial images into a single morphed image. The vulnerability of FRSs to many types of attacks, including both direct and indirect attacks, as well as face-morphing attacks, has garnered significant attention from the biometric field. A morphing attack aims to undermine the security of an FRS at an automated border control (ABC) gate by using an electronic machine-readable travel document (eMRTD) or e-passport that is acquired using a morphed face image. Most countries require applicants for an e-passport to present a passport photograph throughout the application process. A person with malicious intent and a collaborator can create a morphed facial image to illegally get an e-passport. A fraudulent individual, together with their accomplice, can exploit an e-passport with a morphed facial image to successfully travel through a border. Both individuals can authenticate the altered facial image, making it possible. A malicious individual could enter the border undetected, concealing their criminal history, while the access control system's log records information about their accomplice, posing a significant risk. This paper aims to provide a comprehensive overview of face morphing attacks and the developments happening in this field. We will go over the difficulties encountered, the methods for generating morphing images, and the pros and cons of these approaches. Along with the most important performance metrics that measure the efficiency of the algorithms used. The paper also covers the types of techniques used in deep learning and machine learning to detect and determine the attack of mutant faces. Indeed, it provides an overview of the most significant results from studies done in this area of research.</p>Essa Mokna NamisKhalid Shakir JasimSufyan Al-Janabi
Copyright (c) 2024 Essa Mokna Nomis, Khalid Shakir Jasim, Sufyan Al-Janabi
https://creativecommons.org/licenses/by/4.0
2024-07-202024-07-2020248210110.58496/MJBD/2024/007Advanced Machine Learning Approaches for Enhanced GDP Nowcasting in Syria Through Comprehensive Analysis of Regularization Techniques
https://journals.mesopotamian.press/index.php/bigdata/article/view/476
<p>This study addresses the challenge of nowcasting Gross Domestic Product (GDP) in data-scarce environments, with a focus on Syria, a country facing significant economic and political instability. Utilizing a dataset from 2010 to 2022, three machine learning algorithms Elastic Net, Ridge, and Lasso were applied to model GDP dynamics based on macroeconomic indicators, commodity prices, and high-frequency internet search data from Google Trends. Among these, the Lasso regression model, noted for its variable selection and sparsity promotion, proved most effective in capturing Syria's complex economic realities, achieving the lowest Root Mean Squared Error (RMSE) and Mean Absolute Percentage Error (MAPE). This accuracy highlights the Lasso model's capability to identify robust economic relationships despite limited data, thereby reducing overfitting and improving forecast generalizability. The study underscores the significant impact of non-traditional indicators, such as Google Trends Agriculture (GTA) and Google Trends Consumption (GTC), on GDP growth, offering valuable insights for policymakers and analysts in data-scarce environments. The findings support the use of machine learning techniques, particularly Lasso regression, as powerful tools for economic forecasting, enhancing informed decision-making in challenging settings.</p>Khder AlakkariBushra AliMostafa Abotaleb Rana Ali AbttanPushan Kumar Dutta
Copyright (c) 2024 Khder Alakkari, Bushra Ali, Mostafa Abotaleb , Rana Ali Abttan, Pushan Kumar Dutta
https://creativecommons.org/licenses/by/4.0
2024-08-022024-08-02202410211710.58496/MJBD/2024/008Using Data Anonymization in big data analytics security and privacy
https://journals.mesopotamian.press/index.php/bigdata/article/view/479
<p>Big Data and Analytics mean an enormous and complex collection of very diverse information, which is processed with various technologies and methods to produce and deliver useful and valuable insights. Analytics is the science of using data, or information to extract useful and actionable insights, facts and knowledge from a collection of data it could be stated that Big Data Analytics is the best thing since every commercial data system ever built, although everybody with a more optimistic vision of technology would like to take note that there is a fine line where Everything Data crosses the boundary to something else, especially with regard to privacy and security of the world as we know it. Privacy and security are two distinct but closely related phenomena. Whereas privacy refers to the control over access to the individual, security refers to the stability or strength of controls designed to protect the individual’s privacy. There are many obvious considerations and obstacles when attempting to securely share data. During big data analytics, many invasive techniques such as data fusion, cross-correlation, and algorithm training are often conducted over shared data, which can lead to severe privacy leaks. This means that every enterprise, organization, and individual maintaining large data repositories are in danger of being breached. Our study teaches us that security, privacy, and ethical concerns in big data analytics do not exist in parallel to the business cycle, but must be wisely and ethically managed in coherence throughout all emerging processes of the big data and information systems.</p>Abdulatif Ali HussainIsmael KhaleelTahsien Al-Quraishi
Copyright (c) 2024 Abdulatif Ali Hussain, Ismael Khaleel, Tahsien Al-Quraishi
https://creativecommons.org/licenses/by/4.0
2024-08-102024-08-10202411812710.58496/MJBD/2024/009Harnessing the Potential of Artificial Intelligence in Managing Viral Hepatitis
https://journals.mesopotamian.press/index.php/bigdata/article/view/484
<p>Viral hepatitis continues to be a serious global health concern, impacting millions of people, putting a strain on healthcare systems across the world, and causing significant morbidity and mortality. Traditional diagnostic, prognostic, and therapeutic procedures to address viral hepatitis are successful but have limits in accuracy, speed, and accessibility. Artificial intelligence (AI) advancement provides substantial opportunities to overcome these challenges. This study investigates the role of AI in revolutionizing viral hepatitis care, from early detection to therapy optimization and epidemiological surveillance. A comprehensive literature review was conducted using predefined keywords in the Nature, PLOS ONE, PubMed, Frontiers, Wiley Online Library, BMC, Taylor & Francis, Springer, ScienceDirect, MDPI, IEEE Xplore Digital Library, and Google Scholar databases. Peer-reviewed publications written in English between January 2019 and August 2024 were examined. The data of the selected research papers were synthesized and analyzed using thematic and narrative analysis techniques. The use of AI-driven algorithms in viral hepatitis control involves many significant aspects. AI improves diagnostic accuracy by integrating machine learning (ML) models with serological, genomic, and imaging data. It enables tailored treatment plans by assessing patient-specific characteristics and predicting therapy responses. AI-powered technologies aid in epidemiological modeling, and AI-powered systems effectively track treatment adherence, identify medication resistance, and control complications associated with chronic hepatitis infections. It is vital in identifying new antiviral medicines and vaccines, speeding the development pipeline through high-throughput screening and predictive modeling. Despite its transformational promise, using AI in viral hepatitis care presents various challenges, including data privacy concerns, the necessity for extensive and varied datasets, and the possibility of algorithmic biases. Ethical considerations, legal frameworks, and multidisciplinary collaboration are required to resolve these issues and ensure AI technology’s safe and successful use in clinical practice. Exploiting the full AI’s potential for viral hepatitis management provides unparalleled prospects to improve patient outcomes, optimize public health policies, and, eventually, and alleviate the disease’s negative impact worldwide. This study seeks to provide academics, medics, and policymakers with the fundamental knowledge they need to harness AI’s potential in the fight against viral hepatitis.</p>Guma AliMaad M. MijwilIoannis AdamopoulosBosco Apparatus BurugaMurat GökMalik Sallam
Copyright (c) 2024 Guma Ali, Maad M. Mijwil, Ioannis Adamopoulos, Bosco Apparatus Buruga, Murat Gök, Malik Sallam
https://creativecommons.org/licenses/by/4.0
2024-08-152024-08-15202412816310.58496/MJBD/2024/010Hybrid Model for Forecasting Temperature in Khartoum Based on CRU data
https://journals.mesopotamian.press/index.php/bigdata/article/view/490
<p>This consider leverages verifiable climatic data from the Climatic Research Unit (CRU), traversing from 1901 to 2022, to create progressed temperature forecasting models for Khartoum, Sudan. By applying state-of-the-art machine learning techniques, including Hybrid model, we aim to progress the precision of temperature forecasts in a semi-arid climate. The integration of long-term CRU data permits for the recognizable proof of climate patterns and patterns, upgrading the unwavering quality of short- and long-term forecasts. Moved forward temperature forecasting can altogether advantage basic segments empowering way better adjustment to climatic changes and extraordinary climate occasions. Our approach illustrates the potential of combining authentic climate data with machine learning to supply noteworthy experiences for climate flexibility.</p>Hussein Alkattan Alhumaima Ali Subhi Laith Farhan Ghazwan Al-mashhadani
Copyright (c) 2024 Hussein Alkattan , Alhumaima Ali Subhi , Laith Farhan , Ghazwan Al-mashhadani
https://creativecommons.org/licenses/by/4.0
2024-08-202024-08-20202416417410.58496/MJBD/2024/011A Framework for Automated Big Data Analytics in Cybersecurity Threat Detection
https://journals.mesopotamian.press/index.php/bigdata/article/view/544
<p>This research presents a novel framework designed to enhance cybersecurity through the integration of Big Data analytics, addressing the critical need for scalable and real-time threat detection in large-scale environments. Utilizing technologies such as Apache Kafka for efficient data ingestion, Apache Flink for stream processing, and advanced machine learning models like LSTM and Autoencoders, the framework offers robust anomaly detection capabilities. It also includes automated response mechanisms using SOAR and XDR systems, significantly improving response times and accuracy in threat mitigation. The proposed solution not only addresses current challenges in handling vast and complex data but also paves the way for future advancements, such as the integration of more sophisticated AI techniques and application across various domains, including IoT and cloud security. This research contributes to the field by providing a comprehensive, adaptive, and scalable framework that meets the demands of modern cybersecurity landscapes.</p>Mohamed Ariff AmeedeenRula A. HamidTheyazn H H AldhyaniLaith Abdul Khaliq Mohammed Al-NassrSunday Olusanya OlatunjiPriyavahani Subramanian
Copyright (c) 2024 Mohamed Ariff Ameedeen, Rula A. Hamid, Theyazn H H Aldhyani, Laith Abdul Khaliq Mohammed Al-Nassr
https://creativecommons.org/licenses/by/4.0
2024-09-252024-09-25202417518410.58496/MJBD/2024/012Deep Learning Approaches for Gender Classification from Facial Images
https://journals.mesopotamian.press/index.php/bigdata/article/view/553
<p>Gender recognition on the facial level is considered one of the most important technologies that finds use in such fields as a personalized marketing plan, safe systems of authentication, and effective human-computer interfaces. However, it has the following challenges; variation of lighting, facial movement, and ethnic/age face images. AI and DL has been improving on the effectiveness, flexibility, and speed of the gender classification system. AI enables complex and automatic feature learning in Data, while DL is tailored for handle variants in vision-based data. In this paper, we evaluated several architectures including Efficient Net_B2, ResNet50, ResNet18, and Lightning whilst determining the performance of the architectures in gender classification tasks. Self-assessment criteria included accuracy, precision, recall, and the F1-score. As for the performance, we found that ResNet18 had the highest scores on all the metrics, with the validation accuracy of above 98%, closely accompanied by the ResNet50 that, although it performed well as well, needed more epochs for convergence. The implications of this study for the development of future work in the gender classification technology include the discovery of ethnical, dependable, and effective techniques. Through the consideration of the state of the art and case studies, stakeholders can optimise the efficacy and the accountability of such systems, and thus support societal gains as a result of the improvement in technology.</p>Mustafa Abdulfattah HabeebYahya Layth KhaleelReem D. IsmailZ.T. Al-QaysiFatimah N. Ameen
Copyright (c) 2024 Mustafa Abdulfattah Habeeb, Yahya Layth Khaleel, Reem D. Ismail, Z.T. Al-Qaysi, Fatimah N. Ameen
https://creativecommons.org/licenses/by/4.0
2024-10-112024-10-11202418519810.58496/MJBD/2024/013Hybrid Spotted Hyena based Load Balancing algorithm (HSHLB)
https://journals.mesopotamian.press/index.php/bigdata/article/view/332
<p>In this paper, we delve into the core of our proposed dynamic load balancing mechanism, the Hybrid Spotted Hyena based Load Balancing algorithm (HSHLB). We begin by presenting a comprehensive overview of the Spotted Hyena Optimization Algorithm (SHOA) and the fundamental behaviors it encompasses, namely migration and attacking. We then introduce the Load Balancing Algorithm (LB) and outline its principles, emphasizing its distinct characteristics. Recognizing the strengths and weaknesses of both SHOA and LB, we embark on a journey to hybridize these two optimization techniques, resulting in the potent HSHLB. We elucidate the intricate process of combining these algorithms, elucidating how HSHLB harnesses their respective strengths while mitigating their limitations. This hybridization is pivotal to the overarching goal of achieving dynamic load balancing within cloud computing environments. As we progress through this section, we provide invaluable insights into the inner workings of HSHLB, offering readers a comprehensive understanding of its algorithmic steps, parameters, and intricacies. For clarity and enhanced comprehension, we incorporate pseudocode or flowcharts to illustrate the practical implementation of HSHLB. In sum, Section 3 lays the foundation for the practical application of HSHLB in achieving dynamic load balancing, setting the stage for subsequent sections where we delve into implementation details, results, and analysis. HSHLB emerges as a promising solution to the multifaceted challenges of load balancing in cloud computing, leveraging the unique strengths of SHOA and LB to optimize resource allocation and enhance Quality of Service (QoS).</p>Raed A.HasanMustafa M. AkaweeEnas F. AzizOmar A. HammoodAws Saad ShowketHusniyah JasimMohammed A. AlkhafajiOmar K. AhmedKhalil Yasin
Copyright (c) 2024 Raed A.Hasan, Mustafa M. Akawee, Enas F. Aziz, Omar A. Hammood, Aws Saad Showket, Husniyah Jasim, Mohammed A. Alkhafaji, Omar K. Ahmed, Khalil Yasin
https://creativecommons.org/licenses/by/4.0
2024-11-022024-11-02202419921010.58496/MJBD/2024/014Automated Water Quality Assessment Using Big Data Analytics
https://journals.mesopotamian.press/index.php/bigdata/article/view/582
<p>Water is one of the world's most precious resources, essential to life. Industrial waste, agricultural runoff, and urban discharge degrade water, rendering it unfit for consumption. Water quality monitoring and evaluation are more important than ever. Big Data analytics is used to examine water quality utilizing enormous datasets of pH, hardness, solids concentration, chloramine, sulfate, conductivity, organic carbon, trihalomethanes, and turbidity. This work classifies water potability, which is vital for human consumption, using strong machine learning on massive datasets. Classifiers were Random Forest, Gradient Boosting, and Support Vector Machine on 3,276 water bodies. The Random Forest classifier obtained the highest accuracy at 66.77% after significant data preparation and training, followed by Gradient Boosting at 66.01% and SVM at 62.80%. This shows that Big Data analytics and machine learning algorithms can interpret complex water quality data for public health and natural resource management.</p> <p>The Random Forest classifier and SVM in this study accurately calculate water potability. Prediction algorithms consider water cleanliness data and may aid public safety and water resource monitoring.</p> <p> </p>Yasmin Makki Mohialden Nadia Mahmood Hussien Saba Abdulbaqi Salman
Copyright (c) 2024 Yasmin Makki Mohialden , Nadia Mahmood Hussien , Saba Abdulbaqi Salman
https://creativecommons.org/licenses/by/4.0
2024-11-072024-11-07202421122210.58496/MJBD/2024/015