ArticlePDF Available

Artificial Intelligence in Computer Science

Authors:

Abstract

Artificial Intelligence (AI) has emerged as a cornerstone of modern computer science, exerting a profound influence on diverse sectors of society. This article offers a comprehensive overview of the evolution and impact of AI in computer science. Beginning with its historical roots and development, the article traces AI's trajectory from its inception to its current state of advancement. Moreover, the article elucidates the techniques of AI in computer science, offering illuminating insights into a spectrum of methodologies including machine learning, deep learning, natural language processing, computer vision, knowledge representation and reasoning, recommender systems, and optimization techniques. In this context, the article explores the current applications of AI across various sectors, including engineering, medical fields, technology, military affairs, economy, education, entertainment, transportation, emphasizing its role in enhancing efficiency, productivity, and decision-making. In doing so, the article delves into the potential future impact of AI, envisioning a world where AI-powered technologies continue to revolutionize human-computer interaction, automation, and artificial general intelligence. To sum up, the article underscores the pivotal role of AI in shaping the future of computer science and society at large, advocating for continued research, development, and ethical stewardship of AI technologies.
IJEES
International Journal of Electrical Engineering
and Sustainability (IJEES)
ISSN (online): 2959-9229
https://ijees.org/index.php/ijees/index
Volume 2 | Number 2 | April-June 2024 | Pages 01-21
age | 1
Research Article
Artificial Intelligence in Computer Science
Mohamed Khaleel 1,2*, Abdullatif Jebrel 3
1 Department of Aeronautical Engineering, College of Civil Aviation, Misrata, Libya
2 Dept. of Electrical-Electronics Engineering, Faculty of Engineering, Karabuk University, Karabuk 78050, Turkey
3 Department of Computer science, College of Technical Sciences, Misrata, Libya
*Corresponding author: [email protected]; Tel.: +218 91 0123472
Received: January 20, 2024
Accepted: March 10, 2024
Published: March 28, 2024
This is an open access article under the BY-CC license
Abstract: Artificial Intelligence (AI) has emerged as a cornerstone of modern computer science, exerting a
profound influence on diverse sectors of society. This article offers a comprehensive overview of the evolution and
impact of AI in computer science. Beginning with its historical roots and development, the article traces AI's
trajectory from its inception to its current state of advancement. Moreover, the article elucidates the techniques of
AI in computer science, offering illuminating insights into a spectrum of methodologies including machine
learning, deep learning, natural language processing, computer vision, knowledge representation and reasoning,
recommender systems, and optimization techniques. In this context, the article explores the current applications of
AI across various sectors, including engineering, medical fields, technology, military affairs, economy, education,
entertainment, transportation, emphasizing its role in enhancing efficiency, productivity, and decision-making.
In doing so, the article delves into the potential future impact of AI, envisioning a world where AI-powered
technologies continue to revolutionize human-computer interaction, automation, and artificial general intelligence.
To sum up, the article underscores the pivotal role of AI in shaping the future of computer science and society at
large, advocating for continued research, development, and ethical stewardship of AI technologies.
Keywords: AI, Computer science, Classification, Techniques, Application, Challenges and limitations.
1. Introduction
Artificial Intelligence (AI) refers to the development of computer systems capable of performing
tasks that typically require human intelligence. These tasks may include learning from experience,
understanding natural language, recognizing patterns, reasoning, problem-solving, and adapting to
new situations. AI systems are designed to emulate human cognitive functions, enabling them to
perform complex tasks autonomously or with minimal human intervention [1,2]. In addition to that, AI
serves as the cornerstone of innovation in contemporary computing, driving value creation for
individuals and enterprises alike. A prime illustration of AI's impact is evident in optical character
recognition (OCR) technology, which leverages AI algorithms to extract textual information from
images and documents. By converting unstructured content into structured data, OCR not only
facilitates the organization and management of information but also unlocks invaluable insights for
businesses [3,4]. This transformative capability underscores AI's capacity to revolutionize traditional
workflows, streamline processes, and empower decision-making across diverse sectors.
AI constitutes a diverse field of scientific inquiry aimed at constructing computers and machines
capable of emulating human intelligence. This encompasses the ability to reason, learn, and make
decisions, tasks traditionally associated with human cognition, as well as handling data at scales beyond
human capacity for analysis [5]. In this regard, AI encompasses a broad spectrum of disciplines,
including computer science, data analytics, statistics, hardware and software engineering, linguistics,
Khaleel et al.
IJEES
Page | 2
neuroscience, philosophy, and psychology. It draws upon insights and methodologies from these fields
to advance its understanding and capabilities. In practical business applications, AI manifests as a
collection of technologies predominantly rooted in machine learning and deep learning. These
technologies are deployed for various purposes such as data analytics, predictive modeling, object
recognition, natural language processing, recommendation systems, intelligent data retrieval, and
more. By harnessing AI technologies, businesses can automate processes, gain actionable insights from
data, and enhance decision-making capabilities, thereby driving innovation and competitive advantage
in today's data-driven landscape [6-10].
A few scholarly investigations delve into the domain of artificial intelligence within the field of
computer science. According to [11], the study focused on analyzing the efficacy of integrating deep
learning and computer vision technologies. Deep learning achieved a groundbreaking milestone
through the construction of hierarchical neural networks, facilitating end-to-end feature learning and
semantic comprehension of images. The successful applications within the realm of computer vision
provided robust support for training deep learning algorithms. The seamless integration of these two
domains spawned a new generation of sophisticated computer vision systems, markedly surpassing
conventional methodologies in tasks such as machine vision image classification and object detection.
Within this study, typical image classification scenarios were examined to elucidate the superior
performance of deep neural network models, while also acknowledging their limitations in terms of
generalization and interpretability
The study [12] delved into the prospective ramifications of Generative Artificial Intelligence
(Generative AI) on developing nations, meticulously evaluating both advantageous and adverse
implications spanning multiple spheres encompassing information, culture, and industry. Generative
Artificial Intelligence denotes AI systems tasked with generating content, including text, audio, or
video, with the objective of generating innovative and imaginative outputs rooted in training data. In
contrast to conversational artificial intelligence, generative artificial intelligence systems possess the
distinctive capacity not only to furnish responses but also to author the substance of those retorts.
According to [13], the opaque nature inherent in artificial intelligence (AI) models has engendered
numerous apprehensions regarding their application in critical contexts. Explainable Artificial
Intelligence (XAI) emerged as a burgeoning research domain, dedicated to formulating machine
learning models capable of furnishing lucid and interpretable rationales for their decisions and actions.
Within the domain of network cybersecurity, XAI has emerged as a potent force poised to revolutionize
the approach toward network security, affording enhanced insights into the behaviors exhibited by
cyber threats and facilitating the development of more efficacious defensive strategies. This survey
embarked upon a comprehensive examination of the contemporary landscape within XAI for
cybersecurity in network systems, elucidating the diverse array of methodologies proposed to tackle
this pivotal challenge.
According to [14], profound understanding of AI principles and computer science fundamentals is
poised to become indispensable for prospective careers in science and engineering. Anticipating the
imminent future, employment opportunities are projected to predominantly revolve around AI-related
domains. Within this context, proficiency in AI and computer science is anticipated to attain a
significance akin to traditional literacy skills, such as reading and writing. Drawing upon this analogy,
a novel AI education concept has been conceived with the objective of nurturing AI literacy. This
conceptual framework encompasses modules tailored for diverse age groups and educational levels.
Key topics addressed across these modules include problem-solving through search algorithms, sorting
techniques, graph theory, and data structures.
The article presents a significant contribution by delving comprehensively into the historical origins
of AI, thereby offering profound insights into its evolutionary trajectory and foundational principles.
Additionally, it provides an exhaustive and systematic classification of artificial intelligence,
meticulously delineating the diverse paradigms and methodological approaches prevalent within the
domain. Moreover, the article elucidates the techniques of AI in computer science, offering illuminating
insights into a spectrum of methodologies including machine learning, deep learning, natural language
Khaleel et al.
IJEES
Page | 3
processing, computer vision, knowledge representation and reasoning, recommender systems, and
optimization techniques. This comprehensive overview underscores the breadth and depth of AI
applications across various domains. Subsequently, the article undertakes an exhaustive exploration of
the contemporary challenges and limitations of ai in computer science. It scrutinizes the hurdles and
constraints encountered in the practical implementation of AI solutions, particularly focusing on ethical
and societal concerns, interpretability and exploitability, safety and security risks, lack of generalization
and transfer learning, resource constraints and computational complexity, and human-ai collaboration
and interface design. To sum up, the article rigorously analyzes the intricate and multifaceted
relationship between AI and the broader landscape of computer science. This investigation highlights
the reciprocal interaction between AI and diverse subfields within computer science, demonstrating
how AI both derives benefits from and contributes to various disciplinary domains, thereby
underscoring its pervasive influence and interdisciplinary relevance.
Section 2 delves into the origins of AI, providing insights into the historical evolution and
foundational principles of this field. In Section 3, a comprehensive classification of artificial intelligence
is presented, outlining the various paradigms and approaches within the domain. Section 4 elucidates
the Techniques of AI in Computer Science, shedding light on the methodologies and algorithms
employed to develop intelligent systems. Following this, Section 5 offers an in-depth exploration of the
current challenges and limitations of AI in computer science, examining hurdles and constraints
encountered in the practical implementation of AI solutions. In Section 6, the focus shifts to result and
discussion, where the intricate and multifaceted relationship between AI and broader computer science
is analyzed. Lastly, Section 7 encapsulates the conclusion of the article, summarizing key findings and
insights gleaned from the discourse.
2. Origins of Artificial Intelligence
The roots of artificial intelligence can be traced back to the 1950s, when pioneering researchers began
exploring the possibility of creating machines capable of mimicking human intelligence. The term
"artificial intelligence" was first coined by John McCarthy at the Dartmouth Conference in 1956, which
is widely regarded as the birth of AI as a formal field of study. McCarthy, along with fellow visionaries
like Marvin Minsky, Herbert Simon, and Allen Newell, laid the foundation for AI by proposing
ambitious goals such as reasoning, learning, and problem-solving in machines.
A. Key Milestones in AI Development:
The Turing Test (1950): Proposed by British mathematician and computer scientist Alan Turing
in his seminal paper "Computing Machinery and Intelligence," the Turing Test serves as a
benchmark for evaluating a machine's ability to exhibit intelligent behavior indistinguishable
from that of a human. While the Turing Test remains a controversial metric, it sparked
considerable interest in the quest for artificial intelligence [15].
Expert Systems (1970s-1980s): Expert systems represent a pivotal milestone in AI development,
characterized by the creation of specialized programs capable of emulating human expertise in
specific domains. These systems relied on symbolic reasoning and knowledge representation
to solve complex problems and provide decision support in fields such as medicine, finance,
and engineering [16].
Neural Networks (1940s-1950s, resurgence in 1980s): Inspired by the structure and function of
the human brain, neural networks emerged as a foundational concept in AI research. While
early neural network models like the Perceptron showed promise, interest waned during the
"AI winter" of the 1970s and 1980s. However, a resurgence of interest in neural networks in the
late 1980s, fueled by advances in computing power and algorithmic innovations, laid the
groundwork for the modern era of deep learning [17].
Khaleel et al.
IJEES
Page | 4
B. Major Breakthroughs and Challenges:
Breakthroughs: Throughout its history, AI has witnessed numerous breakthroughs, including [18]:
IBM's Deep Blue defeating world chess champion Garry Kasparov in 1997, showcasing the
power of AI in mastering complex games and symbolic reasoning tasks.
The development of machine learning algorithms such as backpropagation and support
vector machines, which revolutionized pattern recognition and predictive modeling.
The emergence of deep learning in the early 21st century, enabling unprecedented
advancements in areas like image recognition, natural language processing, and autonomous
driving.
Challenges: AI has also faced significant challenges and setbacks, including [19,20]:
The "AI winter" periods of the 1970s and 1980s, characterized by diminished funding,
skepticism, and disillusionment with the promises of AI.
Ethical and societal concerns surrounding AI, including issues of bias, transparency, privacy,
and job displacement.
Technical challenges related to scalability, interpretability, and robustness of AI systems,
particularly in high-stakes applications such as healthcare and autonomous vehicles.
Despite these challenges, AI continues to evolve rapidly, driven by interdisciplinary research,
technological advancements, and growing societal demand for intelligent systems capable of
addressing complex problems and enhancing human capabilities.
3. Classification of artificial intelligence
The classification of artificial intelligence (AI) can be delineated through multiple paradigms,
contingent upon developmental phases or operational modalities. To clarify, four distinct stages of AI
evolution are widely acknowledged within scholarly discourse [21-25].
A. Reactive machines
Reactive machines represent a paradigm of AI characterized by their constrained functionality,
which is contingent upon predefined rules governing their response to stimuli. Lacking memory, these
AI systems operate without the capacity for learning from new data. A quintessential exemplar of this
category is IBM's Deep Blue, celebrated for its triumph over chess grandmaster Garry Kasparov in 1997,
thereby illustrating the prowess of reactive machines within a narrowly defined domain.
B. Limited memory
Limited memory AI epitomizes a contemporary manifestation of artificial intelligence, prevalent in
modern AI frameworks. These systems possess a memory component, facilitating incremental
improvement over time through exposure to new data. Typically, such advancement is achieved via
training methodologies employing artificial neural networks or analogous models. Deep learning, a
prominent subset of machine learning methodologies, epitomizes this form of AI, wherein incremental
enhancements are realized through the assimilation of new data, thereby exemplifying the essence of
limited memory artificial intelligence.
C. Theory of mind
Theory of mind AI represents a theoretical frontier within the realm of artificial intelligence,
currently absent in empirical instantiation but subject to ongoing exploration and inquiry. This
conceptual framework envisages AI systems endowed with cognitive faculties analogous to the human
mind, thereby manifesting decision-making capabilities commensurate with human cognition. Central
to this construct is the AI's capacity to recognize and retain emotional cues, alongside the ability to
navigate social contexts with the nuance and adaptability characteristic of human behavior. While
nascent, the pursuit of Theory of mind AI underscores a profound ambition to imbue machines with a
depth of understanding and responsiveness akin to human cognitive faculties.
Khaleel et al.
IJEES
Page | 5
D. Self-aware
Self-aware AI represents a conceptual leap beyond Theory of mind AI, postulating a hypothetical
realm wherein machines attain a level of existential awareness akin to human consciousness. This
theoretical construct envisions AI endowed with a profound cognition of its own existence, alongside
intellectual and emotional faculties comparable to those of humans. However, akin to Theory of mind
AI, empirical instantiation of self-aware AI remains elusive, existing solely within the realm of
speculative discourse and serving as a testament to the aspirational horizons of artificial intelligence
research.
4. Techniques AI in Computer Science
Artificial Intelligence (AI) encompasses a diverse array of techniques aimed at enabling machines to
exhibit intelligent behavior and perform tasks traditionally requiring human intelligence. These
techniques leverage mathematical, statistical, and computational methodologies to analyze data, make
decisions, and solve complex problems. Some of the key techniques in AI include:
A. Machine Learning (ML)
Machine learning techniques enable computers to learn from data and improve their performance
over time without explicit programming. Illustrated in Figure 1 are distinct classifications of machine
learning. Supervised Learning involves the training of models on labeled datasets to facilitate accurate
predictions or classifications. Conversely, Unsupervised Learning focuses on the identification of latent
patterns and structures inherent within unlabeled data. Additionally, Reinforcement Learning
embodies the process of acquiring optimal behaviors through iterative trial and error, leveraging
feedback obtained from interactions with the environment [26].
The landscape of Machine Learning (ML) has undergone a transformative shift from a realm
primarily characterized by artistic and scientific endeavors to one marked by its accessibility as a
technology ubiquitous to developers across diverse platforms. Foreseeably, a paradigm shift is
anticipated wherein trained models will become an integral component of every application, facilitating
data-driven decision-making processes beyond the capacity of developers to manually engineer.
However, this transition poses a formidable engineering hurdle, given the prevailing disconnect
between data science and modeling practices and conventional software development methodologies.
Figure 1. Type of machine learning
In response to these challenges, this paper [27] introduces ML.NET, a framework meticulously
developed at Microsoft over the past decade. ML.NET serves as a strategic response aimed at
streamlining the integration of machine learning models within large-scale software applications,
thereby mitigating the aforementioned obstacles and fostering a more seamless adoption of ML within
the developer community.
Khaleel et al.
IJEES
Page | 6
In [28], the ultimate objective of industrial machine learning (ML) projects invariably revolves
around the development and expedited deployment of ML products into operational environments.
Nevertheless, the process of automating and operationalizing ML products presents formidable
challenges, leading to the failure of numerous ML initiatives to meet their anticipated outcomes.
Addressing this predicament, the paradigm of Machine Learning Operations (MLOps) has emerged.
MLOps encompasses a multifaceted framework comprising best practices, conceptual frameworks, and
a development culture aimed at streamlining the operationalization of ML endeavors.
B. Deep Learning
Deep learning is a subset of machine learning that utilizes artificial neural networks with multiple
layers (deep architectures) to learn complex representations of data [29,30]. It has revolutionized tasks
such as image and speech recognition, natural language processing, and autonomous driving as
illustrated in Figure 2.
Figure 2. Deep learning.
Janiesch [31] posits that deep learning, rooted in artificial neural networks, epitomizes a paradigm
within machine learning. The efficacy of deep learning models often surpasses that of shallow machine
learning models and conventional data analysis methodologies across numerous applications. This
article endeavors to encapsulate the foundational tenets of both machine learning and deep learning,
thereby fostering a comprehensive comprehension of the systematic framework underpinning
contemporary intelligent systems.
Sarker's [32] article also provided a structured and comprehensive examination of deep learning
(DL) techniques, encompassing a taxonomy that considered various real-world tasks such as supervised
and unsupervised learning. Within our taxonomy, deep networks were delineated for supervised or
discriminative learning, unsupervised or generative learning, alongside hybrid learning and other
pertinent categories. Additionally, a synopsis of real-world application domains suitable for deep
learning techniques was outlined. Furthermore, ten potential avenues for future DL modeling, along
with corresponding research directions, were highlighted. The overarching objective of this article was
to present an overarching perspective on DL modeling, intended to serve as a reference guide for both
academic scholars and industry professionals.
In accordance with LeCun [33], deep learning facilitated the development of computational models
comprising multiple layers of processing, enabling the acquisition of data representations characterized
by diverse levels of abstraction. These methodologies have notably advanced the forefront of various
domains, including but not limited to speech recognition, visual object recognition, and object detection.
Deep learning techniques excel in discerning intricate structures within extensive datasets, employing
the backpropagation algorithm to guide adjustments to internal parameters, thereby refining the
computation of representations within each layer based on preceding layer representations. Specifically,
deep convolutional networks have heralded breakthroughs in the processing of images, video, speech,
and audio, while recurrent networks have illuminated sequential data modalities such as text and
speech.
Khaleel et al.
IJEES
Page | 7
C. Natural Language Processing (NLP)
Natural Language Processing (NLP) represents a multifaceted domain within artificial intelligence
(AI) and computational linguistics, dedicated to equipping computers with the capability to understand
and manipulate human language in various forms, including written text, spoken language, and
gestures [34]. Through the application of sophisticated algorithms and linguistic models, NLP enables
machines to perform a wide range of tasks, such as language translation, sentiment analysis, text
summarization, question answering, and speech recognition. Figure 3 presents natural language
processing.
Figure 3. Natural Language Processing (NLP)
Moreover, one of the fundamental challenges in NLP is the inherent ambiguity and complexity of
natural language, which often contains nuances, idiomatic expressions, and context-dependent
meanings. NLP algorithms must contend with these intricacies to accurately interpret and generate
human language. Techniques such as syntactic and semantic analysis, part-of-speech tagging, named
entity recognition, and sentiment analysis are employed to extract meaningful information from text
data [35]. Despite significant advancements, NLP still faces several challenges, including the need for
improved language understanding and context awareness, handling of domain-specific language and
jargon, and ensuring fairness and transparency in language processing algorithms. Additionally,
multilingual NLP and low-resource language processing present unique challenges, particularly in
regions with limited linguistic resources.
Kang [36] expounded upon the array of toolkits accessible and the procedural methodologies
requisite for the utilization of NLP as an analytical tool, delineating both its merits and demerits. In
elucidating these aspects, emphasis was placed on elucidating the managerial and technological
impediments inherent in integrating NLP within the ambit of management research, thereby furnishing
a framework to orient future investigations in this domain.
However, A recent study [37] introduced PyThaiNLP, an open-source natural language processing
(NLP) library tailored for the Thai language and implemented in Python. This comprehensive resource
encompasses an extensive array of software components, models, and datasets specifically designed for
Thai linguistic analysis. The study commenced by furnishing a succinct historical overview of preceding
tools developed for the Thai language preceding the advent of PyThaiNLP. Subsequently, an exposition
was provided regarding the manifold functionalities afforded by PyThaiNLP, including detailed
descriptions of the datasets and pre-trained language models it offered. Furthermore, the researchers
delineated the significant developmental milestones achieved throughout the evolution of PyThaiNLP,
while also offering insights into their experiential journey during its conceptualization and realization.
Recent advancements in natural language processing (NLP) have unveiled promising outcomes by
amplifying model parameters and augmenting training datasets. Nonetheless, the sole reliance on
scalability to enhance performance necessitates a commensurate increase in resource consumption,
encompassing data, time, storage, and energy, all of which are intrinsically finite and disparately
accessible [38]. This impetus has instigated a surge of research endeavors directed towards devising
efficacious methodologies that demand fewer resources while maintaining comparable performance
Khaleel et al.
IJEES
Page | 8
levels. In response to this exigency, this survey endeavors to amalgamate and elucidate current
methodologies and discoveries in the realm of efficient NLP. The principal objective of this study is
twofold: firstly, to furnish practical guidance for conducting NLP tasks under resource constraints, and
secondly, to delineate prospective avenues for pioneering more resource-efficient methodologies.
D. Computer Vision
Computer Vision is a field of artificial intelligence that empowers machines to interpret and analyze
visual data obtained from the real world. This encompasses a range of techniques including Image
Classification, Object Detection, and Semantic Segmentation. Image Classification involves the
categorization of images into predefined classes or categories based on their visual content [39]. Object
Detection, on the other hand, entails the identification and localization of objects within images or
videos. This involves not only recognizing the presence of objects but also determining their precise
locations within the visual scene [40].
Object detection algorithms enable machines to detect and locate multiple objects of interest within
complex visual environments. Semantic Segmentation involves the assignment of semantic labels to
each pixel in an image, effectively partitioning the image into regions corresponding to different objects
or areas of interest. This fine-grained analysis allows machines to understand the spatial layout and
semantic content of images in detail, facilitating tasks such as scene understanding and image
understanding [41].
E. Knowledge Representation and Reasoning
Knowledge Representation and Reasoning encompasses methodologies aimed at structuring
knowledge in a manner comprehensible to machines, enabling them to deduce logical conclusions. This
domain encompasses techniques such as Knowledge Graphs and Rule-based Systems [42]. Moreover,
Knowledge Graphs serve as graph-based structures designed to encapsulate relationships between
entities and concepts. By organizing information in a network of interconnected nodes and edges,
knowledge graphs facilitate the representation of complex relationships and dependencies within a
knowledge domain. This structured representation enables machines to traverse and manipulate the
graph to extract meaningful insights and infer new knowledge [43,44].
Besides, rule-based Systems rely on logical rules to infer new knowledge from existing knowledge
bases. These systems utilize a set of predefined rules encoded in a formal logical language, allowing
machines to perform deductive reasoning and draw conclusions based on the given premises. Rule-
based reasoning enables machines to make informed decisions and derive logical implications from the
available knowledge. Together, these techniques provide a framework for encoding and processing
knowledge in a manner conducive to automated reasoning and decision-making. By leveraging
structured representations and logical inference mechanisms, knowledge representation and reasoning
empower machines to emulate human-like reasoning capabilities, facilitating the development of
intelligent systems capable of understanding and acting upon complex information [45,46].
F. Recommender Systems
Recommender systems have garnered substantial scholarly attention since the inception of the
inaugural paper on collaborative filtering in the mid-1990s. Despite a notable surge in academic inquiry
into recommender systems over the past decade [47]. In this context, recommender systems encompass
a suite of AI techniques aimed at providing personalized recommendations to users based on their
preferences and historical interactions. Figure 4 presents recommender systems.
Figure 4. Recommender systems [48].
Recommender
systems
Content-based
filtering technique Collaborative
filtering technique Hybrid filtering
technique
Khaleel et al.
IJEES
Page | 9
This domain encompasses methods such as Collaborative Filtering, Content-based Filtering, and
Hybrid Approaches. Collaborative Filtering involves analyzing user-item interaction data to identify
patterns and similarities among users or items. By leveraging the preferences and behaviors of similar
users or items, collaborative filtering algorithms generate recommendations for users based on the
preferences of other users with similar tastes. Content-based Filtering, on the other hand, relies on the
intrinsic characteristics of items to generate recommendations. This approach analyzes the attributes or
features of items and matches them with the user's preferences or past interactions [49,50]. Content-
based filtering algorithms recommend items that are similar to those that the user has liked or interacted
with in the past. Hybrid Approaches combine collaborative filtering and content-based filtering
techniques to leverage the strengths of both methods. By integrating user preferences, item attributes,
and historical interaction data, hybrid recommender systems can provide more accurate and diverse
recommendations, enhancing the overall user experience.
This article [51] introduces the Eshop recommender, a hybrid recommendation system designed for
e-commerce applications. The Eshop recommender integrates a recommender module comprising three
subsystems, leveraging collaborative-filtering and content-based methodologies, alongside a fuzzy
expert system. This hybrid architecture aims to offer personalized product recommendations tailored
to individual user preferences and browsing behavior within the e-commerce platform. The
recommendation process involves the utilization of the fuzzy expert system, which operates on various
parameters such as similarity metrics with previously rated items, purchase history coefficients, and
product rating averages.
Gasparic [52] notes that numerous software engineering methodologies exist to facilitate the creation
of high-caliber software, yet the investment of effort and resources necessary for mastering and
implementing these methodologies can be substantial. In response to these challenges, the software
engineering community endeavors to develop supportive tools to aid practitioners in their endeavors.
Among these tools are those designed to recommend optimal solutions tailored to the specific
requirements of the user. This approach bears resemblance to the methodologies employed by search
engines and e-commerce recommendation systems. However, it is noteworthy that recommendation
systems have only recently been adapted for application within the realm of software engineering.
G. Optimization Techniques
Optimization Techniques encompass a spectrum of algorithms devised for enhancing the efficiency
and performance of intricate systems and processes. This domain encompasses methodologies such as
Evolutionary Algorithms, Gradient Descent, and Stochastic Optimization Methods. Evolutionary
Algorithms draw inspiration from principles observed in biological evolution to iteratively improve
solutions to optimization problems [53,54]. By simulating evolutionary processes such as selection,
crossover, and mutation, evolutionary algorithms explore the solution space and converge towards
optimal or near-optimal solutions. Gradient Descent is a ubiquitous optimization technique employed
in machine learning and mathematical optimization. Table 1 discusses the summarized recent
optimization techniques. It involves iteratively adjusting model parameters in the direction of the
steepest descent of a cost or objective function.
Table1: The summarized recent optimization techniques.
Year
Optimization
Techniques
Summarized
2023
SFP, ML,
KNN, NB,
LDA, LR, DT,
SVM, and RF
Software Fault Prediction (SFP) holds significant
importance in identifying faulty components within
software, thereby enabling early detection of faulty classes
or modules during the software development life cycle.
This paper introduces a machine learning framework
tailored for SFP. Initially, pre-processing and re-sampling
techniques are employed to preprocess the SFP datasets,
Khaleel et al.
IJEES
Page | 10
rendering them suitable for utilization by machine learning
(ML) techniques. Subsequently, seven classifiers are
evaluated, namely K-Nearest Neighbors (KNN), Naive
Bayes (NB), Linear Discriminant Analysis (LDA), Linear
Regression (LR), Decision Tree (DT), Support Vector
Machine (SVM), and Random Forest (RF). The findings
reveal that the Random Forest classifier surpasses all other
classifiers, particularly excelling in the elimination of
irrelevant or redundant features.
2023
PSO, ML,
SVM, NB, &
RF
Particle Swarm Optimization (PSO) is employed to
enhance the performance of machine learning (ML)
models. Our evaluation of model performance
encompasses various metrics including precision,
accuracy, recall, F-measure, performance error metrics,
and a confusion matrix. The results demonstrate that both
the standard ML models and their optimized counterparts
achieve optimal performance; nevertheless, the Support
Vector Machine (SVM) model and its optimized version
exhibit superior performance, achieving the highest
accuracy rates of 99% and 99.80%, respectively.
Furthermore, the accuracy rates of other models are as
follows: Naive Bayes (NB) - 93.90%, Optimized NB -
93.80%, Random Forest (RF) - 98.70%, Optimized RF -
99.50%, and ensemble approaches - 98.80% and 97.60%,
respectively.
2023
DL & SBT
The research findings indicate that architectures
founded upon artificial intelligence exhibit enhanced
proficiency in swiftly and accurately identifying
developers responsible for recently reported software
bugs. Notably, deep learning (DL)-based methodologies
exhibit notable advancements in the development of
software bug triaging (SBT) systems, showcasing superior
learning rates, scalability, and overall performance in
comparison to conventional approaches.
2022
Metaheuristic
Optimization
Techniques
The investigation encompassed the exploration of ten
distinct metaheuristic techniques. These methodologies
encompassed spider monkey optimization, shuffled frog
leaping algorithm, cuckoo search algorithm, ant lion
optimization technique, lion optimization technique, moth
flame technique, bat-inspired algorithm, grey wolf
algorithm, whale optimization algorithm, and dragonfly
optimization technique. These techniques were evaluated
for their efficacy in feature selection and optimization tasks
pertinent to the prediction of various medical conditions,
including heart disease, Alzheimer's disease, brain
disorders, diabetes, chronic diseases, liver disease, and
COVID-19.
5. Current Challenges and Limitations of AI in Computer Science
As artificial intelligence (AI) continues to advance rapidly, it brings forth a multitude of
opportunities to revolutionize various aspects of society and industry. However, alongside these
Khaleel et al.
IJEES
Page | 11
advancements come a host of challenges and limitations that must be addressed to realize the full
potential of AI in computer science. From ethical concerns surrounding bias and privacy to technical
hurdles related to interpretability and scalability, navigating the complexities of AI presents a
multifaceted landscape of obstacles. This introduction provides an overview of the current challenges
and limitations faced by AI in computer science, highlighting key areas of concern and the need f or
innovative solutions and interdisciplinary collaboration.
A. Ethical and Societal Concerns:
AI systems can perpetuate biases present in training data, leading to unfair outcomes and reinforcing
societal inequalities.
Privacy concerns arise from the extensive collection and use of personal data by AI algorithms,
raising questions about data security and individual autonomy.
The potential for job displacement due to automation and AI-driven technologies poses
challenges for workforce adaptation and economic stability.
B. Interpretability and Explainability:
Deep learning models, while effective, are often seen as black boxes, making it difficult to
understand how they arrive at their decisions.
Lack of interpretability and explainability in AI systems can hinder trust, accountability, and
regulatory compliance, particularly in high-stakes applications such as healthcare and finance.
C. Data Limitations and Quality:
AI algorithms require large volumes of high-quality data to train effectively, posing challenges
in domains where data collection is costly, limited, or biased.
Noisy or incomplete data can adversely affect the performance and generalization of AI models,
leading to suboptimal outcomes and unreliable predictions.
D. Safety and Security Risks:
AI systems are susceptible to adversarial attacks, where malicious actors manipulate inputs to
deceive or sabotage the system's performance.
Autonomous AI systems, such as self-driving cars and drones, raise concerns about safety and
security risks in real-world deployment scenarios, including potential accidents and cyber
threats.
E. Lack of Generalization and Transfer Learning:
AI algorithms often struggle to generalize knowledge learned from one domain to another,
limiting their adaptability and scalability across diverse tasks and environments.
Transfer learning, which aims to leverage knowledge from related domains to improve
performance, remains a challenging research area with limited practical implementations.
F. Resource Constraints and Computational Complexity:
Training and deploying sophisticated AI models require substantial computational resources,
including high-performance hardware and energy consumption.
Addressing scalability and efficiency concerns is crucial for democratizing AI access and
mitigating environmental impacts associated with large-scale computation.
G. Human-AI Collaboration and Interface Design:
Designing effective human-AI interfaces and facilitating seamless collaboration between
humans and machines remain key challenges.
AI systems often lack contextual understanding and nuanced communication skills, leading to
frustration and inefficiencies in human-AI interactions.
Navigating these challenges requires interdisciplinary collaboration, ethical considerations, and
continuous innovation in AI research and development. Addressing these limitations will be
crucial for realizing the full potential of AI in enhancing human productivity, advancing
scientific knowledge, and addressing societal challenges in the years to come.
Khaleel et al.
IJEES
Page | 12
In conclusion, the current challenges and limitations of AI in Computer Science highlight several
critical areas that demand attention and innovative solutions. Ethical and societal concerns, including
bias perpetuation and privacy issues, underscore the importance of responsible AI development and
regulation. Interpretability and explainability challenges raise questions about trust, accountability, and
regulatory compliance, particularly in high-stakes applications. Data limitations and quality issues
underscore the necessity for robust data collection practices and data preprocessing techniques to
ensure the reliability and generalization of AI models. Safety and security risks emphasize the need for
robust defenses against adversarial attacks and careful consideration of deployment scenarios for
autonomous AI systems. Additionally, the lack of generalization and transfer learning capabilities,
resource constraints, and interface design challenges necessitate interdisciplinary collaboration and
continuous innovation in AI research and development. Addressing these limitations is essential for
unlocking the full potential of AI in enhancing human productivity, advancing scientific knowledge,
and addressing societal challenges in the future.
6. Result and Discussion
In this section, the interaction between artificial intelligence (AI) and broader computer science is
profound and multifaceted, with AI both benefiting from and contributing to various subfields within
computer science. This interaction fosters cross-pollination of ideas, techniques, and methodologies,
driving innovation and advancement across the discipline. Several key aspects characterize the
interaction between AI and broader computer science:
H. Algorithm Development
Algorithm development plays a pivotal role in the intersection between AI and computer science.
AI heavily relies on algorithmic techniques borrowed from computer science disciplines such as
optimization algorithms, data structures, and computational complexity theory. These algorithmic
foundations provide the framework for AI systems to process, analyze, and derive insights from data,
ultimately enabling intelligent decision-making and problem-solving capabilities. In this context,
optimization algorithms, for instance, are fundamental to AI applications such as machine learning,
where models are trained to optimize objective functions by adjusting parameters iteratively.
Techniques like gradient descent, genetic algorithms, and simulated annealing are widely used to
optimize the performance of AI models and improve their predictive accuracy.
Conversely, advancements in AI often stimulate the development of novel algorithms and data
structures that transcend traditional AI domains. For instance, the rise of deep learning has spurred
innovations in neural network architectures, optimization techniques, and parallel computing
algorithms. These advancements have found applications not only in AI tasks such as image recognition
and natural language processing but also in diverse domains including healthcare, finance, and robotics.
In essence, algorithm development serves as the linchpin of the symbiotic relationship between AI and
computer science. As AI continues to evolve, algorithmic innovations will remain essential for pushing
the boundaries of what AI can achieve and driving interdisciplinary collaboration across computer
science disciplines.
I. Data Processing and Management
Data processing and management play crucial roles in the realm of AI, serving as the backbone for
handling the vast volumes of data required for training, validating, and inferring from AI models. As
AI systems rely heavily on data, they inherently drive research and development efforts in data
processing and management, thereby shaping advancements in fields such as databases, distributed
computing, and data mining. Conversely, innovations in data processing techniques contribute to
enhancing the scalability, efficiency, and effectiveness of AI algorithms. Moreover, AI systems require
access to large and diverse datasets to learn patterns, extract insights, and make informed decisions.
Consequently, the field of data processing encompasses techniques and methodologies for collecting,
storing, cleaning, and preprocessing data to make it suitable for AI tasks. This includes the design and
optimization of databases capable of efficiently storing and retrieving massive datasets, as well as data
Khaleel et al.
IJEES
Page | 13
preprocessing techniques such as normalization, feature engineering, and outlier detection.
Furthermore, the scale and complexity of AI applications often necessitate distributed computing
frameworks capable of processing data in parallel across multiple nodes or clusters. Distributed
computing techniques enable AI systems to harness the computational power of distributed
architectures, facilitating the training of large-scale AI models and the processing of real-time data
streams. Technologies such as Apache Hadoop, Spark, and TensorFlow distributed processing enable
scalable and fault-tolerant data processing for AI applications.
In addition, data mining techniques play a crucial role in uncovering meaningful patterns,
correlations, and insights from large datasets. Data mining algorithms, including clustering,
classification, regression, and association rule mining, enable AI systems to extract valuable knowledge
from raw data, thereby informing decision-making and driving predictive analytics. Conversely,
advancements in data processing techniques contribute to the scalability and efficiency of AI
algorithms. Innovations in database systems, such as distributed databases, columnar storage, and in-
memory databases, provide the infrastructure for storing and accessing large-scale datasets with low
latency and high throughput. Moreover, improvements in parallel and distributed computing
architectures enhance the speed and scalability of AI training and inference processes, enabling the
development of more complex and sophisticated AI models. In summary, the symbiotic relationship
between AI and data processing and management is evident, with each driving advancements in the
other. As AI continues to evolve and proliferate across various domains, further innovations in data
processing techniques will be essential for supporting the scalability, efficiency, and effectiveness of AI
systems in tackling real-world challenges and opportunities.
J. Software Engineering
AI systems are complex software systems that require robust engineering practices for
development, deployment, and maintenance. Software engineering principles and methodologies play
a crucial role in ensuring the reliability, maintainability, and scalability of AI solutions. In this direction,
software engineering is integral to the development, deployment, and maintenance of artificial AI
systems, which are inherently complex software systems. Robust software engineering practices are
essential for ensuring the reliability, maintainability, and scalability of AI solutions, ultimately
contributing to their effectiveness and sustainability. Thus, AI systems encompass a wide range of
components, including data preprocessing pipelines, machine learning models, decision-making
algorithms, and user interfaces. Developing and integrating these components into cohesive, functional
systems requires adherence to established software engineering principles and methodologies.
First and foremost, software engineering principles guide the design and architecture of AI systems,
ensuring that they are modular, scalable, and maintainable. Well-defined architectures facilitate
component reuse, code modularity, and separation of concerns, enabling developers to manage
complexity effectively and accommodate evolving requirements over time. Moreover, rigorous testing
practices are essential for verifying the correctness and robustness of AI systems. Test-driven
development, unit testing, integration testing, and end-to-end testing help identify and address
software defects, ensuring that AI solutions meet specified requirements and deliver reliable results in
diverse operational scenarios. Additionally, software engineering methodologies such as agile and
DevOps facilitate iterative development, continuous integration, and rapid deployment of AI systems.
Agile methodologies emphasize collaboration, adaptability, and customer feedback, enabling teams to
respond quickly to changing requirements and deliver incremental improvements to AI solutions.
Furthermore, version control systems and code repositories are essential tools for managing the
complexity and evolution of AI projects. Version control enables collaborative development, code
review, and rollback mechanisms, ensuring transparency, accountability, and reproducibility
throughout the software development lifecycle. Finally, considerations for deployment, monitoring,
and maintenance are paramount in ensuring the long-term success of AI systems. Deployment
pipelines, configuration management tools, and monitoring frameworks facilitate the deployment and
operation of AI solutions in production environments, while continuous monitoring and feedback
mechanisms enable proactive maintenance and optimization. In summary, software engineering plays
Khaleel et al.
IJEES
Page | 14
a critical role in the development, deployment, and maintenance of AI systems, providing the
foundation for their reliability, maintainability, and scalability. By applying sound software engineering
practices and methodologies, organizations can maximize the value and impact of AI solutions while
mitigating risks and ensuring their long-term sustainability.
K. Human-Computer Interaction (HCI)
Human-Computer Interaction (HCI) plays a crucial role in shaping the design and evaluation of
interfaces that facilitate interaction between humans and AI systems. HCI research focuses on
understanding how users interact with technology and aims to design interfaces that are intuitive,
efficient, and user-friendly. In the context of AI systems, HCI research is essential for creating interfaces
that effectively leverage AI capabilities while aligning with user needs and preferences. To elaborate,
understanding human behavior, cognitive psychology, and user experience principles is foundational
to HCI research in the context of AI. By studying how humans perceive, process, and interact with
information, HCI researchers can design interfaces that accommodate cognitive limitations, support
natural interactions, and enhance user engagement. Cognitive psychology insights inform interface
design decisions related to information presentation, task flow, and interaction design, ensuring that AI
systems are intuitive and easy to use.
Moreover, HCI research explores the user experience (UX) aspects of interacting with AI systems,
focusing on aspects such as usability, satisfaction, and engagement. By conducting user studies,
usability tests, and UX evaluations, HCI researchers can identify usability issues, gather feedback from
users, and iteratively improve AI interfaces to enhance user satisfaction and adoption. Effective HCI in
the context of AI also involves considering the unique capabilities and limitations of AI technologies.
Designing AI interfaces requires careful consideration of how AI functionalities, such as natural
language processing, computer vision, and machine learning, can be integrated into the user interface
to enhance user interactions and support user goals. HCI researchers collaborate with AI experts to
explore novel interaction paradigms, such as conversational interfaces, gesture recognition, and
personalized recommendations, that leverage AI capabilities to provide seamless and personalized user
experiences.
Furthermore, HCI research emphasizes the importance of ethical considerations and human-
centered design principles in AI interface design. Designing AI interfaces ethically involves ensuring
transparency, accountability, and fairness in AI systems, as well as respecting user privacy and
autonomy. Human-centered design approaches prioritize user needs, preferences, and values
throughout the design process, fostering empathy and inclusivity in AI interface design. In summary,
HCI research plays a vital role in shaping the design and evaluation of interfaces for AI systems,
enabling effective interaction between humans and AI technologies. By integrating insights from human
behavior, cognitive psychology, user experience principles, and AI capabilities, HCI researchers can
create intuitive, user-friendly interfaces that enhance user satisfaction, engagement, and productivity in
interacting with AI systems.
L. Security and Privacy
The integration of AI systems into various domains introduces unique security and privacy
challenges that demand specialized attention. These challenges encompass vulnerabilities to adversarial
attacks, risks of data breaches, and the potential propagation of algorithmic biases. To address these
concerns, ongoing research in cybersecurity, cryptography, and privacy-preserving technologies is
critical to fortify AI systems and safeguard sensitive data.
Adversarial Attacks
AI systems are susceptible to adversarial attacks, where malicious actors manipulate input data to
deceive or mislead the system's decision-making process. These attacks can compromise the integrity
and reliability of AI models, leading to erroneous predictions or decisions. Research in adversarial
machine learning aims to develop robust defense mechanisms against such attacks, including
adversarial training, input sanitization, and model robustification techniques.
Khaleel et al.
IJEES
Page | 15
Data Breaches
AI systems often rely on large volumes of data for training and inference, raising concerns about the
security and privacy of sensitive information. Data breaches can occur due to unauthorized access,
malicious insider threats, or inadequate data protection measures. Research in cybersecurity focuses on
enhancing data encryption, access control mechanisms, and data anonymization techniques to mitigate
the risks of data breaches and unauthorized access to sensitive data.
Algorithmic Biases
AI algorithms may exhibit biases inherent in the training data, resulting in unfair or discriminatory
outcomes, particularly in sensitive domains such as healthcare, finance, and criminal justice.
Addressing algorithmic biases requires interdisciplinary research efforts combining AI, ethics, and
social sciences. Techniques such as fairness-aware machine learning, bias detection, and mitigation
strategies aim to identify and rectify biases in AI models, ensuring equitable and unbiased decision-
making processes.
Privacy-Preserving Technologies
Protecting user privacy is paramount in AI systems that handle personal or sensitive data. Privacy-
preserving technologies, such as differential privacy, homomorphic encryption, and federated learning,
enable AI systems to perform computations on sensitive data while preserving individual privacy.
These techniques allow data to be analyzed and utilized for AI applications without compromising the
confidentiality or anonymity of individual users. By advancing research in cybersecurity, cryptography,
and privacy-preserving technologies, the AI community can mitigate security and privacy risks
associated with AI systems and foster trust among users and stakeholders. Collaboration between AI
researchers, cybersecurity experts, policymakers, and industry stakeholders are essential to develop
robust security measures and privacy-enhancing technologies that safeguard AI systems and uphold
ethical principles in AI deployment.
M. Theoretical Foundations
Theoretical foundations serve as the bedrock upon which AI research builds its methodologies,
algorithms, and models. Drawing from disciplines such as mathematics, logic, and theoretical computer
science, these foundational concepts provide the theoretical underpinnings necessary for understanding
and developing AI systems. Key theoretical topics that contribute to AI research include probability
theory, linear algebra, and formal logic.
Probability Theory
Probability theory plays a central role in AI, particularly in the field of probabilistic reasoning and
decision-making. Bayesian networks, Markov models, and probabilistic graphical models are examples
of AI techniques that rely on probabilistic principles to represent uncertainty and make predictions.
Probability theory enables AI systems to model and reason under uncertainty, facilitating tasks such as
probabilistic inference, Bayesian learning, and probabilistic reasoning.
Linear Algebra
Linear algebra is fundamental to many aspects of AI, particularly in the realm of machine learning
and data analysis. Techniques such as matrix operations, eigenvalue decomposition, and singular value
decomposition are widely used in AI algorithms for tasks such as dimensionality reduction, feature
extraction, and optimization. Linear algebra provides the mathematical framework for manipulating
and transforming data, enabling AI systems to process and analyze large-scale datasets efficiently.
Formal Logic
Formal logic provides the foundation for symbolic reasoning and knowledge representation in AI.
Propositional logic, predicate logic, and first-order logic are formal languages used to express
knowledge and infer logical conclusions in AI systems. Logical reasoning techniques enable AI systems
to represent and manipulate knowledge in a structured and systematic manner, facilitating tasks such
as theorem proving, logical inference, and automated reasoning. In addition to these core topics, AI
research also draws upon concepts from other disciplines such as calculus, graph theory, information
theory, and computational complexity theory. These theoretical foundations provide the framework for
understanding the capabilities and limitations of AI algorithms, guiding the development of novel
Khaleel et al.
IJEES
Page | 16
techniques and methodologies. By leveraging theoretical insights from mathematics, logic, and
theoretical computer science, AI researchers can develop robust and principled algorithms that exhibit
intelligent behavior and solve complex problems in diverse domains. Continual advancements in
theoretical foundations are essential for pushing the boundaries of AI research and enabling the
development of AI systems that are more capable, efficient, and reliable.
N. Applications and Domains
Artificial intelligence (AI) has permeated numerous domains, revolutionizing processes, and
offering solutions to complex challenges across various industries. From healthcare to finance,
education, and entertainment, AI applications have transformed how tasks are performed, decisions are
made, and experiences are delivered. Collaborations between AI researchers and domain experts have
been instrumental in developing tailored AI solutions to address specific challenges and opportunities
in each domain.
Engineering
In the engineering domain, AI assumes a pivotal role in optimizing manufacturing processes,
refining product design methodologies, and augmenting operational efficiency across a spectrum of
industries, including the energy sector. Within manufacturing processes, AI technologies such as
machine learning and predictive analytics are harnessed to optimize production workflows, streamline
supply chain logistics, and minimize production downtime. By analyzing vast amounts of data
generated from sensors, IoT devices, and production equipment, AI systems can forecast equipment
failures, identify inefficiencies, and optimize resource allocation, thereby enhancing overall
manufacturing productivity and reducing operational costs. Moreover, in product design, AI-driven
techniques like generative design and simulation algorithms facilitate rapid prototyping, iterative
design iterations, and optimization of complex engineering structures. These AI-powered design tools
enable engineers to explore a broader design space, uncover innovative solutions, and accelerate the
development of high-performance products while minimizing material wastage and time-to-market.
In the energy sector, AI is instrumental in optimizing various aspects of energy production,
distribution, and consumption. AI algorithms are deployed in smart grid systems to analyze energy
consumption patterns, predict demand fluctuations, and optimize energy distribution networks in real-
time. Additionally, AI-driven predictive maintenance solutions are employed in power generation
facilities to monitor equipment health, detect anomalies, and schedule maintenance activities
proactively, thereby ensuring reliable and efficient energy production. Furthermore, AI-enabled
optimization algorithms are utilized in energy management systems to optimize energy usage,
minimize wastage, and reduce costs in industrial and commercial facilities. These AI-driven solutions
leverage advanced analytics and machine learning techniques to identify energy-saving opportunities,
optimize energy storage systems, and facilitate demand response programs, thereby promoting
sustainability and cost-efficiency in the energy sector.
Medical fields
The applications and domains of Artificial Intelligence (AI) in both medical fields encompass a wide
array of innovative and transformative solutions. In the medical domain, AI is revolutionizing various
aspects of healthcare delivery, including diagnosis, treatment, personalized medicine, and patient care
management. AI-powered medical imaging technologies, such as computer-aided diagnosis (CAD),
enable more accurate and efficient interpretation of medical images, leading to early detection of
diseases such as cancer, cardiovascular disorders, and neurological conditions. Additionally, AI-driven
predictive analytics models assist healthcare providers in identifying patients at risk of developing
certain diseases, allowing for timely intervention and preventive measures.
Furthermore, AI algorithms are being employed in drug discovery and development processes,
accelerating the identification of potential therapeutic compounds and optimizing drug formulations.
In clinical decision support systems, AI assists healthcare professionals in making evidence-based
decisions by analyzing patient data, medical records, and relevant literature to provide tailored
treatment recommendations. Moreover, AI-powered virtual health assistants and telemedicine
Khaleel et al.
IJEES
Page | 17
platforms enhance access to healthcare services, enabling remote monitoring of patients,
teleconsultations, and remote diagnosis.
Technology
In the realm of technology, Artificial Intelligence (AI) indeed assumes a pivotal role, serving as a
driving force behind numerous advancements and innovations. AI's multifaceted applications
permeate various domains within technology, revolutionizing processes, enhancing capabilities, and
unlocking new possibilities. One prominent area where AI plays a pivotal role is in automation. AI-
powered automation systems streamline workflows, optimize operations, and eliminate repetitive tasks
across diverse industries. This not only increases efficiency but also reduces human error and frees up
valuable human resources to focus on more strategic and creative endeavors. In addition, AI drives
innovation in robotics and autonomous systems. From self-driving cars to robotic manufacturing lines,
AI enables machines to perceive, reason, and act in complex and dynamic environments. These
advancements have the potential to revolutionize transportation, healthcare, manufacturing, and other
industries, leading to safer, more efficient, and more sustainable processes. Furthermore, AI facilitates
the development of intelligent assistants and chatbots that interact with users in natural language,
providing personalized assistance, information retrieval, and task automation. These AI-powered
interfaces enhance user experiences and support a wide range of applications, from customer service to
virtual personal assistants.
Military affairs
In the realm of military affairs, Artificial Intelligence (AI) indeed assumes a pivotal role, exerting a
transformative influence on modern warfare and defense strategies. AI technologies are increasingly
integrated into military systems and operations, offering capabilities that enhance situational
awareness, decision-making, and operational effectiveness. One significant application of AI in the
military is in autonomous systems and unmanned vehicles. AI-powered drones, unmanned ground
vehicles, and autonomous submarines enable militaries to conduct reconnaissance, surveillance, and
combat missions in hostile environments without risking human lives. These autonomous systems can
navigate complex terrain, identify targets, and execute missions with precision and efficiency, thereby
augmenting military capabilities and reducing operational risks. Moreover, AI enables the development
of intelligent command and control systems that facilitate real-time decision-making and coordination
across military operations. AI algorithms analyze vast amounts of data from sensors, satellites, and
other sources to provide commanders with actionable insights and predictive intelligence. This
enhances situational awareness, enables rapid response to threats, and improves the overall
effectiveness of military operations.
Additionally, AI enhances cybersecurity and defense capabilities by detecting and mitigating cyber
threats in real-time. AI-driven cybersecurity systems can identify patterns of malicious activity,
anticipate cyber-attacks, and automatically respond to emerging threats, thereby safeguarding military
networks, systems, and data from cyber adversaries. These AI-powered technologies provide militaries
with a competitive edge on the battlefield, enabling them to deter adversaries, project power, and
achieve strategic objectives with greater precision and efficiency.
Economy
In the realm of economics, Artificial Intelligence (AI) assumes a pivotal role in shaping economic
outcomes and driving progress. AI technologies offer significant potential to optimize resource
allocation, enhance decision-making processes, and foster innovation across various sectors of the
economy. One key aspect of AI's impact on economics is its ability to streamline operational processes
and increase productivity. By automating repetitive tasks and leveraging data-driven insights, AI
systems can optimize production processes, reduce labor costs, and improve overall efficiency in
economic activities. This increased efficiency translates into cost savings and improved competitiveness
for businesses operating in various industries.
Education
AI technologies are transforming education through personalized learning, adaptive assessment,
and intelligent tutoring systems. AI-powered educational platforms tailor learning experiences to
Khaleel et al.
IJEES
Page | 18
individual student needs, adapting content and pacing based on student performance and preferences.
Additionally, AI-driven assessment tools provide real-time feedback and insights to educators, enabling
data-driven decision-making and instructional improvement.
Entertainment
In the entertainment industry, AI is used for content recommendation, content creation, and
immersive experiences. AI algorithms analyze user preferences and behavior to recommend movies,
music, and other media content personalized to individual tastes. Furthermore, AI-driven content
generation tools, such as natural language processing (NLP) and generative adversarial networks
(GANs), enable the creation of virtual characters, special effects, and interactive narratives.
Transportation
The transportation industry employs AI for route optimization, autonomous vehicles, traffic
management, and predictive maintenance. AI algorithms analyze traffic patterns, weather conditions,
and historical data to optimize transportation routes, reduce congestion, and improve fuel efficiency.
Additionally, AI-driven autonomous vehicles promise to revolutionize mobility, offering safer, more
efficient, and environmentally friendly transportation solutions. These examples highlight the diverse
applications and domains where AI in computer science is making a significant impact. Collaborations
between AI researchers and domain experts are essential for tailoring AI solutions to specific industry
needs, driving innovation, and unlocking new possibilities for enhancing productivity, efficiency, and
quality of life across various sectors.
7. Conclusion
In conclusion, Artificial Intelligence (AI) stands as a transformative force within the realm of
computer science, with a rich history of evolution and an ever-expanding array of applications across
various sectors. From its origins to its current state of advancement, AI has continually pushed the
boundaries of innovation and revolutionized industries worldwide. As we look toward the future, the
potential impact of AI appears boundless, promising further breakthroughs in human-computer
interaction, automation, and the realization of artificial general intelligence. However, with great power
comes great responsibility, and it is imperative that the development and deployment of AI technologies
are accompanied by ethical considerations and regulatory frameworks to ensure their responsible and
beneficial integration into society. By continuing to explore the potential of AI while prioritizing ethical
principles and societal well-being, we can harness its transformative potential to shape a brighter and
more equitable future for all.
Author Contributions: Conceptualization, M. K.; methodology, M. K. & A. J.; validation, M. K. & A. J.; formal
analysis, M. K. & A. J.; investigation M. K.; resources, author; data curation, author; writingoriginal draft
preparation, M. K. & A. J.; writingreview and editing, author; visualization, author; supervision, author; project
administration, M. K. & A. J.; author has read and agreed to the published version of the manuscript.
Funding: This article received no external funding.
Data Availability Statement: Not applicable.
Acknowledgments: The author(s) would like to extend their sincere gratitude to the Department of Research and
Development, College of Civil Aviation, Misrata, Libya, for their support and guidance throughout the research
process.
Conflicts of Interest: The author(s) declare no conflict of interest.
ORCID
Mohamed Khaleel https://orcid.org/0000-0002-3468-3220
Abdullatif Jebrel https://orcid.org/0009-0006-9151-3435
Khaleel et al.
IJEES
Page | 19
References
[1]
L. Messeri and M. J. Crockett, “Artificial intelligence and illusions of understanding in scientific
research,” Nature, vol. 627, no. 8002, pp. 4958, 2024. [Google Scholar]
[2]
P. Giudici, M. Centurelli, and S. Turchetta, “Artificial Intelligence risk measurement,” Expert Syst. Appl., vol.
235, no. 121220, p. 121220, 2024. [Google Scholar]
[3]
M. Khaleel, Y. Nassar, and H. J. El-Khozondar, “Towards utilizing Artificial Intelligence in scientific
writing,” Int. J. Electr. Eng. and Sustain., pp. 4550, 2024. [Google Scholar]
[4]
A. Sinha, D. Sapra, D. Sinwar, V. Singh, and G. Raghuwanshi, “Assessing and mitigating bias in Artificial
Intelligence: A review,” Recent Advances in Computer Science and Communications, vol. 17, no. 1, pp. 110,
2024. [Google Scholar]
[5]
M. Khaleel, A. A. Ahmed, and A. Alsharif, “Artificial Intelligence in Engineering,” Brilliance, vol. 3, no. 1,
pp. 3242, 2023. [Google Scholar]
[6]
S. Nyholm, “Artificial intelligence and human enhancement: Can AI technologies make us more (artificially)
intelligent?,” Camb. Q. Healthc. Ethics, vol. 33, no. 1, pp. 7688, 2024. [Google Scholar]
[7]
J. Shuford and M. M. Islam, “Exploring the latest trends in artificial intelligence technology: A
comprehensive review,” Journal of Artificial Intelligence General science (JAIGS), vol. 2, no. 1, 2024. [Google
Scholar]
[8]
M. Khaleel, “Intelligent Control Techniques for Microgrid Systems,” Brilliance, vol. 3, no. 1, pp. 5667, 2023.
[9]
Z. Li, Z. A. Pardos, and C. Rena, “Aligning open educational resources to new taxonomies: How AI
technologies can help and in which scenarios,” Comput. Educ., no. 105027, p. 105027, 2024. [Google Scholar]
[10]
M. Khaleel, S. A. Abulifa, and A. A. Abulifa, “Artificial intelligent techniques for identifying the cause of
disturbances in the power grid,” Brilliance, vol. 3, no. 1, pp. 1931, 2023. [Google Scholar]
[11]
B. Liu, L. Yu, C. Che, Q. Lin, H. Hu, and X. Zhao, “Integration and performance analysis of artificial
intelligence and computer vision based on deep learning algorithms,” 2023. [Google Scholar]
[12]
N. R. Mannuru et al., “Artificial intelligence in developing countries: The impact of generative artificial
intelligence (AI) technologies for development,” Inf. Dev., 2023. [Google Scholar]
[13]
G. Rjoub et al., “A survey on explainable Artificial Intelligence for cybersecurity,” 2023. [Google Scholar]
[14]
M. Kandlhofer, G. Steinbauer, S. Hirschmugl-Gaisch, and P. Huber, “Artificial intelligence and computer
science in education: From kindergarten to university,” in 2016 IEEE Frontiers in Education Conference (FIE),
2016. [Google Scholar]
[15]
E. Y. Zhang, A. D. Cheok, Z. Pan, J. Cai, and Y. Yan, “From Turing to transformers: A comprehensive review
and tutorial on the evolution and applications of generative transformer models,” Sci, vol. 5, no. 4, p. 46,
2023. [Google Scholar]
[16]
Y. Cao, S. Tang, R. Yao, L. Chang, and X. Yin, “Interpretable hierarchical belief rule base expert system for
complex system modeling,” Measurement (Lond.), vol. 226, no. 114033, p. 114033, 2024. [Google Scholar]
[17]
D. D. Cox and T. Dean, “Neural networks and neuroscience-inspired computer vision,” Curr. Biol., vol. 24,
no. 18, pp. R921R929, 2014. [Google Scholar]
[18]
S. Sharma and P. Chaudhary, “Chapter 4 Machine learning and deep learning,” in Quantum Computing and
Artificial Intelligence, De Gruyter, 2023, pp. 7184. [Google Scholar]
[19]
V. Galanos, “Expectations and expertise in artificial intelligence: specialist views and historical perspectives
on conceptualisation, promise, and funding.” The University of Edinburgh, 2023. [Google Scholar]
[20]
H. Hirsch-Kreinsen, “Artificial intelligence: a ‘promising technology,’” AI Soc., 2023. [Google Scholar]
[21]
S. Lins, K. D. Pandl, H. Teigeler, S. Thiebes, C. Bayer, and A. Sunyaev, “Artificial intelligence as a service:
Classification and research directions,” Bus. Inf. Syst. Eng., vol. 63, no. 4, pp. 441456, 2021. [Google Scholar]
[22]
B. Jena, S. Saxena, G. K. Nayak, L. Saba, N. Sharma, and J. S. Suri, “Artificial intelligence-based hybrid deep
learning models for image classification: The first narrative review,” Comput. Biol. Med., vol. 137, no. 104803,
p. 104803, 2021. [Google Scholar]
[23]
R. Rojas, “The first code for computer chess,” in Konrad Zuse’s Early Computers, Cham: Springer Nature
Switzerland, 2023, pp. 191201. [Google Scholar]
[24]
B. Goertzel, “Artificial General Intelligence: Concept, state of the art, and future prospects,” J. Artif. Gen.
Intell., vol. 5, no. 1, pp. 148, 2014. [Google Scholar]
[25]
A. M. Barrett and S. D. Baum, “A model of pathways to artificial superintelligence catastrophe for risk and
decision analysis,” J. Exp. Theor. Artif. Intell., vol. 29, no. 2, pp. 397414, 2017. [Google Scholar]
Khaleel et al.
IJEES
Page | 20
[26]
J. Bell, “What is machine learning?,” Machine Learning and the City. Wiley, pp. 207216, 21-May-2022. [Google
Scholar]
[27]
Z. Ahmed et al., “Machine learning at Microsoft with ML.NET,” in Proceedings of the 25th ACM SIGKDD
International Conference on Knowledge Discovery & Data Mining, 2019. [Google Scholar]
[28]
D. Kreuzberger, N. Kühl, and S. Hirschl, “Machine learning operations (MLOps): Overview, definition, and
architecture,” IEEE Access, vol. 11, pp. 3186631879, 2023. [Google Scholar]
[29]
S. Secinaro, D. Calandra, A. Secinaro, V. Muthurangu, and P. Biancone, “The role of artificial intelligence in
healthcare: a structured literature review,” BMC Med. Inform. Decis. Mak., vol. 21, no. 1, 2021. [Google
Scholar]
[30]
Y. Guo, Y. Liu, A. Oerlemans, S. Lao, S. Wu, and M. S. Lew, “Deep learning for visual understanding: A
review,” Neurocomputing, vol. 187, pp. 2748, 2016. [Google Scholar]
[31]
C. Janiesch, P. Zschech, and K. Heinrich, “Machine learning and deep learning,” Electron. Mark., vol. 31, no.
3, pp. 685695, 2021. [Google Scholar]
[32]
I. H. Sarker, “Deep learning: A comprehensive overview on techniques, taxonomy, applications and
research directions,” SN Comput. Sci., vol. 2, no. 6, 2021. [Google Scholar]
[33]
Y. LeCun, Y. Bengio, and G. Hinton, “Deep learning,” Nature, vol. 521, no. 7553, pp. 436444, 2015. [Google
Scholar]
[34]
D. Khurana, A. Koli, K. Khatter, and S. Singh, “Natural language processing: state of the art, current trends
and challenges,” Multimed. Tools Appl., vol. 82, no. 3, pp. 37133744, 2023. [Google Scholar]
[35]
S. Chung, S. Moon, J. Kim, J. Kim, S. Lim, and S. Chi, “Comparing natural language processing (NLP)
applications in construction and computer science using preferred reporting items for systematic reviews
(PRISMA),” Autom. Constr., vol. 154, no. 105020, p. 105020, 2023. [Google Scholar]
[36]
Y. Kang, Z. Cai, C.-W. Tan, Q. Huang, and H. Liu, “Natural language processing (NLP) in management
research: A literature review,” J. Manag. Anal., vol. 7, no. 2, pp. 139172, 2020. [Google Scholar]
[37]
W. Phatthiyaphaibun et al., “PyThaiNLP: Thai natural language processing in Python,” arXiv [cs.CL], 2023.
[38]
M. Treviso et al., “Efficient methods for natural language processing: A survey,” Trans. Assoc. Comput.
Linguist., vol. 11, pp. 826860, 2023. [Google Scholar]
[39]
T. Ige, A. Kolade, and O. Kolade, “Enhancing border security and countering terrorism through computer
vision: A field of artificial intelligence,” arXiv [cs.CV], 2023. [Google Scholar]
[40]
L. Kaiyue, L. Lei “Research and application of artificial intelligence in the field of vision system and
network,” in 2019 5th International Conference on Advanced Computing, Networking and Security (ADCONS
2019), 2019. [Google Scholar]
[41]
P. Abichandani, C. Iaboni, D. Lobo, and T. Kelly, “Artificial intelligence and computer vision education:
Codifying student learning gains and attitudes,” Computers and Education: Artificial Intelligence, vol. 5, no.
100159, p. 100159, 2023. [Google Scholar]
[42]
G.-J. Hwang, H. Xie, B. W. Wah, and D. Gašević, “Vision, challenges, roles and research issues of Artificial
Intelligence in Education,” Computers and Education: Artificial Intelligence, vol. 1, no. 100001, p. 100001, 2020.
[Google Scholar]
[43]
S. B. Rajasekaran and NVIDIA, “Literature review: Recent advances in computer vision and language AI,” J
Arti Inte & Cloud Comp, pp. 13, 2023. [Google Scholar]
[44]
Y. Pan, J. Liu, L. Zhang, and Y. Huang, “Incorporating logic rules with textual representations for
interpretable knowledge graph reasoning,” Knowl. Based Syst., vol. 277, no. 110787, p. 110787, 2023. [Google
Scholar]
[45]
Z. Jiang, “Representation and reasoning of empirical engineering knowledge,” Int. J. Knowl. Eng., vol. 5, no.
2, pp. 5356, 2019. [Google Scholar]
[46]
S. Wang and D. Liu, “Knowledge representation and reasoning for qualitative spatial change,” Knowl. Based
Syst., vol. 30, pp. 161171, 2012. [Google Scholar]
[47]
L. Lü, M. Medo, C. H. Yeung, Y.-C. Zhang, Z.-K. Zhang, and T. Zhou, “Recommender systems,” Phys. Rep.,
vol. 519, no. 1, pp. 149, 2012. [Google Scholar]
[48]
F. O. Isinkaye, Y. O. Folajimi, and B. A. Ojokoh, “Recommendation systems: Principles, methods and
evaluation,” Egypt. Inform. J., vol. 16, no. 3, pp. 261273, 2015. [Google Scholar]
[49]
A. Gunawardana, G. Shani, and S. Yogev, “Evaluating Recommender Systems,” in Recommender Systems
Handbook, New York, NY: Springer US, 2022, pp. 547601. [Google Scholar]
Khaleel et al.
IJEES
Page | 21
[50]
C. Gao, Y. Zheng, W. Wang, F. Feng, X. He, and Y. Li, “Causal inference in recommender systems: A survey
and future directions,” ACM Trans. Inf. Syst., vol. 42, no. 4, pp. 132, 2024. [Google Scholar]
[51]
D. H. Park, H. K. Kim, I. Y. Choi, and J. K. Kim, “A literature review and classification of recommender
systems research,” Expert Syst. Appl., vol. 39, no. 11, pp. 1005910072, 2012. [Google Scholar]
[52]
M. Gasparic and A. Janes, “What recommendation systems for software engineering recommend: A
systematic literature review,” J. Syst. Softw., vol. 113, pp. 101113, 2016. [Google Scholar]
[53]
M. Khaleel, Z. Yusupov, Y. Nassar, H. J. El-khozondar, A. Ahmed, and A. Alsharif, “Technical challenges
and optimization of superconducting magnetic energy storage in electrical power systems,” e-Prime -
Advances in Electrical Engineering, Electronics and Energy, vol. 5, no. 100223, p. 100223, 2023. [Google Scholar]
[54]
M. Khaleel, E. Yaghoubi, E. Yaghoubi, and M. Z. Jahromi, “The role of mechanical energy storage systems
based on artificial intelligence techniques in future sustainable energy systems,” Int. J. Electr. Eng. and
Sustain., pp. 0131, 2023. [Google Scholar]
[55]
M. Mafarja et al., “Classification framework for faulty-software using enhanced exploratory whale
optimizer-based feature selection scheme and random forest ensemble learning,” Appl. Intell., vol. 53, no. 15,
pp. 1871518757, 2023. [Google Scholar]
[56]
A. Khalid, G. Badshah, N. Ayub, M. Shiraz, and M. Ghouse, “Software defect prediction analysis using
machine learning techniques,” Sustainability, vol. 15, no. 6, p. 5517, 2023. [Google Scholar]
[57]
N. K. Nagwani and J. S. Suri, “An artificial intelligence framework on software bug triaging, technological
evolution, and future challenges: A review,” International Journal of Information Management Data Insights,
vol. 3, no. 1, p. 100153, 2023. [Google Scholar]
[58]
S. Kaur, Y. Kumar, A. Koul, and S. Kumar Kamboj, “A systematic review on metaheuristic optimization
techniques for feature selections in disease diagnosis: Open issues and challenges,” Arch. Comput. Methods
Eng., vol. 30, no. 3, pp. 18631895, 2023. [Google Scholar]
Open Access This article is licensed under a Creative Commons Attribution 4.0 International
License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long
as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons
licence, and indicate if changes were made. The images or other third-party material in this article are included in
the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not
included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation
or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy
of this licence, visit http://creativecommons.org/licenses/by/4.0/.
© The Author(s) 2024
ResearchGate has not been able to resolve any citations for this publication.
Article
Full-text available
This paper focuses on the analysis of the application effectiveness of the integration of deep learning and computer vision technologies. Deep learning achieves a historic breakthrough by constructing hierarchical neural networks, enabling end-to-end feature learning and semantic understanding of images. The successful experiences in the field of computer vision provide strong support for training deep learning algorithms. The tight integration of these two fields has given rise to a new generation of advanced computer vision systems, significantly surpassing traditional methods in tasks such as machine vision image classification and object detection. In this paper, typical image classification cases are combined to analyze the superior performance of deep neural network models while also pointing out their limitations in generalization and interpretability, proposing directions for future improvements. Overall, the efficient integration and development trend of deep learning with massive visual data will continue to drive technological breakthroughs and application expansion in the field of computer vision, making it possible to build truly intelligent machine vision systems. This deepening fusion paradigm will powerfully promote unprecedented tasks and functions in computer vision, providing stronger development momentum for related disciplines and industries.
Article
Full-text available
Artificial intelligence (AI) has become increasingly pervasive across various domains, including smartphones, social media platforms, search engines, and autonomous vehicles, among others. This study undertakes a scoping review of the current landscape of AI technologies, following the PRISMA framework, with the aim of identifying the most advanced technologies utilized in different domains of AI research. Three reputable journals within the artificial intelligence and machine learning domain, namely the Journal of Artificial Intelligence Research, the Journal of Machine Learning Research, and Machine Learning, were selected for this review. Articles published in 2022 were scrutinized against certain criteria: the technology must be tested against comparable solutions, employ commonly approved or well-justified datasets, and demonstrate improvements over comparable solutions. A crucial aspect of technology development identified in this review is the processing and exploitation of data collected from diverse sources. Given the highly unstructured nature of data, technological solutions should minimize the need for manual intervention by humans. The review indicates that creating labeled datasets is a labor-intensive process, leading to increased research focus on solutions leveraging unsupervised or semi-supervised learning technologies. Efficient updating of learning algorithms and the interpretability of predictions emerge as key considerations in the development of AI technologies. Moreover, in real-world applications, ensuring safety and providing explainable predictions are imperative before widespread adoption can be achieved. Thus, this review underscores the importance of addressing these factors to facilitate the responsible and effective integration of AI technologies into various domains.
Article
Full-text available
The integration of Artificial Intelligence (AI) tools in scientific writing processes has emerged as a topic of increasing interest and exploration. This article encapsulates the multifaceted dimensions associated with the utilization of AI in scientific writing endeavors. The contemporary landscape underscores the potential of AI systems, such as ChatGPT, in facilitating various stages of scientific writing, including literature review, draft generation, language refinement, and formatting. While AI-powered tools offer unparalleled efficiency and efficacy in streamlining writing tasks, ethical considerations regarding plagiarism detection, bias mitigation, and the preservation of academic integrity necessitate careful deliberation and regulatory frameworks. Furthermore, the synergistic collaboration between human expertise and AI assistance is advocated to optimize the quality and authenticity of scientific publications. This article provides a foundational overview of the evolving role of AI in scientific writing, inviting further research and discourse to harness its transformative potential while upholding scholarly standards and ethical principles.
Article
Full-text available
In recent years, generative transformers have become increasingly prevalent in the field of artificial intelligence, especially within the scope of natural language processing. This paper provides a comprehensive overview of these models, beginning with the foundational theories introduced by Alan Turing and extending to contemporary generative transformer architectures. The manuscript serves as a review, historical account, and tutorial, aiming to offer a thorough understanding of the models’ importance, underlying principles, and wide-ranging applications. The tutorial section includes a practical guide for constructing a basic generative transformer model. Additionally, the paper addresses the challenges, ethical implications, and future directions in the study of generative models.
Conference Paper
Full-text available
We present PyThaiNLP, a free and open-source natural language processing (NLP) library for Thai language implemented in Python. It provides a wide range of software, models, and datasets for Thai language. We first provide a brief historical context of tools for Thai language prior to the development of PyThaiNLP. We then outline the functionalities it provided as well as datasets and pre-trained language models. We later summarize its development milestones and discuss our experience during its development. We conclude by demonstrating how industrial and research communities utilize PyThaiNLP in their work. The library is freely available at https://github.com/pythainlp/pythainlp.
Article
Scientists are enthusiastically imagining ways in which artificial intelligence (AI) tools might improve research. Why are AI tools so attractive and what are the risks of implementing them across the research pipeline? Here we develop a taxonomy of scientists' visions for AI, observing that their appeal comes from promises to improve productivity and objectivity by overcoming human shortcomings. But proposed AI solutions can also exploit our cognitive limitations, making us vulnerable to illusions of understanding in which we believe we understand more about the world than we actually do. Such illusions obscure the scientific community's ability to see the formation of scientific monocultures, in which some types of methods, questions and viewpoints come to dominate alternative approaches, making science less innovative and more vulnerable to errors. The proliferation of AI tools in science risks introducing a phase of scientific enquiry in which we produce more but understand less. By analysing the appeal of these tools, we provide a framework for advancing discussions of responsible knowledge production in the age of AI.
Article
Recommender systems have become crucial in information filtering nowadays. Existing recommender systems extract user preferences based on the correlation in data, such as behavioral correlation in collaborative filtering, feature-feature, or feature-behavior correlation in click-through rate prediction. However, unfortunately, the real world is driven by causality , not just correlation, and correlation does not imply causation. For instance, recommender systems might recommend a battery charger to a user after buying a phone, where the latter can serve as the cause of the former; such a causal relation cannot be reversed. Recently, to address this, researchers in recommender systems have begun utilizing causal inference to extract causality, thereby enhancing the recommender system. In this survey, we offer a comprehensive review of the literature on causal inference-based recommendation. Initially, we introduce the fundamental concepts of both recommender system and causal inference as the foundation for subsequent content. We then highlight the typical issues faced by non-causality recommender system. Following that, we thoroughly review the existing work on causal inference-based recommender systems, based on a taxonomy of three-aspect challenges that causal inference can address. Finally, we discuss the open problems in this critical research area and suggest important potential future works.
Article
This comprehensive literature review examines the latest breakthroughs in computer vision and natural language processing (NLP), two rapidly evolving fields with applications across search, human-computer interaction, robotics, and more. It synthesizes key findings, trends, limitations, and open challenges from cutting-edge research at their intersection. The dramatic progress driven by deep neural networks is analysed in depth, along with issues like generalization, context handling, reasoning, uncertainty, and human-centric evaluation. Although remarkable advances have been made, especially in computer vision, core problems remain to be addressed. This review provides a thorough overview of the state-of-the-art, reflecting the most recent innovations, and promising future directions in this dynamic research domain.