15067
15067
Abstract:
This article provides an in-depth analysis of the use of Artificial Intelligence (AI) in various aspects of
biology, including healthcare, agriculture, and environmental monitoring. It highlights AI's ability to
mimic human intelligence and analyze large datasets for predictions and tasks. The article also discusses
its integration into Chinese medicine, where AI-guided diagnostic and therapeutic systems optimize
clinical treatments and health management. AI is also used in disease management, analyzing data on
diseases and pests, predicting their impact on ecosystems, and implementing preventative measures. The
article also highlights the role of integrated information systems in environmental monitoring.
Artificial intelligence (AI) has significant potential in healthcare research and chemical discoveries.
Pharmaceutical companies are using AI to improve drug development by utilizing computational
biology and machine learning systems to predict molecular behavior and the likelihood of finding a
useful drug. This saves time and money on unnecessary tests. Clinical studies, electronic medical
records, high-resolution medical images, and genomic profiles can be used as resources for drug
development. Strong AI systems can analyze extensive data sets in pharmaceutical and medical research.
This review focuses on integrating knowledge of cancer drugs, drug resistance, next-generation
sequencing, genetic variants, and structural biology in cancer precision drug discovery.
in AI for extracting structure from data: supervised learning, which predicts the labels of novel items,
and unsupervised learning, which has recently leap-frogged all previous algorithms in competition. More
breakthroughs in AI are expected in the coming years, in various domains. AI systems have the ability to
see patterns in high-dimensional data and serve as powerful tools to assist researchers rather than replace
them.
Artificial intelligence (AI) is a complex concept, with its definition largely based on Alan Turing's
Turing test. AI refers to a machine's ability to communicate, reason, and operate independently in both
familiar and novel scenarios, similar to a human. However, this is far beyond the scope of current
methods and is not commonly used. AI is often used interchangeably with machine learning or deep
learning, a specific form of machine learning that involves algorithms and statistical models learning
from labelled training data to recognize and infer patterns. This makes AI a complex and challenging
concept to define precisely.
Artificial intelligence is a branch of computer science that simulates intelligent behavior in computers
using algorithms established by humans or learned through computer methods. Machine learning is a
subfield of AI that improves performance by continuously incorporating new data into an existing
model. Deep learning (DL) is a subfield of ML that uses mathematical algorithms deployed on multi-
layered computational units, such as neural networks with different architectures, to simulate human
cognition. Artificial neural networks, with their different architectures, can effectively analyze
unstructured data, a common type of medical data. These unstructured data, often acquired through
patient-provider interactions or imaging, can be processed using natural language processing (NLP)
techniques and recurrent neural networks. Convolutional neural networks are the most promising AI
architectures for exploring imaging files, despite their potential in other data types. Machine learning
(ML) models are developed and validated through various steps, including problem identification, data
collection, pre-processing, training, internal validation, testing, optimization, evaluation, and external
validation[3]. These steps are crucial for creating reliable models that can be applied in clinical practice.
After deployment, results and application should be monitored for drift checking to ensure model
consistency. Clinical utility of ML models must be assessed using specific metrics, such as the receiver
operating characteristic curve (ROC curve) and confusion matrix, which plot the true positive rate and
false positive rate, and the accuracy level.
Figure 1: Artificial Intelligence Flywheel: Graphic representation of the artificial Intelligence and data
cycle for building effective and responsible machine learning models for healthcare.
The last decade has seen significant advances in the ability of AI to solve difficult problems once
thought intractable for artificial systems. Neuroscientists have also begun using ANNs as models of real
neural computation, with new experimental techniques facilitating direct comparisons between ANNs
and real brains. This correspondence has been observed in feedforward and recurrent neural networks
across various brain areas, including visual, language, motor, and prefrontal areas.
For example, the inferotemporal cortex, a critical region for representing object identity in primates, is
spatially organized into a map of object space whose two axes are the same as those in a late layer of the
deep network trained on object classification. Computational neuroscientists interested in understanding
learning and plasticity in the brain began looking at the techniques used to train ANNs and found that
some of the same principles could theoretically be in operation in the brain.
This has led to a huge increase in research using AI systems to model various aspects of animal behavior
and cognition. There is also a renewed hope that neuroscience will provide additional insights for the
development of new ANN approaches that can further advance the state of AI moving forward.
IJFMR240215067 Volume 6, Issue 2, March-April 2024 3
International Journal for Multidisciplinary Research (IJFMR)
E-ISSN: 2582-2160 ● Website: www.ijfmr.com ● Email: [email protected]
An example of this is the evaluation of ecological and economic sustainability via the nutrient cycle. A
balanced and appropriate nutrient cycle is essential for efficient, productive, and sustainable production
of crop and livestock products. In dairy farming, data on feeding and milk yield can be collected to
optimize nutrient cycling. However, optimizing nutrient cycling requires complete and sufficient quality
data. Interoperability, uniform ontologies, and cognitive processing of data are crucial components.
AI in agriculture can provide food security by adapting agricultural management to changing climates,
including identifying resistant crops. Molecular biology tools, such as molecular breeding, genetic
manipulation of DNA to improve animal or plant traits, and plant tissue culture techniques have been
applied for rapid plant production, disease resistance, conservation of endangered species, and genetic
transformation. These technologies aim to improve crop performance in agriculture, resulting in genetic
preservation, uniform plant growth, increased efficiency, and year-round production regardless of
season.
Training
Prediction of High- Model Computatio Prediction of Gene
Artificial
Throughput Crop Intelligence n Function Analysis
phynotyping
Figure 2: Artificial Intelligence used as a powerful tool for the prediction of high-throughput crop
phenotyping and gene functional analysis in modern crop breeding. The high-throughput phenotypic and
genotypic data were collected from large crop germplasm and breeding populations. The massive
comprehensive database could integrate various resources with AI technology, such as phenotypic
diversity of crops, SNPs polymorphisms, QTL analysis, GWAS analysis, genomics selection, and
genome sequence. AI technologies are applied to predict the crop phenotype with whole genome
prediction, the novel breeding strategies are produced through AI related to computation and training
models.
• Predictive modeling AI can use data from satellite and drone imagery to predict tree growth and yield,
optimizing forest planting and management. Advanced imaging techniques enable non-invasive
visualization of tumor extent and metabolic activity, playing a crucial role in oncology patient diagnostic
work-ups and surveillance. However, current tumor staging and surveillance criteria are primarily based
on anatomic criteria. Clinical evaluations are often basic, relying on size measurements and response to
treatment, with qualitative assessments of tumor characteristics like homogeneity and shape. This is
possible due to the ease of making size measurements and the universality of image display and analysis
platforms. However, there is a risk of under-utilization of substantial information, potentially wasting
valuable information for tumor evaluation and treatment planning for oncology patients.
The skeleton is crucial for human motion and activities, but bone density decreases with age. Common
bone diseases like osteoporosis and Paget's are common in the elderly, and a combination of multiple
drugs is the optimal treatment. Artificial intelligence (AI) can evaluate drug-drug interactions (DDIs)
and their underlying adverse effects. This research developed an AI-based machine-learning model to
predict outcomes of interactions between drugs used for osteoporosis and Paget's treatment, reducing the
cost and time of implementing the best combination of medications in clinical practice. A DDI dataset
was collected from the DrugBank database, and chemical features were extracted from the simplified
molecular-input line-entry system (SMILES) of defined drug pairs. Machine-learning algorithms were
used to learn these features. The stack ensemble model from Random Forest and XGBoost achieved an
average accuracy of 74% in predicting DDIs, showing the potential of AI models in predicting DDIs for
osteoporosis-Paget's disease and other diseases.
increase in productivity in all agricultural production areas is needed, while water, energy, pesticides,
and fertilizers should be used diligently and efficiently.
Agriculture faces challenges and uncertainties, necessitating new solutions for all aspects of production.
Integrated Pest Management (IPM) is the most popular and sustainable approach to controlling pests, but
requires extensive knowledge, expertise, and observation. IPM involves monitoring pests in the field,
determining their sensitive stages, and choosing the most appropriate control tactic. This requires
intensive field observation, trained staff, and data mining. Artificial intelligence (AI) algorithms are now
essential for controlling, tracking, and using agricultural inputs optimally. This chapter examines the
impact of Industry 4.0 on modern agriculture and the innovations in IPM in information technologies.
The development of Industry 4.0 has significantly impacted modern agriculture, necessitating new
solutions for improved agricultural production.
Agriculture faces challenges and uncertainties, necessitating new solutions for all aspects of production.
Integrated Pest Management (IPM) is the most popular and sustainable approach to controlling pests, but
requires extensive knowledge, expertise, and observation. IPM involves monitoring pests in the field,
determining their sensitive stages, and choosing the most appropriate control tactic. This requires
intensive field observation, trained staff, and data mining. Artificial intelligence (AI) algorithms are now
essential for controlling, tracking, and using agricultural inputs optimally. This chapter examines the
impact of Industry 4.0 on modern agriculture and the innovations in IPM in information technologies.
The development of Industry 4.0 has significantly impacted modern agriculture, necessitating new
solutions for improved agricultural production.
Deep learning technology has been used in various fields, including image-based pest and disease
recognition research. However, building deep learning models requires large amounts of high-quality
training data. Challenges include collecting large amounts of data on crop pests and diseases, which are
seasonal and difficult to collect for rare, destructive pests like fire blight.
Building high-quality data is crucial for AI training, but labeling is essential for accurate diagnoses.
However, there is a shortage of experts to handle cases where different pests and diseases have similar
damage symptoms, leading to inaccurate diagnoses.
To ensure consistent data quality, a comprehensive data management system is needed. Multiple data
sources, such as image, location, plant, disease, and pesticide info, are essential for efficient pest and
disease monitoring and image-recognition-based diagnosis. This data must be managed in a standardized
format using an integrated data management system that can expand for long-term use.
V. Environmental monitoring:
This paper presents a novel integrated information system (IIS) that combines Internet of Things (IoT),
Cloud Computing, Geoinformatics, and e-Science for environmental monitoring and management[7].
The IIS uses multi-sensors and web services to collect data, while public and private networks access
and transport mass data. Key technologies and tools include real-time operational database (RODB),
extraction–transformation–loading (ETL), on-line analytical processing (OLAP) and relational OLAP
(ROLAP), naming, addressing, and profile server (NAPS), application gateway (AG), application
software for different platforms and tasks (APPs), IoT application infrastructure (IoT-AI), GIS and e-
Science platforms, and representational state transfer/Java database connectivity (RESTful/JDBC).
Application Program Interfaces (APIs) are implemented in the middleware layer of the IIS, providing
functions for storing, organizing, processing, and sharing data and information, as well as functions in
environmental monitoring and management.
The case study on regional climate change and its ecological effects in Xinjiang reveals an increasing
trend in air temperature and precipitation over the last 50 years. Water resource availability is a decisive
factor for the terrestrial ecosystem in the area. The IIS greatly benefits research work, improving data
collection, web services, and applications based on cloud computing and e-Science platforms. This paper
provides a prototype IIS for environmental monitoring and management, providing a new paradigm for
future research and practice, especially in the era of big data and IoT.
As AlphaFold is a relative newcomer to the field, its full implications are not yet fully understood, but
they are likely to be far-reaching and long-lasting.
Furthermore, the software is available for researchers to use. So, if you are working on a protein in the
lab whose structure is unknown, why not discover its predicted structure using AlphaFold.
circuits in the brain responsible for visual processing. This resource is ideal for students and researchers
looking to bridge different approaches to studying and developing visual systems.
Procedures and challenges to protein structure prediction are discussed in numerous review articles.
After modeling, the structure is checked using tools like PROCHECK, MOLPROBITY, VERIFY_3D,
and WHAT_CHECK. 3D structures of proteins allow for structural analyses like functional annotation,
interaction pathway analyses, and target drug identification. Protein-protein interactions are mediated by
hydrophobic interactions and H-bonds, which are crucial for designing and engineering functional
assemblies. Numerous algorithms have been developed to understand these interactions.
The DeepMind AlphaFold2 (AF2) deep learning method demonstrated exceptional performance in blind
protein structure predictions during the 2020 Critical Assessment of Protein Structure Prediction
(CASP14) [21]. The RosettaFold method also demonstrated excellent modeling accuracy on natural
proteins and was particularly successful in modeling de novo designed proteins. Modified approaches
with AI platforms can often reliably model protein-protein complexes and multimeric assemblies.
The Protein Structure Validation Software suite (PSVS) integrates multiple tools for protein structure
validation, with a particular focus on protein NMR structure validation. PSVS provides knowledge-
based protein structure validation tools, such as Molprobity clash and Ramachandran backbone analyses,
Verify3D and ProsaII protein fold analysis tools, and PDBStat software for distance and dihedral angle
restraint violation analysis.
holistically and integratively, focuses on analyzing large and complex datasets, developing molecular-
level models, and assisting in drug discovery and development.
Figure 4: Schematic diagram of biological and artificial computing systems. a) The human brain.
b) The biological neural network. c) A biological synapse. d) A biological neuron. e) An AI chip.
Image-based risk models are being developed by researchers, but they require extensive scientific proof
and computational advancement. Factor models assign a person the probability of a future unfavorable
outcome over a specific time period. AI, ML, and DL are useful in early cancer prediction, improving
the chances of cure and survival.
Aberrant DNA methylation could be a significant biomarker for tumor diagnosis, therapy, and
prognosis, as it is closely connected to various disorders and is typically more stable than gene
regulation. These tools can help improve early cancer diagnosis, therapy, and prognosis.
inference and estimation, while newer methods often emphasize prediction. Common statistical
applications in plant pathology include analyzing designed experiments to understand the relationships
between predictors and responses.
Machine learning communities are developing improved algorithms for various applications, such as
surveillance strategies for Xylellafastidiosa. Algorithms often include wrappers to find optimal values of
hyperparameters and build new models based on the shortcomings of earlier models. Cross-validation is
used to evaluate predictions for data not used in learning the model.
However, optimization options in machine learning can result in overfitting, which is constructing a
predictive model that is overly specific to the data used for learning [31]. A common weakness in
machine learning is the inability to perform with new types of data, particularly in the context of climate
change. Global analyses generally include interpolation and extrapolation, but do these methods provide
predictions useful for finer-scale economic decision-making?
Advances in deep learning methods have made new options available for extremely detailed data sets,
such as images and videos, weather prediction, and disease management [32]. Neural networks have also
been used in plant pathology for decades, but greater computational speed broadens the opportunities for
using deep learning.
Combining deep learning methods with other approaches has the potential to provide the best of both
worlds. Digital agriculture has the potential to incorporate robotics into disease management through
ground robots, offering value through automated detection and targeted treatment.
overcome these challenges by processing and analyzing large amounts of data quickly and accurately.
AI can detect patterns that humans may not be able to detect, identify endangered species, monitor their
movements, and predict their behavior [35].
This paper discusses the concept of smart cities and how AI-based systems are being used for their
management. It also discusses how blockchain technology can empower smart cities, and how AI-based
drones can enhance surveillance of events. The paper also discusses the security concerns in cities and
how they are addressed with AI-based drones. In conclusion, emerging technology such as blockchain
can help improve the management of smart cities.
including access controls, security measures, data anonymization, and encryption techniques.
Various approaches to maintain data integrity include data validation and cleaning techniques, anomaly
detection and pattern analysis, and federated learning techniques. Data validation and cleaning ensure
quality and eliminate biases, while anomaly detection and pattern analysis identify potential
manipulations or malicious attacks. Federated learning techniques keep data in their original locations,
minimizing exposure of sensitive data. Overall, data protection is essential for building trust in AI
systems and ensuring the accuracy, fairness, and reliability of AI models [39].
The primary topics discussed are concepts, basic principles, and methods for performance assessment of
ML algorithms in regulatory science, but not regulatory policy[42]. The question of which components
should be included in a specific regulatory submission depends on factors such as device risk, impact on
clinical practice, technology complexity, and precedents.
The topics selected are based on experience and expertise but are not intended to be comprehensive. ML
algorithms are developed for both imaging and non-imaging modalities for treating brain disorders, with
most discussions applicable to ML algorithms in general imaging applications [43]. Assessment methods
are well established, but better methods may become available and adopted by researchers, developers,
and regulatory agencies in the future.
To provide readers with a more specific sense of the scope of applications relevant to these discussions,
the American College of Radiology (ACR) and FDA public databases were reviewed to summarize
major scope characteristics including imaging modalities, functionalities, and types of ML algorithms.
Social acceptance of AI technologies is crucial, as they can be used for border control, generating new
knowledge on migration and migrants, detecting emerging trends in migration, understanding migrants'
experiences, recognizing their material and social needs, and detecting human rights violations. While
some argue that AI technologies are always harmful when applied to migrants, it is essential to consider
whether it is possible to block the use of these technologies while reducing risks while expanding their
potential benefits.
Identifying the actors who are asked to accept AI technologies applied to migration is also problematic
[45]. With the growing fragmentation and diversification of social life in late modernity, identifying
well-defined social groups and dominant "public opinion" has become increasingly difficult. Personal
orientations, attitudes, choices, or lifestyles have an increasing weight in defining how individuals deal
with public issues such as technologies or migration, while other social variables like professional group
affiliation, income class, living conditions, or religious beliefs have less influence on people's thinking
and behaviors.
In conclusion, the use of AI technologies in migration is a complex and politically sensitive area, and it
is essential to consider how risks can be reduced while enlarging their potential benefits [46].
Artificial Intelligence (AI) technologies, including machine learning, rule-based systems, natural
language processing, and speech recognition, are gaining momentum as disruptive technologies in
various industries [47]. They are also beginning to be adopted in the public sector, such as education,
social policy, and regulation. AI systems can predict which school teacher will have the greatest value
added, while social policy uses AI to predict high-risk youth for targeted interventions [48].
However, the hype surrounding AI adoption in the public sector is accompanied by uncertainty. While
AI applications are seen as enablers of increased efficiency and effectiveness, they also pose challenges
such as job destruction, privacy infringements, and reinforcement of biases in policy-making [49]. The
healthcare sector is a key area of adoption of AI technologies, with high investments in new
technologies. AI has the potential for transformative work, such as mining medical records, assisting
repetitive jobs, and designing treatment plans [50]. However, the healthcare sector presents unique
characteristics due to its ecosystem of stakeholders, including government agencies, service-delivering
public organizations, and private IT firms providing technology.
Despite the media hype, there is still little research on AI in the public sector. There is little empirical
research to test assumptions or provide guidelines for the governance of this emerging phenomenon.
This study aims to fill this gap by addressing the perceived challenges of AI adoption in the public
healthcare sector.
This study explores the challenges of adopting IBM Watson, an AI system, in China's public healthcare
ecosystem. The research focuses on government policy-makers, hospital managers/doctors, and IT firm
managers. China's healthcare services are heavily influenced by public sector intervention in regulation,
strategy, and financing. Watson, developed by IBM, is being introduced in some regions to answer
natural language questions and design personalized treatment plans based on medical literature [52]. The
study aims to map the challenges of AI adoption by key stakeholders in the public sector, allowing for a
better understanding of AI impacts beyond speculation..
AI's hallmark lies in its ability to learn from data, recognize patterns, and make predictions or decisions
with remarkable accuracy [56]. This prowess has propelled AI into numerous domains, transcending
traditional boundaries and reshaping industries. For example, AI-driven diagnostics and personalized
treatment recommendations promise more effective patient care in healthcare, predictive analytics and
algorithmic trading systems revolutionize investment strategies in finance, and autonomous vehicles in
However, ethical considerations and societal implications remain large, with questions regarding data
privacy, algorithmic bias, job displacement due to automation, and the ethical use of AI. As AI systems
become more autonomous and embedded in critical decision-making processes, ensuring transparency,
fairness, and accountability becomes imperative. Striking a balance between innovation and ethical
responsibility remains a central challenge in the ongoing development and deployment of AI
technologies [58].
IoT is a crucial component in the communication paradigm of everyday objects, enabling devices to
communicate and become essential for smart cities [59]. It is used by national governments and private
organizations in ICT solutions to manage the concept of smart cities, aiming to improve community
resources, service quality, and operational and management costs. IoT technologies are essential in
developing the landscape of current smart cities and guiding the standard to the enormous data scale
[60]. According to Statista Research, by 2030, the global count of IoT-enabled devices is projected to
exceed 29 billion, nearly three times the figure in 2020. IoT is considered one of the most valuable
emerging technologies, bringing new opportunities and challenges in implementing intelligent
applications and offerings. It is essential for the advancement and progression of diverse smart city
applications, driving sustainable development and fostering innovative solutions.
IoT is a vital component in everyday objects, enabling communication and becoming cities for essential
for smart [61]. It is used by governments and private organizations in ICT solutions to improve
community resources, service quality, and operational costs. By 2030, the global count of IoT-enabled
devices is projected to exceed 29 billion, nearly three times the 2020 figure. IoT is considered a valuable
emerging technology, presenting new opportunities and challenges in implementing intelligent
applications. It is crucial for the advancement of diverse smart city applications, driving sustainable
development, and fostering innovative solutions [62].
This review article explores the concept of smart cities and the role of IoT in their development. It
discusses the role of IoT in enabling devices to communicate and becoming essential components of
smart cities [63]. The article also explores the potential of AI in analyzing the vast amount of data
produced by IoT sensors in smart cities. It highlights how this technology can enhance city
administration and lead to better living standards for residents.
The article is structured into several sections, including an introduction to smart cities, a definition of
smart cities, IoT-based technologies necessary for creating smart cities, AI algorithms suitable for smart
cities, future trends of smart cities, such as the integration of 5G communication technology, and the
impact of AI on various applications.
The review article aims to provide a comprehensive overview of the current state and future directions of
smart city development, highlighting the potential benefits of this emerging field for urban residents,
governments, and businesses [64]. It also discusses the challenges and opportunities of integrating AI
and IoT in smart cities, such as reducing human interaction and enhancing city administration. The
article concludes by providing conclusions and prospects for the future of smart cities.
Recent progress has been made in the development of powerful AI techniques, with machine learning
models increasingly useful for identifying hidden relationships within data [66]. The study of machine
learning techniques is constantly evolving, with a growing focus on automated machine learning,
reinforcement learning, and artificial neural networks.
Recurrent neural networks and convolutional neural networks are developing, mainly applied to
processing sounds, images, and videos. In production systems, ANN and machine learning techniques
can analyze historical data of production processes and identify optimal parameter combinations to
improve efficiency, reduce production time, or minimize material waste [67].
In conclusion, AI is a rapidly evolving field that aims to improve efficiency, reduce production time, and
minimize material waste. Advances in machine learning, artificial neural networks, and reinforcement
learning are driving the development of more powerful AI techniques.
Deep learning is a machine learning method based on artificial neural networks, utilizing multiple layers
to improve performance [68]. With increased computational power and data availability, it has made
significant progress in predictive maintenance and demand forecasting. However, interpretability
remains a challenge. New techniques have been developed to understand and visualize algorithms'
workings, enabling more informed decisions.
Computer vision, a field of computer science, enables computers to observe, identify, and process
images like human vision. Advances have been made in detecting specific objects within images,
improving product tracking and inventory tracking. The field is also advancing in the generation of
realistic, three-dimensional images [69].
Cognitive computing aims to solve complex problems with uncertainty and ambiguity, integrating
different sources for informed, ethical, and responsible decisions. With the exponential increase of data,
it is mainly used for complex resolution problems [70].
Natural language processing, which involves computer programs understanding spoken and written
human language, has been developed to generate coherent and controlled text like ChatGPT [71]. It can
be used to process supply contracts or order documents, reduce processing times, and analyze customer
feedback for valuable information on preferences and performance improvements.
3D printing:
3D printing is a process that creates physical objects from CAD models or digital 3D models using
layers of material fused together. This technology has improved manufacturing productivity by reducing
lead times in prototyping and creating auxiliary devices for new product production. Rapid prototyping
allows for greater innovation and experimentation while reducing costs. However, the speed of 3D prints
is not high, making it less used in mass production [72].
There are various 3D printing techniques depending on the material layer and equipment used. For
example, powder bed fusion uses a laser beam to create physical objects from powder layers,
stereolithography creates three-dimensional objects from a liquid polymer, and polyjet builds parts by
throwing droplets of photopolymer onto a build platform and solidifying them with UV light. These
technologies are particularly useful in aerospace, automotive, and medical sectors, but are mainly used
for small series or prototype production.
Bioprinting combines cells and biomaterials to create biomedical parts, simulating the natural features of
tissues [73]. Fused filament fabrication and fused deposition modelling add layers of molten plastic to
create models or products, mainly used for spare parts and replacement of damaged or obsolete
components. Directed energy deposition uses focused energy sources to melt materials simultaneously
deposited by a nozzle, reducing machine downtime and producing large parts that cannot be created with
traditional methods. Binder jetting is a 3D printing process where an industrial print head selectively
deposits a liquid binding agent on a thin layer of dust particles to build high-value parts.
The value of big data technology increases significantly when well implemented within production
systems, allowing companies to better customize products and achieve high levels of satisfaction [75].
However, it presents challenges such as data protection, relevant data identification, and cleaning
techniques for analyzing data with different formats. Data mining is a process that analyzes hidden
patterns of data from different perspectives to obtain useful information [76]. Sentiment analysis and
user behavior analytics are data mining techniques used to measure people's opinions and extract
personal information from the web. These techniques help monitor customer satisfaction, make informed
decisions on product and service improvements, support marketing strategies, and monitor competitors.
User behavior analytics is a powerful methodology used to analyze and monitor the behavior of
operators in the workplace, involving closely examining and interpreting data related to their actions,
activities, and time spent working with machines.
Blockchain technology:
Blockchain technology is a distributed ledger that records data in blocks linked together using
cryptography, ensuring linear and chronological storage. It is known for its role in cryptocurrencies and
is increasingly used in production systems for its benefits such as product traceability, security, and
efficiency [77]. Blockchain eliminates intermediaries and costs associated with transactions, allowing
for faster completion and transparency.
Another tool connected to blockchain is the smart contract, which is a code that can be integrated into
the blockchain to facilitate and verify contractual agreements [78]. Smart contracts operate on
predefined conditions agreed by users and are automatically executed when these conditions occur.
These tools can facilitate various activities within supply chain management (SCM) and automate
payments between parties involved only upon fulfilling predefined conditions, such as successful
product delivery.
In conclusion, blockchain technology offers numerous benefits, including cost reduction, efficiency,
transparency, and collaboration in various sectors. However, challenges such as block size debates and
unclear government regulations remain.
Computing:
Cloud-related technology involves data storage on multiple servers and online access from any device
[79]. It offers applications such as email, archiving, data backup, analytics, and on-demand software.
Cloud-based software offers companies benefits such as cost savings in information management
technologies and infrastructures, improved monitoring and maintenance of production equipment, and
collaboration between teams through data sharing. However, there are risks associated with cloud
computing, such as data security concerns and potential natural disasters, internal bugs, and power
outages.
To improve inefficiencies, new tools and techniques have been developed, such as edge computing, fog
computing, and cloudlet computing. Edge computing processes data as close to its origin to reduce
network latency by minimizing communication times between client and server. Fog computing is an
infrastructure interposed between edge and cloud computing to enable more efficient data processing,
analysis, and storage. Cloudlets are small-scale data centers designed to quickly deliver cloud computing
services to mobile and wearable devices, increasing response time of applications [80].
Real-time data analysis is performed directly in the proximities of items, reducing latency and enabling
timely responses from data generated by sensors and production machines [81]. These systems are
useful for early detection of anomalies or machine malfunctions. Quantum computing, based on
quantum theory principles, can handle operations at significantly faster speeds and with lower power
consumption, optimizing production processes to solve complex problems such as managing workflows
or reducing production times.
Digital applications:
Digital applications, including social media & network, mobile applications, and web applications, have
become essential tools for people and companies. Social media & network enable users to share content
and interact with others, increasing sales through advertising, promotions, and customer service [82].
Mobile applications are software designed for smartphones and tablets, providing functions similar to
PC applications. They are used in production systems for monitoring machines, updating order statuses,
tracking material stocks and inventory, and facilitating real-time communication.
Web applications are software programs that run on a web server and require a web browser to access.
They offer several benefits over desktop applications, such as not being developed for multiple
platforms and not having direct access to updated versions. Web apps allow users to access data from
multiple devices, planning production, allocating resources, managing inventory, and improving
coordination of activities. They can also facilitate supplier relationship management through shared
order management systems.
Robotic process automation is another digital application that automates routine business practices with
software agents (bots) that perform tasks automatically. Bots digitally replicate human activities,
automating activities without direct human intervention [83]. Technologies necessary for digitization in
this area include digital signatures, electronic invoices, and digital contracts. Overall, digital applications
have become indispensable tools for people and companies, enabling efficient use of resources, better
coordination, and improved productivity.
Geospatial technologies:
Geospatial technologies are modern tools that aid in mapping and analyzing Earth's surface and human
activities. Satellites have enabled the detection of Earth's surface images and human activities, while
computers have enabled the storage and transfer of GIS data[84]. These technologies have been widely
used in industrial engineering, agricultural, and environmental monitoring in the last decade. They map
manufacturing resources, such as equipment, warehouses, and facilities, to identify efficient routes for
resource management. The data collected can also identify areas of environmental importance for
marketing and influence product demand. Geospatial technologies are also being used to improve safety
at work by identifying high-risk areas and preventing employee health and safety [85].
Several geospatial technologies include remote sensing for data acquisition, web mapping, GPS for
position, speed, and synchronization, and the global navigation satellite system for global Earth
coverage. These technologies are used in logistics to track goods, monitor outbreaks, assess crop status,
and target advertisements to specific users.
enormous amounts of data and to perform labor-intensive tasks in short times, makes AI extremely
appealing, stimulating researchers to explore applications in several scientific fields.
Molecular biology could benefit considerably from AI. There is a lot of interest in trying to predict the
three-dimensional structure of proteins from DNA sequences. Similarly, many efforts are being made in
astronomy and the study of the universe [86].
Space observations in fact produce huge amounts of data where instrumental artifacts need to be
identified and removed before real data can be analyzed. AI and machine learning can perform this
challenging task, “cleaning” data for later analyses.
Advocating for ethics and diversity is crucial to account for biased models and avoid stereotypes
perpetuated by AI systems [89]. Interdisciplinary approaches, including art and science, and ensuring
minorities are well represented among users and evaluators of the latest eXplainable AI techniques can
make AI more accessible and inclusive to otherwise unreachable communities.
The AI revolution is rapidly advancing in research, healthcare, and industry, but its long-term impact on
society will not be immediately apparent. AI methods may be applied to problems that are not yet ready,
leading to ethically questionable applications such as predicting sexual orientations from people's faces
or using facial recognition in law enforcement or commercial use. AI can improve data privacy and
threat identification, but it is often seen as a threat to IT systems.
In industry, AI chatbots have been racist due to training data presented to the algorithm, recruitment
software has been gender-biased, and risk assessment tools developed by a US contractor have sent
innocent people to jail. A more careful consideration of the impact of AI is needed by following global
and local ethics guidelines for trustworthy and responsible AI.
drug targets is crucial for increasing the likelihood of developing clinically effective therapies. Target
identification, the process of identifying the right biological molecules or cellular pathways that can be
modulated by drugs to achieve therapeutic benefits, is increasingly important in modern drug discovery.
Although innovations in experimental and omic technologies have been growing over the past few
decades, identifying actionable therapeutic targets remains challenging. The integration of multiomic
data with AI (see Glossary) algorithms has recently emerged as a promising approach for target
identification ii, iii. We discuss here the conventional target identification approaches with a focus on
the application of AI algorithms to target identification. This paper aims to offer a progressive outlook
on the emergence of the AI-driven drug discovery era and encourage the integration of AI technologies
into drug discovery pipelines. Strategies in target identification: from experiments to machine learning
Target identification can be classified into three distinct strategies – experimental, multiomic, and
computational approaches. Using these methods collaboratively can generate novel therapeutic
hypotheses in exploratory target identification, thus significantly enhancing our understanding of
complex diseases.
The emergence of artificial intelligence (AI) in early drug development. (Upper panel) Key
technological advances in the history of target identification are classified into three types: experiment-
based (red), multiomic (blue), and computational (green) approaches. Traditionally, experiment-based
methods have been the go-to approach for discovering therapeutic targets. However, with the rise of big
data, integrated analysis of multiomic data has become a more efficient strategy for target identification.
In addition, recent advances in AI-driven biological analysis have identified novel targets and AI-
designed drugs are now entering clinical trials. (Lower panel) AI applications in the early stages of drug
discovery. Abbreviations: AGC chemistry, affinity-guided catalyst chemistry; ALS, amyotrophic lateral
sclerosis; DL, deep learning; EGFR, epidermal growth factor receptor; GAN, generative adversarial
network; GWAS, genome-wide association study; LD chemistry, ligand-directed chemistry; MTOR,
mammalian target of rapamycin; NSCLC, non-small cell lung cancer; SILAC, stable isotope labeling
with amino acids in cell culture; TID, target identification. Figure created with BioRender.com [92].
(Industry 4.0 or 4IR), with the capability of incorporating human behavior and intelligence into
machines or systems [94]. Thus, AI-based modeling is the key to build automated, intelligent, and smart
systems according to today’s needs. To solve real-world issues, various types of AI such as analytical,
functional, interactive, textual, and visual AI can be applied to enhance the intelligence and capabilities
of an application. However, developing an effective AI model is a challenging task due to the dynamic
nature and variation in real-world problems and data. In this paper, we present a comprehensive view on
“AI-based Modeling” with the principles and capabilities of potential AI techniques that can play an
important role in developing intelligent and smart systems in various real-world application areas
including business, finance, healthcare, agriculture, smart cities, cyber security and many more. We also
emphasize and highlight the research issues within the scope of our study. Overall, the goal of this paper
is to provide a broad overview of AI-based modeling that can be used as a reference guide by academics
and industry people as well as decision-makers in various real-world scenarios and application domains.
Nowadays, we live in a technological age, the Fourth Industrial Revolution, known as Industry 4.0 or
4IR , which envisions fast change in technology, industries, societal patterns, and processes as a
consequence of enhanced interconnectivity and smart automation. This revolution is impacting almost
every industry in every country and causing a tremendous change in a non-linear manner at an
unprecedented rate, with implications for all disciplines, industries, and economies. Three key terms
Automation, i.e., reducing human interaction in operations, Intelligent, i.e., ability to extract insights or
usable knowledge from data, and Smart computing, i.e., self-monitoring, analyzing, and reporting,
known as self-awareness, have become fundamental criteria in designing today’s applications and
systems in every sector of our lives since the current world is more reliant on technology than ever
before. The use of modern smart technologies enables making smarter, faster decisions regarding the
business process, ultimately increasing the productivity and profitability of the overall operation, where
Artificial Intelligence (AI) is known as a leading technology in the area. The AI revolution, like earlier
industrial revolutions that launched massive economic activity in manufacturing, commerce,
transportation, and other areas, has the potential to lead the way of progress. As a result, the impact of
AI on the fourth industrial revolution motivates us to focus briefly on “AI-based modeling” in this paper.
Artificial intelligence (AI) is a leading technology of the current age of the Fourth Industrial Revolution,
with the capability of incorporating human behavior and intelligence into machines or systems [95].
Thus, AI-based modeling is the key to build automated, intelligent, and smart systems according to
today’s needs. To solve real-world issues, various types of AI such as analytical, functional, interactive,
textual, and visual AI can be applied to enhance the intelligence and capabilities of an application.
However, developing an effective AI model is a challenging task due to the dynamic nature and
variation in real-world problems and data. In this paper, we present a comprehensive view on “AI-based
Modeling” with the principles and capabilities of potential AI techniques that can play an important role
in developing intelligent and smart systems in various real-world application areas including business,
finance, healthcare, agriculture, smart cities, cyber security and many more. We also emphasize and
highlight the research issues within the scope of our study. Overall, the goal of this paper is to provide a
broad overview of AI-based modeling that can be used as a reference guide by academics and industry
people as well as decision-makers in various real-world scenarios and application domains.
Ethical AI:
AI ethicists have been formulating ethical principles and frameworks to responsibly guide the
development and implementation of AI systems [97]. A key contribution came from who synthesized a
set of core ethical principles like beneficence, non-maleficence, autonomy, justice, and explicability for
cultivating a “Good AI Society.” However, as principles alone they are in sufficient they need to be
translated into concrete practices. Much work remains to develop these into practical tools and
methodologies. Recent work has begun exploring approaches for embedding ethical values directly into
AI system design, from which Value-sensitive Design (VSD) has emerged as a candidate to bridge the
principle-practice gap. A recent critical review indicates that VSD may be effective but limited. Some
limitations include the inadequate elicitation of values, a tendency to depend on pre-established values
over context-specific ones, and a lack of precise instructions for embedding values.
AI Alignment:
Considering the potential for harm done by AI, some refer to such systems as misaligned with human
values. For example, referencing social media, ethicist Tristan Harris says that by optimizing for
attention, these platforms are misaligned with human wellbeing and dignity. “AI alignment” is a field of
research that aims to develop systems that are aligned with human values and intent. Alignment has been
earlier studied as the principal-agent problem in economics and law, where an agent must achieve the
objectives and interests of the principal . For example, in a car repair scenario, the car owner (principal)
expects the mechanic (agent) to fix the car efficiently and affordably. However, the mechanic might
suggest unnecessary repairs to increase the bill (misalignment), contrary to the owner’s desire for cost-
effective service. Taking this framework to AI systems, alignment means ensuring that AI agents
effectively and reliably pursue the goals and preferences set by their designers and users. One successful
example of technical alignment work is the use of Reinforcement Learning with Human Feedback
(RLHF) , which uses human preference data to align the behavior of Large Language Models (LLMs).
Related techniques include Constitutional AI and inverse reinforcement learning (IRL). There are many
new techniques in the expanding field of technical AI alignment.2However, while these technical efforts
show tremendous progress, their technology-centered perspective risks missing broader sociotechnical
considerations, such as the design of human systems to effectively respond to AI .
engineering design, including problem definition, scoping, and rapid prototyping, combined with
methodologies from social sciences like conducting ethnographic research, interviews, deriving
understanding from qualitative data, and engaging in empathetic practices. For a comprehensive
overview, readers are referred to the live agenda summarizing ongoing alignment efforts posted to the
AI Alignment Forum, which has been founded by prominent alignment researcher Eliezer Yudkowsky.
Why wellbeing?
A growing movement of scholars advocates for the incorporation of wellbeing metrics into AI systems
so that optimization efforts can measurably contribute to social benefit. Specifically, they argue that
measures of wellbeing can help manage AI’s effects on society. Indeed, wellbeing has a strong
methodological foundation , and there is extensive research on defining and measuring wellbeing; this
suggests that algorithmic systems may be able to systematically optimize wellbeing. Wellbeing’s
complexity captures many relevant societal concerns AI systems should address. This combination of
rich meaning and inherent measurability supports the operationalizing wellbeing as an optimization
objective for AI systems. This sentiment is also expressed by a recent IEEE standards review that argues
for the adoption of holistic wellbeing frameworks (like IEEE 7010) to guide the design, deployment, and
evaluation of AI systems. However, significant questions remain regarding whether available wellbeing
frameworks are fully sufficient, whether existing metrics are sufficient, what the impacts of wellbeing
optimization may be. Some argue that wellbeing is a sort of ultimate objective: in The Moral Landscape,
S. Harris (2010) argues that other values like fairness, transparency, or accountability should be seen as
components that contribute to wellbeing, rather than ends in themselves. From this perspective,
optimizing for wellbeing involves optimizing for all values that matter, but only insofar as they
contribute empirically to wellbeing. In so far as AI systems are able to assess their own impact on
human wellbeing, they may be able to potentially maximize all benefits and minimize all harms
experienced by users and society .Wellbeing optimization might then allow for the management of
complex issues like misinformation and inequality associated with AI systems (Stray, 2020).
Key Challenges:
As experienced design researchers in this domain, we have consistently encountered unique challenges
in designing Positive AI. The challenges we outline here are intended to inform and guide other
designers embarking on similarventures.
The far-reaching achievements of deep learning, such as ChatGPT, have led several scholars to raise the
question of whether AI can reach the level of “artificial general intelligence,” namely, whether AI can
reach and then surpass the level of human thought. The milestone where AI, supposing, reaches human
intelligence was called by Vernor Vinge, the “singularity.” This notion has been popularized by the
futurist scholar and director of Google Engineering, Ray Kurzweil.
There is in my opinion a serious methodological problem concerning the possible proof that a machine
has surpassed the human level of intelligence. This would require a proof that for every conceivable
human goal, the machine achieves better performance. So far, this has been accomplished for particular
goals by the direct competition of a human expert and a machine. For example, this happened for the
goals of winning a chess or a Go game. However, it is apparent that such an approach cannot be used for
an uncountable number of possible situations. Hence, unless a different methodology is suggested, the
question of proving whether general AI has been reached is not well defined.
In any case, even if we assume that this question is well-posed, its analysis, necessitates, first, the
introduction of a definition of intelligence. The cosmologist and leading AI exponent, Max Tegmark,
attempted to provide such a definition. He defined intelligence as “the ability to accomplish complex
goals.” In his important book, Life 3.0: being human in the age of artificial intelligence, it is claimed that
this definition encompasses Oxford Dictionary’s definition of intelligence, “as the ability to acquire and
apply knowledge and skills,” as well as several of the definitions proposed in the Nobel Week Dialogue
2015: The Future of Intelligence. In this conference, among the definitions proposed for intelligence
were, “the capacity for problem solving, learning, logic, and planning.” According to Tegmark,
acquiring knowledge, learning new information or a new skill, solving specific problems, employing
logical algorithms, and designing concrete plans, can all be considered as processes subsumed by the
phrase “accomplishing complex goals.” It will be argued below, that although Tegmark’s definition of
intelligence is adequate for machines, it does not capture the essence of human thought. Indeed, I believe
that this definition is appropriate for technology, which can be defined as a collection of devices and
engineering practices as means of achieving a complicated goal.
reported, including cytosine base editors (CBEs)that induce C•G to T•A conversion, adenine base editors
(ABEs)that induce A•T to G•C conversion, and glycosylase base editors (GBEs) with C•G to G•C
conversion.
These BEs provide almost ideal solutions for treating more than half of known pathogenic SNVs.
However, before BE-based genetic therapies can be implemented, it is necessary to construct
mammalian cell disease models for developing and optimizing BEs and enabling applications in gene
therapy. Due to the great number of genetic diseases and known SNVs, it is necessary to develop a
method that enables the construction of a large number of cell models carrying different pathogenic
SNVs, so that extensive research on gene therapies can be performed to find curing solutions. According
to ClinVar, approximately 50% of total human pathogenic SNVs are C•G to T•A conversion, which can
be corrected by ABEs. However, it is currently hard to obtain large numbers of cell models carrying
these SNVs with reasonable labor and funding investment.
Mammalian cell lines are usually used for developing and optimizing the BEs, and research for
prediction of base editing performance, such as efficiency and specificity, requires a large amount of
editing data [102]. To solve this problem, methods based on a target-locus integration library were
developed, such as the Be-Hive, which provided the data for AI to learn and predict the editing
performance. However, currently such data was obtained from integrated editing loci which lacked in
situ information. Previous research has shown a strong correlation between the performance of nuclease
and the chromatin accessibility properties. Kristopher et al. demonstrated that gene editing was more
efficient in euchromatin than in heterochromatin. Large-scale genetic screens in human cell lines
indicated that highly active sgRNAs for Cas9 and dCas9 were found in regions of low nucleosome
occupancy, and the nucleosomes directly impeded Cas9 binding and cleavage in vitro. We previously
found that pioneer factor, such as Vp64, improved CRISPR-based C-to-G and C-to-T base editing by
changing local chromatin environment. However, current studies on deep learning employed the editing
data from lentiviral integrated target sequences, while the real chromosomal environment of the target
sequence was ignored. One of the reasons is that it is difficult to obtain a large set of editing data from
endogenous target sites. For large scale samples, manual operations are not only time-consuming, but
also error prone, less consistency and expensive. An automatic platform would make it possible to get
large-scale editing dataset of endogenous targets. And with the large-scale in situ editing data and
sequence information, combined with local chromatin accessibility, a machine learning model with in
situ data might be able to better predict the actual base editing efficiency.
In this study, we devise an automated platform that performs the whole genome editing process from
guide RNA (gRNA) design to the analysis of the editing results, which comprehensively characterizes
the relationships of the in situ base editing outcomes with the sequence and chromatin environment for
BEs.
(I). Conclusion:
Artificial Intelligence (AI) is a double-edged sword: on one hand, AI promises to provide great advances
that could benefit humanity, but on the other hand, AI poses substantial (even existential) risks[96]. With
advancements happening daily, many people are increasingly worried about AI's impact on their lives.
To ensure AI progresses beneficially, some researchers have proposed "wellbeing" as a key objective to
govern AI. This article addresses key challenges in designing AI for wellbeing. We group these
challenges into issues of modeling wellbeing in context, assessing wellbeing in context, designing
interventions to improve wellbeing, and maintaining AI alignment with wellbeing over time. The
identification of these challenges provides a scope for efforts to help ensure that AI developments are
aligned with human wellbeing. It is not a naturally observable quantity, but rather a multifaceted
construct that is based, at least in part, on conscious human experiences. Therefore, designing Positive
AI requires understanding and shaping human experiences. This situates the challenge squarely in the
domain of human-centered design .Before reviewing the possibilities for HCD in designing AI for well-
being, we will briefly address other fields associated with the creation of positive human outcomes in
AI. The current article is not the venue for reviewing them in depth. Yet, we find it important that
Positive AI designers are aware broadly aware of their contributions.
Acknowledgement: Authors would like to thank R. R. Institute of Modern Technology for providing a
platform.
References:
1. Konar A. Artificial intelligence and soft computing: behavioral and cognitive modeling of the
human brain. CRC press; 2018 Oct 8.
2. Feng C, Zhou S, Qu Y, Wang Q, Bao S, Li Y, Yang T. Overview of artificial intelligence
applications in Chinese medicine therapy. Evidence-based complementary and alternative
medicine. 2021 Mar 17;2021.
3. Rahmani AM, Yousefpoor E, Yousefpoor MS, Mehmood Z, Haider A, Hosseinzadeh M, Ali Naqvi
R. Machine learning (ML) in medicine: Review, applications, and challenges. Mathematics. 2021
Nov 21;9(22):2970.
4. Vatansever S, Schlessinger A, Wacker D, Kaniskan HÜ, Jin J, Zhou MM, Zhang B. Artificial
intelligence and machine learning‐aided drug discovery in central nervous system diseases: State‐
of‐the‐arts and future directions. Medicinal research reviews. 2021 May;41(3):1427-73.
5. Javaid M, Haleem A, Khan IH, Suman R. Understanding the potential applications of Artificial
Intelligence in Agriculture Sector. Advanced Agrochem. 2023 Mar 1;2(1):15-30.
6. Sandino J, Pegg G, Gonzalez F, Smith G. Aerial mapping of forests affected by pathogens using
UAVs, hyperspectral sensors, and artificial intelligence. Sensors. 2018 Mar 22;18(4):944.
7. Fang S, Da Xu L, Zhu Y, Ahati J, Pei H, Yan J, Liu Z. An integrated system for regional
environmental monitoring and management based on internet of things. IEEE Transactions on
Industrial Informatics. 2014 Jan 27;10(2):1596-605.
8. Müller H, Holzinger A, Plass M, Brcic L, Stumptner C, Zatloukal K. Explainability and causability
for artificial intelligence-supported medical image analysis in the context of the European In Vitro
Diagnostic Regulation. New Biotechnology. 2022 Sep 25;70:67-72.
9. Bhardwaj A, Kishore S, Pandey DK. Artificial intelligence in biological sciences. Life. 2022 Sep
14;12(9):1430.
10. van Houten F, Wertheim R, Ayali A, Poverenov E, Mechraz G, Eckert U, Rentzsch H, Dani I,
Willocx M, Duflou JR. Bio-based design methodologies for products, processes, machine tools and
production systems. CIRP journal of manufacturing science and technology. 2021 Jan 1;32:46-60.
11. Hassoun S, Jefferson F, Shi X, Stucky B, Wang J, Rosa Jr E. Artificial intelligence for biology.
Integrative and Comparative Biology. 2021 Dec;61(6):2267-75.
12. Mahmud M, Kaiser MS, McGinnity TM, Hussain A. Deep learning in mining biological data.
Cognitive computation. 2021 Jan;13:1-33.
13. Mahmud M, Kaiser MS, McGinnity TM, Hussain A. Deep learning in mining biological data.
Cognitive computation. 2021 Jan;13:1-33.
14. Smith KP, Wang H, Durant TJ, Mathison BA, Sharp SE, Kirby JE, Long SW, Rhoads DD.
Applications of artificial intelligence in clinical microbiology diagnostic testing. Clinical
Microbiology Newsletter. 2020 Apr 15;42(8):61-70.
15. Gong Z, Hu G, Li Q, Liu Z, Wang F, Zhang X, Xiong J, Li P, Xu Y, Ma R, Chen S. Compound
libraries: recent advances and their applications in drug discovery. Current drug discovery
technologies. 2017 Dec 1;14(4):216-28.
16. Schmidhuber J. Deep learning in neural networks: An overview. Neural networks. 2015 Jan
1;61:85-117.
17. Kreiman G. Biological and computer vision. Cambridge University Press; 2021 Feb 4.
18. Dhawan AP. Medical image analysis. John Wiley & Sons; 2011 Mar 29.
19. Johnson KB, Wei WQ, Weeraratne D, Frisse ME, Misulis K, Rhee K, Zhao J, Snowdon JL.
Precision medicine, AI, and the future of personalized health care. Clinical and translational
science. 2021 Jan;14(1):86-93.
20. Floudas CA, Fung HK, McAllister SR, Mönnigmann M, Rajgaria R. Advances in protein structure
prediction and de novo protein design: A review. Chemical Engineering Science. 2006 Feb
1;61(3):966-88.
21. Schauperl M, Denny RA. AI-based protein structure prediction in drug discovery: impacts and
challenges. Journal of Chemical Information and Modeling. 2022 Jun 21;62(13):3142-56.
22. Gliwińska A, Czubilińska-Łada J, Więckiewicz G, Świętochowska E, Badeński A, Dworak M,
Szczepańska M. The role of brain-derived neurotrophic factor (BDNF) in diagnosis and treatment
of epilepsy, depression, schizophrenia, anorexia nervosa and Alzheimer’s disease as highly drug-
resistant diseases: a narrative review. Brain Sciences. 2023 Jan 18;13(2):163.
23. Helmy M, Smith D, Selvarajoo K. Systems biology approaches integrated with artificial
intelligence for optimized metabolic engineering. Metabolic Engineering Communications. 2020
Dec 1;11:e00149.
24. Paul D, Sanap G, Shenoy S, Kalyane D, Kalia K, Tekade RK. Artificial intelligence in drug
discovery and development. Drug discovery today. 2021 Jan;26(1):80.
25. Cioffi R, Travaglioni M, Piscitelli G, Petrillo A, De Felice F. Artificial intelligence and machine
learning applications in smart production: Progress, trends, and directions. Sustainability. 2020 Jan
8;12(2):492.
26. Schork NJ. Artificial intelligence and personalized medicine. Precision medicine in Cancer
therapy. 2019:265-83.
27. Al-Amin M, Ali MS, Salam A, Khan A, Ali A, Ullah A, Alam MN, Chowdhury SK. History of
generative Artificial Intelligence (AI) chatbots: past, present, and future development. arXiv
preprint arXiv:2402.05122. 2024 Feb 4.
28. Fosch-Villaronga E, Drukarch H. AI for healthcare robotics. CRC Press; 2022 Jun 7.
29. Xu L, Sanders L, Li K, Chow JC. Chatbot for health care and oncology applications using artificial
intelligence and machine learning: systematic review. JMIR cancer. 2021 Nov 29;7(4):e27850.
30. The healthcare industry is undergoing a transformation with the integration of AI technologies,
particularly chatbots and virtual assistants. Design characteristics play a crucial role in optimizing
user experience.
31. Seyedzadeh S, Rahimian FP, Rastogi P, Glesk I. Tuning machine learning models for prediction of
building energy loads. Sustainable Cities and Society. 2019 May 1;47:101484.
32. Sengupta S, Basak S, Saikia P, Paul S, Tsalavoutis V, Atiah F, Ravi V, Peters A. A review of deep
learning with special emphasis on architectures, applications and recent trends. Knowledge-Based
Systems. 2020 Apr 22;194:105596.
33. Dertien JS, Negi H, Dinerstein E, Krishnamurthy R, Negi HS, Gopal R, Gulick S, Pathak SK,
Kapoor M, Yadav P, Benitez M. Mitigating human–wildlife conflict and monitoring endangered
tigers using a real-time camera-based alert system. BioScience. 2023 Oct;73(10):748-57.
34. Kideghesho J, Rija A, Mwamende K, Selemani I. Emerging issues and challenges in conservation
of biodiversity in the rangelands of Tanzania. Nature Conservation. 2013 Nov 18;6:1-29.
35. Congdon JV, Hosseini M, Gading EF, Masousi M, Franke M, MacDonald SE. The Future of
Artificial Intelligence in Monitoring Animal Identification, Health, and Behaviour. Animals. 2022
Jul 1;12(13):1711.
36. Hatcher WG, Yu W. A survey of deep learning: Platforms, applications and emerging research
trends. IEEE Access. 2018 Apr 27;6:24411-32.
37. Taddeo M, McCutcheon T, Floridi L. Trusting artificial intelligence in cybersecurity is a double-
edged sword. Nature Machine Intelligence. 2019 Dec;1(12):557-60.
38. Li B, Qi P, Liu B, Di S, Liu J, Pei J, Yi J, Zhou B. Trustworthy AI: From principles to practices.
ACM Computing Surveys. 2023 Jan 13;55(9):1-46.
39. Truong N, Sun K, Wang S, Guitton F, Guo Y. Privacy preservation in federated learning: An
insightful survey from the GDPR perspective. Computers & Security. 2021 Nov 1;110:102402.
40. Bolukbasi T, Chang KW, Zou JY, Saligrama V, Kalai AT. Man is to computer programmer as
woman is to homemaker? debiasing word embeddings. Advances in neural information processing
systems. 2016;29.
41. Zhou SK, Greenspan H, Davatzikos C, Duncan JS, Van Ginneken B, Madabhushi A, Prince JL,
Rueckert D, Summers RM. A review of deep learning in medical imaging: Imaging traits,
technology trends, case studies with progress highlights, and future promises. Proceedings of the
IEEE. 2021 Feb 26;109(5):820-38.
42. Chen W, Krainak D, Sahiner B, Petrick N. A Regulatory Science Perspective on Performance
Assessment of Machine Learning Algorithms in Imaging. Machine Learning for Brain Disorders.
2023 Jul 23:705-52.
43. Cenek M, Hu M, York G, Dahl S. Survey of image processing techniques for brain pathology
diagnosis: Challenges and opportunities. Frontiers in Robotics and AI. 2018 Nov 2;5:120.
44. Beduschi A. International migration management in the age of artificial intelligence. Migration
Studies. 2021 Sep 1;9(3):576-96.
45. Vohra A. Social order in the age of artificial intelligence: the use of technology in migration
governance and decision-making (Doctoral dissertation, University of British Columbia).
46. Molnar P. Technology on the margins: AI and global migration management from a human rights
perspective. Cambridge International Law Journal. 2019 Dec 1;8(2):305-30.
47. Sarker IH. Ai-based modeling: Techniques, applications and research issues towards automation,
intelligent and smart systems. SN Computer Science. 2022 Mar;3(2):158.
48. Hunt X, Tomlinson M, Sikander S, Skeen S, Marlow M, du Toit S, Eisner M. Artificial
intelligence, big data, and mHealth: The frontiers of the prevention of violence against children.
Frontiers in artificial intelligence. 2020 Oct 22;3:543305.
49. Sun TQ, Medaglia R. Mapping the challenges of Artificial Intelligence in the public sector:
Evidence from public healthcare. Government Information Quarterly. 2019 Apr 1;36(2):368-83.
50. Dwivedi YK, Hughes L, Ismagilova E, Aarts G, Coombs C, Crick T, Duan Y, Dwivedi R,
Edwards J, Eirug A, Galanos V. Artificial Intelligence (AI): Multidisciplinary perspectives on
emerging challenges, opportunities, and agenda for research, practice and policy. International
Journal of Information Management. 2021 Apr 1;57:101994.
51. Dwivedi YK, Hughes L, Ismagilova E, Aarts G, Coombs C, Crick T, Duan Y, Dwivedi R,
Edwards J, Eirug A, Galanos V. Artificial Intelligence (AI): Multidisciplinary perspectives on
emerging challenges, opportunities, and agenda for research, practice and policy. International
Journal of Information Management. 2021 Apr 1;57:101994.
52. Ahmed MN, Toor AS, O'Neil K, Friedland D. Cognitive computing and the future of health care
cognitive computing and the future of healthcare: the cognitive power of IBM Watson has the
potential to transform global personalized medicine. IEEE pulse. 2017 May 16;8(3):4-9.
53. Rouhiainen L. Artificial Intelligence: 101 things you must know today about our future. Lasse
Rouhiainen; 2018 Jan 31.
54. Kurzweil R, Richter R, Kurzweil R, Schneider ML. The age of intelligent machines. Cambridge:
MIT press; 1990 Sep.
55. Kasula BY. Advancements and Applications of Artificial Intelligence: A Comprehensive Review.
International Journal of Statistical Computation and Simulation. 2016 Aug 17;8(1):1-7.
56. De Bruyn A, Viswanathan V, Beh YS, Brock JK, Von Wangenheim F. Artificial intelligence and
marketing: Pitfalls and opportunities. Journal of Interactive Marketing. 2020 Aug;51(1):91-105.
57. Siddiqui MN. AI REVOLUTION: EMPOWERING THE FUTURE WITH ARTIFICIAL
INTELLIGENCE. Pakistan Journal of International Affairs. 2023 Dec 20;6(4).
58. Konda SR. Ethical Considerations in the Development and Deployment of AI-Driven Software
Systems. INTERNATIONAL JOURNAL OF COMPUTER SCIENCE AND TECHNOLOGY.
2022 Jul 31;6(3):86-101.
59. Petrolo R, Loscri V, Mitton N. Towards a smart city based on cloud of things, a survey on the
smart city vision and paradigms. Transactions on emerging telecommunications technologies. 2017
Jan;28(1):e2931.
60. Syed AS, Sierra-Sosa D, Kumar A, Elmaghraby A. IoT in smart cities: A survey of technologies,
practices and challenges. Smart Cities. 2021 Mar 30;4(2):429-75.
61. Mohanty SP, Choppali U, Kougianos E. Everything you wanted to know about smart cities: The
Internet of things is the backbone. IEEE Consumer Electronics Magazine. 2016 Jul;5(3):60-70.
62. Belli L, Cilfone A, Davoli L, Ferrari G, Adorni P, Di Nocera F, Dall’Olio A, Pellegrini C,
Mordacci M, Bertolotti E. IoT-enabled smart sustainable cities: Challenges and approaches. Smart
Cities. 2020 Sep 18;3(3):1039-71.
63. Talari S, Shafie-Khah M, Siano P, Loia V, Tommasetti A, Catalão JP. A review of smart cities
based on the internet of things concept. Energies. 2017 Mar 23;10(4):421.
64. Sánchez-Corcuera R, Nuñez-Marcos A, Sesma-Solance J, Bilbao-Jayo A, Mulero R, Zulaika U,
Azkune G, Almeida A. Smart cities survey: Technologies, application domains and challenges for
the cities of the future. International Journal of Distributed Sensor Networks. 2019
Jun;15(6):1550147719853984.
65. Sarker IH. Ai-based modeling: Techniques, applications and research issues towards automation,
intelligent and smart systems. SN Computer Science. 2022 Mar;3(2):158.
66. Li L, Rong S, Wang R, Yu S. Recent advances in artificial intelligence and machine learning for
nonlinear relationship analysis and process control in drinking water treatment: A review.
Chemical Engineering Journal. 2021 Feb 1;405:126673.
67. Weichert D, Link P, Stoll A, Rüping S, Ihlenfeldt S, Wrobel S. A review of machine learning for
the optimization of production processes. The International Journal of Advanced Manufacturing
Technology. 2019 Oct;104(5-8):1889-902.
68. Sahoo AK, Pradhan C, Das H. Performance evaluation of different machine learning methods and
deep-learning based convolutional neural network for health decision making. Nature inspired
computing for data science. 2020:201-12.
69. Mathavan S, Kamal K, Rahman M. A review of three-dimensional imaging technologies for
pavement distress detection and measurements. IEEE Transactions on Intelligent Transportation
Systems. 2015 May 12;16(5):2353-62.
70. Appel AP, Candello H, Gandour FL. Cognitive computing: Where big data is driving us.
Handbook of Big Data Technologies. 2017:807-50.
71. Roumeliotis KI, Tselikas ND. ChatGPT and Open-AI Models: A Preliminary Review. Future
Internet. 2023 May 26;15(6):192.
72. Chua CK, Leong KF. 3D Printing and additive manufacturing: Principles and applications (with
companion media pack)-of rapid prototyping. World Scientific Publishing Company; 2014 Aug 6.
73. Xie Z, Gao M, Lobo AO, Webster TJ. 3D bioprinting in tissue engineering for medical
applications: the classic and the hybrid. Polymers. 2020 Jul 31;12(8):1717.
74. Gandomi A, Haider M. Beyond the hype: Big data concepts, methods, and analytics. International
journal of information management. 2015 Apr 1;35(2):137-44.
75. Manyika J, Chui M, Brown B, Bughin J, Dobbs R, Roxburgh C, Hung Byers A. Big data: The next
frontier for innovation, competition, and productivity.
76. Gorunescu F. Data Mining: Concepts, models and techniques. Springer Science & Business Media;
2011 Mar 10.
77. Raja Santhi A, Muthuswamy P. Influence of blockchain technology in manufacturing supply chain
and logistics. Logistics. 2022 Feb 8;6(1):15.
78. Chang SE, Chen YC, Lu MF. Supply chain re-engineering using blockchain technology: A case of
smart contract based tracking process. Technological Forecasting and Social Change. 2019 Jul
1;144:1-1.
79. Shawish A, Salama M. Cloud computing: paradigms and technologies. InInter-cooperative
collective intelligence: Techniques and applications 2013 Aug 14 (pp. 39-67). Berlin, Heidelberg:
Springer Berlin Heidelberg.
100. Ehsan U, Riedl MO. Human-centered explainable ai: Towards a reflective sociotechnical
approach. InHCI International 2020-Late Breaking Papers: Multimodality and Intelligence: 22nd
HCI International Conference, HCII 2020, Copenhagen, Denmark, July 19–24, 2020, Proceedings
22 2020 (pp. 449-466). Springer International Publishing.
101. Li S, An J, Li Y, Zhu X, Zhao D, Wang L, Sun Y, Yang Y, Bi C, Zhang X, Wang M. Automated
high-throughput genome editing platform with an AI learning in situ prediction model. Nature
Communications. 2022 Nov 30;13(1):7386.
102. Giner G, Ikram S, Herold MJ, Papenfuss AT. A systematic review of computational methods for
designing efficient guides for CRISPR DNA base editor systems. Briefings in Bioinformatics.
2023 Jun 7:bbad205.