Emergent Collective Intelligence for Adaptable and Resilient
Emergent Collective Intelligence for Adaptable and Resilient
1. Working Title
2. Introduction
Floods are among the most devastating and recurrent natural disasters, causing widespread destruction,
economic losses, and displacement of populations. In 2023 alone, floods led to over $50 billion USD in
economic damages and displaced more than 20 million people worldwide. Beyond economic implications,
floods result in severe humanitarian crises, including food and water shortages, infrastructure collapse, and
increased health risks due to waterborne diseases. Effective disaster response in such crises requires rapid,
coordinated interventions to facilitate search, rescue, and relief operations. However, traditional human-
centric response strategies often prove inefficient and ineffective, particularly in highly dynamic
environments where unpredictable water currents, unstable structures, and limited visibility hinder rescue
efforts.
The integration of autonomous multi-robot systems into disaster response has emerged as a promising
solution to address these challenges. Robotic systems can operate in hazardous conditions, reducing the risks
1
PhD Research Proposal – Anyeh Ndi-Tah
faced by human responders while improving speed and efficiency in disaster mitigation. These systems,
comprising Unmanned Aerial Vehicles (UAVs), Unmanned Ground Vehicles (UGVs), and Autonomous
Underwater Vehicles (AUVs), provide faster reconnaissance, enhanced victim detection, and improved
resource distribution. UAVs can offer real-time aerial mapping of flooded regions, UGVs can navigate
through debris and transport essential supplies, while AUVs can be deployed to monitor submerged areas
and assess underwater damage. The heterogeneity of these robots enables them to perform complementary
roles, making multi-robot coordination a crucial aspect of effective deployment. However, ensuring seamless
collaboration among such diverse robotic agents in disaster-stricken environments remains a significant
challenge due to communication constraints, unpredictable terrain, and fluctuating environmental
conditions.
Traditionally, centralized control mechanisms have been employed for multi-robot coordination, where a
single control unit assigns tasks and processes incoming data. While effective in controlled environments,
centralized approaches are highly susceptible to failure in disaster scenarios where network infrastructure is
unstable or completely unavailable. Centralized decision-making also introduces computational bottlenecks,
limiting the system’s ability to scale efficiently when deployed in large-scale disaster response operations. In
contrast, decentralized coordination frameworks offer greater resilience and adaptability, allowing robots to
make localized decisions based on real-time environmental feedback. By reducing dependence on a single
decision-making entity, decentralized systems improve fault tolerance, ensuring that failures in individual agents
do not compromise the entire operation.
The concept of Emergent Collective Intelligence (ECI) further enhances decentralized coordination by
enabling robotic agents to self-organize, adapt, and collaborate autonomously. Inspired by biological swarm
behavior, ECI allows robots to operate under minimal supervision, leveraging local interactions and shared
learning to achieve global optimization in response efforts. Recent advancements in Multi-Agent
Reinforcement Learning (MARL) and Collective Intelligence (CI) have provided new opportunities to develop
efficient, scalable robotic coordination strategies that align with real-world disaster response needs.
Among these advancements, Deep Reinforcement Learning (DRL) has emerged as a powerful tool for
teaching autonomous agents adaptive behaviors in unpredictable environments. DRL enables robotic agents
to learn through trial and error, improving their ability to navigate obstacles, allocate resources, and
cooperate with other agents over time. However, despite its potential, existing MARL-based approaches still
face significant challenges, such as:
1. Facilitating real-time cooperation among heterogeneous robots with distinct roles, capabilities, and
sensory inputs.
2. Developing scalable learning models that can generalize across different flood disaster scenarios
while minimizing computational overhead.
2
PhD Research Proposal – Anyeh Ndi-Tah
To address these challenges, this research explores the integration of Deep Multi-Agent Reinforcement
Learning (DMARL) with Emergent Collective Intelligence (ECI) to enhance the resilience, adaptability, and
efficiency of robotic coordination in flood disaster scenarios. By leveraging self-learning capabilities,
decentralized decision-making, and multi-agent collaboration, this study aims to develop a robust, scalable
framework capable of functioning effectively in complex, real-world disaster environments.
Flood disasters present one of the most pressing humanitarian and economic challenges globally, necessitating
efficient and scalable response mechanisms. Traditional disaster response systems, which rely heavily on
human intervention and centralized decision-making, often struggle with communication bottlenecks,
adaptability issues, and resource allocation inefficiencies. As disasters unfold, rapidly evolving conditions,
such as shifting floodwater levels, collapsing infrastructure, and unpredictable weather patterns, render
rigid control mechanisms inadequate. In such high-risk environments, autonomous multi-robot systems
provide a promising solution by enhancing response efficiency, improving situational awareness, and
minimizing human exposure to dangerous conditions. However, the lack of robust coordination strategies
among heterogeneous robotic agents remains a significant barrier to achieving fully autonomous and
resilient disaster response frameworks.
This research seeks to bridge the gap between theory and real-world application in autonomous robotic
disaster response by developing a scalable, resilient, and decentralized coordination framework for
heterogeneous multi-robot teams. By leveraging Emergent Collective Intelligence (ECI) and Deep Multi-
Agent Reinforcement Learning (DMARL), the study aims to contribute to the following key areas:
3
PhD Research Proposal – Anyeh Ndi-Tah
Research Aim
To achieve these objectives, this research proposes a novel Emergent Collective Intelligence (ECI)
framework underpinned by Deep Multi-Agent Reinforcement Learning (DMARL). By integrating swarm
intelligence principles, cooperative reinforcement learning, and decentralized control mechanisms, the
study seeks to:
• Develop an adaptive MARL framework that enables robots to dynamically form and reconfigure
teams based on environmental conditions and mission requirements.
• Introduce novel reward mechanisms to foster sustainable cooperative behaviors, ensuring long-
term efficiency in multi-robot coordination.
• Design a robust communication strategy to mitigate the effects of partial observability, intermittent
connectivity, and uncertainty in large-scale robotic operations.
• Validate the proposed model through comprehensive simulations and real-world field tests,
ensuring its practical applicability in flood disaster response scenarios.
By addressing these challenges, this research will contribute to the next generation of autonomous robotic
systems, offering scalable, resilient, and adaptable solutions for large-scale disaster response operations.
4
PhD Research Proposal – Anyeh Ndi-Tah
Research Problem
The increasing severity and frequency of flood disasters necessitate the development of autonomous,
intelligent, and scalable response solutions that can function effectively in dynamic and high-risk
environments. Traditional disaster response systems, while effective in controlled conditions, struggle to scale
efficiently in the face of widespread infrastructural failures, unpredictable flood patterns, and
communication breakdowns. The integration of heterogeneous multi-robot teams, comprising Unmanned
Aerial Vehicles (UAVs), Unmanned Ground Vehicles (UGVs), and Autonomous Underwater Vehicles
(AUVs), presents a viable alternative to traditional response mechanisms. However, existing centralized
coordination approaches introduce significant bottlenecks, single-point failures, and computational
inefficiencies, making them unsuitable for real-time disaster response.
This research addresses the fundamental challenge of decentralized multi-robot coordination by leveraging
Emergent Collective Intelligence (ECI) and Deep Multi-Agent Reinforcement Learning (DMARL). The core
research question guiding this study is:
"How can Emergent Collective Intelligence, powered by Deep Multi-Agent Reinforcement Learning,
enable decentralized, resilient, and adaptive coordination of heterogeneous multi-robot teams in flood
disaster response?"
To explore this problem comprehensively, the study is structured around the following core research
questions:
a. How can communication constraints and partial observability be mitigated to ensure effective
decentralized decision-making in real-world disaster response scenarios?
b. What role does distributed sensor fusion play in enhancing multi-agent perception and
situational awareness?
4. Comparative Performance Analysis:
a. How does the proposed DMARL-ECI framework compare to traditional centralized and
heuristic-based coordination approaches in terms of scalability, resilience, and operational
efficiency?
b. What are the key performance trade-offs between centralized and decentralized multi-agent
reinforcement learning models in flood disaster response?
By addressing these critical research questions, this study aims to develop an intelligent, scalable, and
resilient multi-robot coordination framework that enhances the effectiveness of flood disaster response
operations.
This research is founded on the principle that Emergent Collective Intelligence (ECI) can serve as a superior
alternative to traditional centralized control mechanisms in robotic disaster response. By leveraging Deep
Multi-Agent Reinforcement Learning (DMARL), robotic agents can learn from experience, adapt to evolving
conditions, and optimize their actions collectively without requiring constant human intervention. This study
seeks to design a fully decentralized decision-making framework that enhances real-time adaptability,
scalability, and robustness in multi-robot coordination for flood disaster scenarios.
Research Objectives
4. Evaluate performance metrics such as response efficiency, adaptability, and resilience through
extensive simulations and real-world testing.
Expected Contributions
This research is expected to make significant theoretical and practical contributions to the field of
autonomous disaster response:
By achieving these contributions, this research will set a foundation for the next generation of autonomous
disaster response systems, significantly improving real-time coordination, efficiency, and resilience in flood
disaster management.
3. Literature Review
The field of multi-robot systems for disaster response has seen significant advancements in
recent years, driven by the growing need for autonomous, efficient, and scalable solutions in
extreme environments. Flood disasters, in particular, present unique challenges that require
coordinated efforts among heterogeneous robotic agents. Multi-robot coordination strategies
have evolved from centralized control systems to decentralized, emergent intelligence
models, with a particular emphasis on leveraging Deep Multi-Agent Reinforcement Learning
(DMARL) for improved adaptability and resilience.
7
PhD Research Proposal – Anyeh Ndi-Tah
This literature review explores the existing research landscape, evaluating both traditional
approaches and cutting-edge methodologies in multi-robot coordination. It also identifies key
limitations, gaps, and challenges in the field while highlighting how this research aims to
address them.
Multi-robot systems have emerged as a viable solution for disaster response operations such as
search and rescue, environmental mapping, and victim localization. Recent studies in
distributed robotic control (Cao et al., 2021) underscore the necessity of scalable and robust
architectures that can function autonomously in high-risk environments. Multi-robot systems
can be classified based on functionality and deployment strategies, with particular attention to
heterogeneous robotic systems.
Heterogeneous robotic teams, consisting of aerial, ground, and underwater robots, play
complementary roles in disaster management:
8
PhD Research Proposal – Anyeh Ndi-Tah
• Limited real-time adaptability: Many traditional planning-based strategies are rigid and
fail to adjust dynamically to rapidly changing disaster conditions.
• Vulnerability to single-point failures: Centralized control architectures suffer from
bottlenecks and system-wide failures if the central unit is compromised.
• High computational costs: Real-time coordination in large-scale disaster environments
requires significant computational resources, making deployment infeasible in certain
scenarios.
These limitations have driven the shift towards decentralized coordination strategies, which
enhance scalability, fault tolerance, and real-time adaptability.
Reinforcement Learning (RL) is a machine learning paradigm where agents learn optimal
policies by interacting with their environment and receiving feedback in the form of rewards or
penalties. Recent advancements in Multi-Agent Reinforcement Learning (MARL) have
demonstrated improved decision-making, adaptability, and cooperation in high-complexity
environments (Foerster et al., 2020).
1. Independent Q-learning (IQL): Each agent learns independently, but this approach
struggles with non-stationarity, as the environment continuously changes due to the
actions of other agents.
9
PhD Research Proposal – Anyeh Ndi-Tah
2. Centralized Training with Decentralized Execution (CTDE): Agents are trained using a
centralized model but make independent decisions during execution, balancing efficiency
and adaptability.
3. Hierarchical Reinforcement Learning (HRL): Learning is divided into multiple
hierarchical levels, enabling long-term planning and strategic decision-making.
• Local Interaction Rules: Instead of following a central command, agents rely on local
interactions to achieve global objectives.
10
PhD Research Proposal – Anyeh Ndi-Tah
• Adaptive Policies: Agents learn and evolve based on real-time environmental feedback,
enhancing long-term cooperation.
• Decentralized Decision-Making: By distributing control, the system becomes resilient to
partial failures and improves overall system robustness.
• Lack of real-world validation: Most ECI research is simulation-based, with limited real-
world disaster response applications.
• Integration with DMARL: Few studies have explored the synergy between emergent
intelligence and deep reinforcement learning for disaster management.
This research seeks to bridge these gaps by integrating DMARL-driven emergent intelligence
for scalable, autonomous, and adaptive robotic coordination in flood disaster scenarios.
Traditional reinforcement learning models such as Q-learning and SARSA face scalability
limitations, especially in environments with high-dimensional state spaces. The development of
Deep Reinforcement Learning (DRL) has led to superior performance in handling complex
robotic tasks.
• Deep Q-Networks (DQN): Utilizes deep neural networks to approximate Q-values, enabling
agents to make better decisions in large action spaces.
• Proximal Policy Optimization (PPO): Balances exploration and exploitation, improving
the stability of policy learning.
• Soft Actor-Critic (SAC): Optimizes both exploration efficiency and learning stability,
making it well-suited for highly uncertain disaster environments.
11
PhD Research Proposal – Anyeh Ndi-Tah
While research in MARL, ECI, and DMARL has made substantial progress, several limitations
remain:
1. Developing novel reward functions that balance short-term and long-term cooperation
incentives, ensuring sustained coordination among robotic agents.
2. Introducing Hierarchical Reinforcement Learning (HRL) to enhance multi-level
decision-making, improving policy optimization in complex disaster scenarios.
3. Implementing real-world-inspired simulations that integrate realistic flood dynamics,
environmental uncertainty, and heterogeneous robot interactions to test the
adaptability of the proposed framework.
By addressing these research gaps, this study contributes to the development of a scalable,
intelligent, and resilient robotic coordination framework, advancing the next generation of
AI-driven disaster response solutions.
12
PhD Research Proposal – Anyeh Ndi-Tah
4. Methodology
The methodology of this research is designed to ensure rigorous validation of the proposed
Emergent Collective Intelligence (ECI) framework integrated with Deep Multi-Agent
Reinforcement Learning (DMARL) for multi-robot coordination in flood disaster response.
This study employs a mixed-methods approach, incorporating simulation-based experiments,
real-world testing, and comparative analysis to evaluate the system’s effectiveness in
scalability, adaptability, resilience, and efficiency.
1. Simulation-Based Experiments:
a. Initial training and evaluation of MARL-based coordination strategies will be
conducted in virtual flood environments.
b. Simulated multi-robot teams (including UAVs, UGVs, and AUVs) will be deployed in
flood scenarios generated using SUMO, OpenAI Gym, and Gazebo.
2. Implementation of Reinforcement Learning Frameworks:
a. The study will integrate Deep Q-Networks (DQN), Proximal Policy Optimization
(PPO), and Multi-Agent Deep Deterministic Policy Gradients (MADDPG) for
policy learning and optimization.
b. Centralized Training with Decentralized Execution (CTDE) will be used to train
models in a structured environment while allowing real-world decentralized
decision-making.
3. Real-World Deployment in Controlled Flood Scenarios:
13
PhD Research Proposal – Anyeh Ndi-Tah
The combination of simulation-based testing and real-world trials ensures that the model is
both theoretically sound and practically applicable, providing a robust evaluation of its
capabilities in flood disaster response.
• OpenAI Gym: A reinforcement learning platform for developing and testing MARL-based
policies.
• SUMO (Simulation of Urban Mobility): Used for simulating urban environments affected
by floods, including road blockages, changing water levels, and traffic obstructions.
• Gazebo: A realistic 3D robotics simulator for testing autonomous robotic behavior in
complex flood scenarios.
The research will employ Deep Multi-Agent Reinforcement Learning (DMARL) techniques to
optimize team-based coordination strategies. Key algorithms include:
• Deep Deterministic Policy Gradient (DDPG): Used for continuous control tasks,
enabling smooth and adaptive decision-making in dynamic environments.
• Multi-Agent Proximal Policy Optimization (MAPPO): Enhances policy robustness and
scalability, ensuring stability in multi-agent learning.
14
PhD Research Proposal – Anyeh Ndi-Tah
• Centralized Training with Decentralized Execution (CTDE): Allows agents to train under
controlled conditions while executing decentralized actions in real-world flood
response scenarios.
The simulation environment will include a heterogeneous fleet of robots, each designed for
specific flood disaster response tasks:
By simulating realistic flood environments with heterogeneous robotic agents, this study
ensures that MARL-based coordination strategies are optimized for real-world disaster
conditions.
The initial phase of experimentation will focus on developing and refining the reinforcement
learning models in simulated environments:
• Flood disaster scenarios will be designed with dynamic water levels, terrain obstacles,
and changing weather conditions.
• Robotic teams will be trained in OpenAI Gym and Gazebo using multi-agent
reinforcement learning.
15
PhD Research Proposal – Anyeh Ndi-Tah
Once the model demonstrates stable and reliable performance in simulations, it will be
deployed in controlled flood environments with:
The reinforcement learning framework will be implemented using PyTorch, a widely used deep
learning library. The model training process will involve:
To assess the effectiveness and robustness of the proposed coordination framework, the
following performance metrics will be used:
4.4.1 Scalability
16
PhD Research Proposal – Anyeh Ndi-Tah
4.4.2 Adaptability
17
PhD Research Proposal – Anyeh Ndi-Tah
• Ensuring that robots operate within predefined safety parameters to prevent harm to
human responders or civilians.
• Establishing emergency override mechanisms to allow human intervention when
necessary.
• Ensuring that the deployment of robotic teams adheres to local and international disaster
management policies.
• Collaboration with emergency response agencies to integrate robotic operations with
existing disaster relief efforts.
By addressing these ethical considerations, this research ensures that autonomous disaster
response solutions are developed responsibly and in compliance with regulatory standards.
Conclusion
This methodology section outlines a structured approach for developing and validating the
ECI-based multi-robot coordination framework. By combining simulation-based learning with
real-world experimentation, the study aims to establish a scalable, adaptable, and resilient
robotic response system for flood disaster management. The research’s rigorous evaluation
framework, grounded in scalability, adaptability, resilience, and ethical compliance, ensures
18
PhD Research Proposal – Anyeh Ndi-Tah
that the proposed solution is both theoretically sound and practically viable for real-world
disaster response applications.
5. Timescales
The research will be conducted over a structured timeline, ensuring a systematic approach to
developing, testing, and evaluating the Deep Multi-Agent Reinforcement Learning (DMARL)
framework for heterogeneous multi-robot coordination in flood disaster response. The study
is divided into six key phases, each designed to build upon the findings of the previous phase
while ensuring efficient progress toward the final research objectives.
The proposed timeline spans approximately three to four years, considering the complexity of
model development, testing, and validation. Below is a detailed breakdown of the timescales,
including estimated durations and key activities associated with each phase.
19
PhD Research Proposal – Anyeh Ndi-Tah
20
PhD Research Proposal – Anyeh Ndi-Tah
The proposed timeline reflects the complexity and depth of the research, ensuring a rigorous
and comprehensive approach to developing and validating the proposed framework.
While the research follows a structured timeline, several challenges may arise during different
phases. Below are some anticipated risks and their corresponding mitigation strategies:
21
PhD Research Proposal – Anyeh Ndi-Tah
5.4 Conclusion
This timeline ensures that the study remains feasible, well-paced, and aligned with high
academic and practical standards, facilitating the successful deployment of AI-driven robotic
systems for flood disaster response.
22
PhD Research Proposal – Anyeh Ndi-Tah
The proposed study is expected to yield several key findings that will enhance the efficiency and
effectiveness of autonomous multi-robot coordination in disaster environments. These
anticipated outcomes include:
The study will produce a multi-agent reinforcement learning (MARL) framework that enables
heterogeneous robotic teams to coordinate autonomously in high-risk flood environments. By
shifting from centralized control to decentralized emergent intelligence, the framework will
improve fault tolerance, adaptability, and scalability.
By integrating Emergent Collective Intelligence (ECI) principles, the system will allow robotic
agents to self-organize and adapt to rapidly changing conditions, such as shifting floodwaters,
infrastructure collapses, and unpredictable terrain. The model will showcase higher flexibility
than traditional pre-programmed strategies.
The research will develop and evaluate optimized reward structures to reinforce cooperative
and goal-oriented behaviors among robotic agents. By balancing short-term efficiency with
23
PhD Research Proposal – Anyeh Ndi-Tah
long-term sustainability, the system will foster more reliable and effective teamwork among
autonomous robots.
One of the critical findings will be the ability of agents to collectively develop intelligent
behaviors without requiring explicit programming. By allowing agents to learn through
continuous interaction and environmental feedback, the study will validate the emergence of
self-organized decision-making patterns.
The proposed framework will enhance the resilience of robotic teams by enabling them to
recover from failures, whether caused by communication breakdowns, hardware
malfunctions, or unexpected environmental disruptions. This adaptability will ensure
continuous operation even under challenging real-world conditions.
Although the primary focus is flood disaster response, the proposed framework will be
generalizable to other disaster scenarios, including earthquakes, wildfires, and industrial
accidents. The adaptability of MARL-based decentralized coordination will be applicable to
various high-risk environments, expanding the potential utility of the system.
The research has the potential to transform disaster response methodologies by introducing
AI-driven autonomous coordination into emergency management. The broader implications
span academia, industry, and policymaking, with long-term benefits for autonomous systems
research and real-world disaster mitigation efforts.
24
PhD Research Proposal – Anyeh Ndi-Tah
1. Academic Contributions
• The study will advance research in MARL, reinforcement learning, and emergent
intelligence for real-world applications.
• It will contribute to AI-driven disaster response literature, providing a new paradigm for
autonomous robotic systems in emergency environments.
• The results will be disseminated through peer-reviewed publications, conferences, and
open-source frameworks to benefit future research.
2. Industrial Applications
• The findings will support the deployment of intelligent robotic teams for disaster
response operations, improving response times, efficiency, and safety.
• The proposed framework could be adapted for use in search-and-rescue missions,
environmental monitoring, and automated crisis intervention.
• The technology could be commercialized for disaster relief organizations, emergency
response agencies, and private security firms.
3. Policy Implications
• The research will provide data-driven insights for government agencies and disaster
management organizations to develop more effective deployment strategies for robotic
emergency response teams.
• It will influence future policy discussions on the integration of AI-driven disaster
response technologies, ensuring regulatory compliance and ethical deployment.
• The framework may assist in establishing international standards for robotic intervention
in disaster scenarios, setting a precedent for future AI-driven relief strategies.
Conclusion
This research will not only advance the theoretical understanding of MARL and emergent
intelligence in disaster response but will also provide a practical, scalable, and resilient
25
PhD Research Proposal – Anyeh Ndi-Tah
References
[1] Y. Du, H. Ding, and J. Xu, "A survey on multi-agent deep reinforcement learning: Methods,
challenges and applications," Neurocomputing, vol. 502, pp. 92-107, 2022. DOI:
10.1016/j.neucom.2021.09.067.
[2] Z. Zhou, Y. Guo, and J. Zhang, "Multi-agent reinforcement learning: State of the art and
challenges," IEEE Transactions on Artificial Intelligence, vol. 3, no. 1, pp. 1-17, 2022. DOI:
10.1109/TAI.2021.3129059.
[4] W. Li, H. Zhang, and B. Li, "Multi-agent reinforcement learning for cooperative control: A review,"
IEEE Transactions on Neural Networks and Learning Systems, vol. 34, no. 4, pp. 1432-1447, 2023. DOI:
10.1109/TNNLS.2022.3176524.
26
PhD Research Proposal – Anyeh Ndi-Tah
[6] R. Garnier, S. Jouffrais, and P. Gaussier, "Self-organizing emergent intelligence for robotic teams in
dynamic environments," Robotics and Autonomous Systems, vol. 161, pp. 104376, 2023. DOI:
10.1016/j.robot.2023.104376.
[7] A. S. Abdullahi, M. Al-Saqqa, and C. Talbot, "Swarm intelligence-based reinforcement learning for
decentralized robotic coordination," Expert Systems with Applications, vol. 206, pp. 117591, 2022. DOI:
10.1016/j.eswa.2022.117591.
[8] F. H. Khan, M. K. Hussain, and S. Ahmed, "Decentralized emergent intelligence for autonomous
robotic collaboration," IEEE Access, vol. 10, pp. 18672-18685, 2022. DOI:
10.1109/ACCESS.2022.3145896.
[9] Chen, H., et al. (2023). Emergent collective intelligence from massive-agent cooperation and
competition. arXiv:2301.01609.
[10] A. P. Mohanty, S. K. Choudhury, and P. N. Suganthan, "Multi-robot systems for search and rescue
operations: A review," IEEE Transactions on Cybernetics, vol. 52, no. 3, pp. 1568-1584, 2022. DOI:
10.1109/TCYB.2021.3059389.
[11] R. Ahmed and J. L. M. Lastra, "Aerial-ground collaborative robotics for disaster response: A
survey," Robotics and Autonomous Systems, vol. 158, pp. 104173, 2022. DOI:
10.1016/j.robot.2022.104173.
[12] M. A. Bais and M. Azam, "Reinforcement learning-driven coordination strategies for multi-robot
disaster management," Journal of Intelligent & Robotic Systems, vol. 106, pp. 31-50, 2022. DOI:
10.1007/s10846-022-01502-3.
[13] C. Feng, D. M. Graham, and P. Yan, "Human-robot teaming in disaster response: Challenges and
solutions," IEEE Robotics & Automation Magazine, vol. 29, no. 2, pp. 56-67, 2022. DOI:
10.1109/MRA.2022.3145878.
[14] Kouzeghar, M., et al. (2023). Multi-Target Pursuit by a Decentralized Heterogeneous UAV Swarm using
Deep Multi-Agent Reinforcement Learning. arXiv:2303.01799.
27
PhD Research Proposal – Anyeh Ndi-Tah
[15] Howell, P., et al. (2023). Generalization of Heterogeneous Multi-Robot Policies via Awareness and
Communication of Capabilities. OpenReview.
[16] Anonymous. (2024). Graph Neural Network-based Multi-agent Reinforcement Learning for Resilient
Coordination. arXiv:2403.13093.
[17] D. Zhang, B. Yang, and X. Chen, "Deep reinforcement learning for dynamic multi-agent
collaboration: A survey," IEEE Transactions on Neural Networks and Learning Systems, vol. 34, no. 1, pp.
27-45, 2023. DOI: 10.1109/TNNLS.2022.3204359.
[18] X. Ma, H. Wang, and J. Liu, "Deep reinforcement learning for multi-agent cooperation: Challenges
and future directions," Neural Networks, vol. 161, pp. 136-149, 2023. DOI: 10.1016/j.neunet.2022.07.014.
[19] H. Kim, C. Lee, and S. Park, "Multi-agent reinforcement learning with shared experience for
cooperative task execution," IEEE Transactions on Cybernetics, vol. 52, no. 4, pp. 2519-2532, 2022. DOI:
10.1109/TCYB.2021.3074568.
[20] Sartoretti, G., et al. (2019). Distributed Reinforcement Learning for Multi-robot Decentralized Collective
Construction. In Distributed Autonomous Robotic Systems (pp. 35–49). Springer.
[21] L. Smith and J. D. Lee, "Ethical concerns in AI-driven autonomous disaster response systems," AI
& Society, vol. 38, pp. 75-92, 2023. DOI: 10.1007/s00146-022-01388-9.
[22] C. T. Nguyen and T. H. Tran, "Secure and reliable communication in multi-agent robotic teams,"
IEEE Communications Surveys & Tutorials, vol. 24, no. 2, pp. 276-298, 2022. DOI:
10.1109/COMST.2022.3145879.
[23] F. A. Rahman, M. S. Alam, and S. K. Roy, "Ethical and policy challenges in deploying AI for
emergency management," IEEE Transactions on Technology and Society, vol. 3, no. 2, pp. 96-110, 2022.
DOI: 10.1109/TTS.2022.3169887.
28