Predictive Link Stability and Trust-Aware Routing in FANETs
Team
- E/20/035, K.C.H.N.A.W.M.R.C.J.N. Bandara, email
- E/20/173, P.A.S.V. Jayasooriya, email
- E/20/178, N.K.D.P. Jayawardena, email
Supervisors
Table of content
- Abstract
- Related works
- Methodology
- Experiment Setup and Implementation
- Results and Analysis
- Conclusion
- Publications
- Links
Abstract
Efficient routing in Flying Ad Hoc Networks remains a critical challenge due to time-varying link quality, node mobility, and unpredictable network topology changes. Traditional routing protocols like OLSR and AODV rely on static metrics and fixed decision rules, which often fail to adapt to rapidly changing conditions. This paper proposes QBR (Q-Learning Based Routing), a novel machine learning-driven routing protocol that leverages Q-Learning with Quantile Regression to make intelligent routing decisions on demand. Unlike conventional protocols, QBR adapts its path selection in real-time by learning from network feedback, including link quality metrics (SNR, packet loss), latency, and throughput. We evaluate QBR against established protocols (OLSR and AODV) in a simulated 40-node wireless network using NS-3. Experimental results demonstrate that QBR achieves higher Packet Delivery Ratio (PDR) and improved throughput while maintaining competitive latency, particularly in scenarios with fluctuating link quality. The findings illustrate the potential of reinforcement learning-based approaches to enhance routing performance in dynamic wireless networks.
Related works
Traditional ad-hoc routing protocols can be broadly categorized into reactive (on-demand) and proactive (table-driven) approaches. AODV and DSR are prominent reactive protocols that discover routes only when needed, while OLSR is a proactive protocol that maintains routing tables periodically. These conventional approaches typically use simple metrics such as hop count, which often prove inadequate in dynamic environments with fluctuating link quality. Recent research has explored machine learning applications in networking. Q-Learning and other reinforcement learning techniques have been successfully applied to congestion control, resource allocation, and load balancing. Some studies have investigated learning-based routing for mobile networks; however, limited work exists on distributional Q-Learning approaches that capture link quality uncertainty. Our work extends this domain by combining Q-Learning with Quantile Regression to provide risk-aware routing decisions that consider both expected performance and uncertainty in link metrics.
Methodology
QBR operates at the network layer by applying Q-Learning to continuously improve routing decisions based on network conditions. The protocol maintains a Q-Table where each state represents a combination of destination and neighboring nodes, and actions correspond to selecting the next hop for packet forwarding. Routing decisions are made using an ε-greedy strategy, which balances exploration of new paths and exploitation of known high-quality routes. The reward mechanism is designed using real-time link metrics such as signal-to-noise ratio (SNR), packet delivery success, and delay, enabling the protocol to learn from network feedback. Additionally, Quantile Regression is incorporated to estimate the distribution of Q-values, allowing the model to account for uncertainty in link quality and make risk-aware decisions. As packets are transmitted, the protocol dynamically updates its knowledge and progressively converges toward optimal routing paths under changing network conditions.
Experiment Setup and Implementation
We evaluate QBR using NS-3, a discrete-event network simulator, with the following configuration:
- Network Topology: 40 mobile nodes deployed in a 1000 m × 1000 m area
- Mobility Model: Random Waypoint Model with velocity 0.5–10 m/s and pause time 5 seconds
- Physical Layer: 802.11g WiFi with 54 Mbps data rate
- Communication: CBR (Constant Bit Rate) traffic: 10 flows, 512-byte packets, 5 packets/second
- Simulation Duration: 300 seconds
- Q-Learning Parameters: Learning rate α = 0.1, discount factor γ = 0.95, ε = 0.1 (decaying)
- Comparison Protocols: OLSR (proactive), AODV (reactive), QBR (learning-based)
- Evaluation Metrics: PDR (Packet Delivery Ratio): Percentage of successfully delivered packets Average Latency: Mean end-to-end delay Throughput: Goodput in Kbps Performance is recorded every 1 second, and results are aggregated for statistical analysis.
Results and Analysis
Conclusion
This work demonstrates that reinforcement learning-based routing can effectively address limitations of traditional routing protocols in dynamic wireless networks. QBR leverages Q-Learning and Quantile Regression to learn adaptive routing policies that improve Packet Delivery Ratio and throughput compared to OLSR and AODV. The experimental evaluation confirms that the protocol rapidly converges to optimal routing decisions as network conditions evolve. However, scalability to larger networks, computational overhead, and Q-Table memory requirements merit further investigation. Future work should explore hierarchical Q-Learning for large-scale networks, integration of multi-agent reinforcement learning for cooperative routing, and deployment on real wireless testbeds to validate simulation findings. The results suggest that machine learning approaches offer promising avenues for developing more intelligent and adaptive routing protocols in next-generation wireless networks.