Ref. | ML Technique | Multiple | Synthetic data from | Features | Output | Evaluation | |
---|---|---|---|---|---|---|---|
Bottlenecka | ns-2 simulation | (action-set for RL) | Settings | Results | |||
PAQM [160] | Supervised: · OLS | ✓ | Topology: · 6-linear · Arbitrary dumbbell Time =50s | · Traffic volume (bytes) | TSF: · Traffic volume | · NMLS algorithm based on LMMSE | Accuracy: ·90−92.3% |
APACE [212] | Supervised: · OLS | ✓ | Topology: · Dumbbell (1-sink) · 6-linear Time =40s | · Queue length | TSF: · Queue length | · NMLS algorithm based on LMMSE | Accuracy: · 92% |
α_SNFAQM [498] | Supervised: · MLP-NN | – | Topology: · Dumbbell (1-sink) Time =300s | · Traffic volume · Predicted traffic volume | TSF: · Traffic volume | · 2 input neurons · 2 hidden layers with 3 neurons · 1 output neuron | Accuracy: ·90−93% |
NN-RED [179] | Supervised: · SLP-NN | – | Topology: · Dumbbell Time =900s | · Queue length | TSF: · Queue length | · 1+N input neurons (N past values) · 0 hidden layers · 1 output neuron · Delta-rule learning | N/A |
DEEP BLUE [298] | Reinforcement: · Q-learning - ε-greedy | – | Topology: · Dumbbell Time =50sOPNET simulator instead of ns-2 | States: · Queue length · Packet drop prob. Reward: · Throughput · Queuing delay | Decision making: · Increment of the packet drop probability (finite: 6 actions) | ·N/A states · 6 actions ·ε-greedy ASSb | Optimal packet drop probability: · Outperforms BLUE [144] |
Neuron PID [428] | Reinforcement: · PIDNN | ✓ | Topology: · Dumbbell Time =100s | · Queue length error | Decision making: · Increment of the packet drop probability(continuous) | · 3 input neurons · 0 hidden layers · 1 output neuron · Hebbian learning · 1 PID component | QLAcc errorc: · 7.15 QLJit: · 20.18 |
AN-AQM [427] | Reinforcement: · PIDNN | ✓ | Topology: · Dumbbell · 6-linear Time =100s | · Queue length error · Sending rate error | Decision making: · Increment of the packet drop probability(continuous) | · 6 input neurons · 0 hidden layers · 1 output neuron · Hebbian learning · 2 PID components | QLAcc errorc: · 6.44 QLJit: · 22.61 |
FAPIDNN [485] | Reinforcement: · PIDNN | ✓ | Topology: · Dumbbell Time =60s | · Queue length error | Decision making: · Increment of the packet drop probability(continuous) | · 3 input neurons · 0 hidden layers · 1 output neuron · 1 PID component · 1 fuzzy component | QLAcc errorc: · 3.73 QLJit: · 31.8 |
NRL [499] | Reinforcement: · SLP-NN | ✓ | Topology: · Dumbbell Time =100s | · Queue length error · Sending rate error | Decision making: · Increment of the packet drop probability(continuous) | · 2 input neurons · 0 hidden layers · 1 output neuron · RL learning | QLAcc errorc: · 38.73 QLJit: · 128.84 |