Adversarial IRL on Re-routing¶
Classes and Functions¶
- cyber_train_airl_ieee123.evaluate_policy(env, model)¶
Evaluating the learned policy by running experiments in the environment
- Parameters
cpenv (Gym.Env) – The cyber RL environment
model (torch.nn.Module) – Trained policy network model
- Returns
average episode length, average reward
- Return type
float
- cyber_train_airl_ieee123.train_and_evaluate(channel_bws, router_qlimits, policy_net_train_len, airl_train_len, exp_trajectory_len)¶
For different combination of channel bandwidths and router queue limits and expert demonstrations, train and test the policy and saves the results, reward and policy network.
- Parameters
channel_bws (list) – List of the channel bandwidths value considered in the communication network
router_qlimits (list) – List of the router queue upper bound considered in the network
policy_net_train_len (int) – Samples considered for training the policy network fed in the generator network as initial policy
airl_train_len (int) – Samples considered for training AIRL network
exp_tajectory_len (int) – Number of expert demonstration samples to be considered
- Returns
Nothing
- Return type
None