Reinforcement learning should be fast, safe and easy to use.
rsrl
provides generic constructs for reinforcement learning (RL)
experiments in an extensible framework with efficient implementations of
existing methods for rapid prototyping.
toml
[dependencies]
rsrl = "0.8"
Note that rsrl
enables the blas
feature of its [ndarray
] dependency, so
if you're building a binary, you additionally need to specify a BLAS backend
compatible with ndarray
. For example, you can add these dependencies:
toml
blas-src = { version = "0.2.0", default-features = false, features = ["openblas"] }
openblas-src = { version = "0.6.0", default-features = false, features = ["cblas", "system"] }
See ndarray
's README
for more information.
The code below shows how one could use rsrl
to evaluate a QLearning agent
using a linear function approximator with Fourier basis projection to solve the
canonical mountain car problem.
See examples/ for more...
```rust let env = MountainCar::default(); let nactions = env.actionspace().card().into();
let mut rng = StdRng::seedfromu64(0); let (mut ql, policy) = { let basis = Fourier::fromspace(5, env.statespace()).withbias(); let qfunc = makeshared(LFA::vector(basis, SGD(0.001), nactions)); let policy = Greedy::new(q_func.clone());
(QLearning {
q_func,
gamma: 0.9,
}, policy)
};
for e in 0..200 { // Episode loop: let mut j = 0; let mut env = MountainCar::default(); let mut action = policy.sample(&mut rng, env.emit().state());
for i in 0.. {
// Trajectory loop:
j = i;
let t = env.transition(action);
ql.handle(&t).ok();
action = policy.sample(&mut rng, t.to.state());
if t.terminated() {
break;
}
}
println!("Batch {}: {} steps...", e + 1, j + 1);
}
let traj = MountainCar::default().rollout(|s| policy.mode(s), Some(500));
println!("OOS: {} states...", traj.n_states()); ```
Pull requests are welcome. For major changes, please open an issue first to discuss what you would like to change.
Please make sure to update tests as appropriate and adhere to the angularjs commit message conventions (see here).