Update README.md
Browse files
README.md
CHANGED
@@ -36,17 +36,57 @@ model-index:
|
|
36 |
---
|
37 |
# Soft-Actor-Critic: Walker2d-v2
|
38 |
|
39 |
-
These are 25 trained models over **seeds (0-4)**
|
40 |
|
41 |
## Model Sources
|
42 |
|
43 |
**Repository:** [https://github.com/dee0512/Sequence-Reinforcement-Learning](https://github.com/dee0512/Sequence-Reinforcement-Learning)
|
44 |
**Paper (ICLR):** [https://openreview.net/forum?id=w3iM4WLuvy](https://openreview.net/forum?id=w3iM4WLuvy)
|
45 |
-
**Arxiv:** [
|
46 |
|
47 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
48 |
|
|
|
49 |
Using the repository:
|
50 |
|
51 |
-
```
|
52 |
-
python
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
36 |
---
|
37 |
# Soft-Actor-Critic: Walker2d-v2
|
38 |
|
39 |
+
These are 25 trained models over **seeds (0-4)** and **J = 1, 2, 4, 8, 16** of **Soft actor critic** agent playing **Walker2d-v2** for **[Sequence Reinforcement Learning (SRL)](https://github.com/dee0512/Sequence-Reinforcement-Learning)**.
|
40 |
|
41 |
## Model Sources
|
42 |
|
43 |
**Repository:** [https://github.com/dee0512/Sequence-Reinforcement-Learning](https://github.com/dee0512/Sequence-Reinforcement-Learning)
|
44 |
**Paper (ICLR):** [https://openreview.net/forum?id=w3iM4WLuvy](https://openreview.net/forum?id=w3iM4WLuvy)
|
45 |
+
**Arxiv:** [arxiv.org/pdf/2410.08979](https://arxiv.org/pdf/2410.08979)
|
46 |
|
47 |
+
# Training Details:
|
48 |
+
Using the repository:
|
49 |
+
|
50 |
+
```
|
51 |
+
python .\train_sac.py --env_name <env_name> --seed <seed> --j <j>
|
52 |
+
```
|
53 |
+
|
54 |
+
# Evaluation:
|
55 |
|
56 |
+
Download the models folder and place it in the same directory as the cloned repository.
|
57 |
Using the repository:
|
58 |
|
59 |
+
```
|
60 |
+
python .\eval_sac.py --env_name <env_name> --seed <seed> --j <j>
|
61 |
+
```
|
62 |
+
|
63 |
+
## Metrics:
|
64 |
+
|
65 |
+
**FAS:** Frequency Averaged Score
|
66 |
+
**j:** Action repetition parameter
|
67 |
+
|
68 |
+
|
69 |
+
# Citation
|
70 |
+
|
71 |
+
The paper can be cited with the following bibtex entry:
|
72 |
+
|
73 |
+
## BibTeX:
|
74 |
+
|
75 |
+
```
|
76 |
+
@inproceedings{DBLP:conf/iclr/PatelS25,
|
77 |
+
author = {Devdhar Patel and
|
78 |
+
Hava T. Siegelmann},
|
79 |
+
title = {Overcoming Slow Decision Frequencies in Continuous Control: Model-Based
|
80 |
+
Sequence Reinforcement Learning for Model-Free Control},
|
81 |
+
booktitle = {The Thirteenth International Conference on Learning Representations,
|
82 |
+
{ICLR} 2025, Singapore, April 24-28, 2025},
|
83 |
+
publisher = {OpenReview.net},
|
84 |
+
year = {2025},
|
85 |
+
url = {https://openreview.net/forum?id=w3iM4WLuvy}
|
86 |
+
}
|
87 |
+
```
|
88 |
+
|
89 |
+
## APA:
|
90 |
+
```
|
91 |
+
Patel, D., & Siegelmann, H. T. Overcoming Slow Decision Frequencies in Continuous Control: Model-Based Sequence Reinforcement Learning for Model-Free Control. In The Thirteenth International Conference on Learning Representations.
|
92 |
+
```
|