devdharpatel commited on
Commit
573035d
·
verified ·
1 Parent(s): 33cf03c

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +92 -3
README.md CHANGED
@@ -1,3 +1,92 @@
1
- ---
2
- license: bsd-3-clause
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: bsd-3-clause
3
+ tags:
4
+ - Walker2d-v2
5
+ - reinforcement-learning
6
+ - Soft Actor Critic
7
+ - SRL
8
+ - deep-reinforcement-learning
9
+ model-index:
10
+ - name: SAC
11
+ results:
12
+ - metrics:
13
+ - type: FAS (J=1)
14
+ value: 0.4419 ± 0.025996
15
+ name: FAS
16
+ - type: FAS (J=2)
17
+ value: 0.423547 ± 0.026536
18
+ name: FAS
19
+ - type: FAS (J=4)
20
+ value: 0.497902 ± 0.034868
21
+ name: FAS
22
+ - type: FAS (J=8)
23
+ value: 0.489516 ± 0.044905
24
+ name: FAS
25
+ - type: FAS (J=16)
26
+ value: 0.32623 ± 0.053239
27
+ name: FAS
28
+ task:
29
+ type: OpenAI Gym
30
+ name: OpenAI Gym
31
+ dataset:
32
+ name: Walker2d-v2
33
+ type: Walker2d-v2
34
+ Paper: https://arxiv.org/pdf/2410.08979
35
+ Code: https://github.com/dee0512/Sequence-Reinforcement-Learning
36
+ ---
37
+ # Soft-Actor-Critic: Walker2d-v2
38
+
39
+ These are 25 trained models over **seeds (0-4)** and **J = 1, 2, 4, 8, 16** of **Soft actor critic** agent playing **Walker2d-v2** for **[Sequence Reinforcement Learning (SRL)](https://github.com/dee0512/Sequence-Reinforcement-Learning)**.
40
+
41
+ ## Model Sources
42
+
43
+ **Repository:** [https://github.com/dee0512/Sequence-Reinforcement-Learning](https://github.com/dee0512/Sequence-Reinforcement-Learning)
44
+ **Paper (ICLR):** [https://openreview.net/forum?id=w3iM4WLuvy](https://openreview.net/forum?id=w3iM4WLuvy)
45
+ **Arxiv:** [arxiv.org/pdf/2410.08979](https://arxiv.org/pdf/2410.08979)
46
+
47
+ # Training Details:
48
+ Using the repository:
49
+
50
+ ```
51
+ python .\train_sac.py --env_name <env_name> --seed <seed> --j <j>
52
+ ```
53
+
54
+ # Evaluation:
55
+
56
+ Download the models folder and place it in the same directory as the cloned repository.
57
+ Using the repository:
58
+
59
+ ```
60
+ python .\eval_sac.py --env_name <env_name> --seed <seed> --j <j>
61
+ ```
62
+
63
+ ## Metrics:
64
+
65
+ **FAS:** Frequency Averaged Score
66
+ **j:** Action repetition parameter
67
+
68
+
69
+ # Citation
70
+
71
+ The paper can be cited with the following bibtex entry:
72
+
73
+ ## BibTeX:
74
+
75
+ ```
76
+ @inproceedings{DBLP:conf/iclr/PatelS25,
77
+ author = {Devdhar Patel and
78
+ Hava T. Siegelmann},
79
+ title = {Overcoming Slow Decision Frequencies in Continuous Control: Model-Based
80
+ Sequence Reinforcement Learning for Model-Free Control},
81
+ booktitle = {The Thirteenth International Conference on Learning Representations,
82
+ {ICLR} 2025, Singapore, April 24-28, 2025},
83
+ publisher = {OpenReview.net},
84
+ year = {2025},
85
+ url = {https://openreview.net/forum?id=w3iM4WLuvy}
86
+ }
87
+ ```
88
+
89
+ ## APA:
90
+ ```
91
+ Patel, D., & Siegelmann, H. T. Overcoming Slow Decision Frequencies in Continuous Control: Model-Based Sequence Reinforcement Learning for Model-Free Control. In The Thirteenth International Conference on Learning Representations.
92
+ ```