Comments (7)
can you be more specific on that?
from ris-miso-deep-reinforcement-learning.
In figure 4, the sum rate is from 5 to 35
However, in figure 6, the reward is less than 10. In my opinion, our reward should be sum rate, so max reward should equal sum rate. Can you help me to solve that issue?
from ris-miso-deep-reinforcement-learning.
yes, there's an inconsistency between the two figures. however, note that the used hyperparameters are different for these figures; otherwise, they would've produced the same results. the authors didn't provide any hyperparameter setting for such particular learning curves, and I don't remember which hyperparameter values I used to produce Fig. 6 unfortunately. I've taken a look at the paper, but still couldn't find any information.
please let me know if anything else, and if you find the used hyperparameter values for Fig. 6.
from ris-miso-deep-reinforcement-learning.
Thank you for your reply. However, there is something weird. In figure 4, when I increase the number of RIS element (N), the result is getting worse, not like what you see in your figure 4. The following is my current figure 4 based on your code.
from ris-miso-deep-reinforcement-learning.
I believe this is expected since you increased the number of users as well. increasing the number of users would degrade the performance.
from ris-miso-deep-reinforcement-learning.
Thx for your explanation. I have changed the configuration, where the only distinction is the number of RIS element (N) like the following. By the way, I consider the sum rate in figure 4 may be the opt_reward rather than current_reward. opt_reward is the SNR rather than SINR. In that case, we will get larger sum rate.
from ris-miso-deep-reinforcement-learning.
yes, this is what's expected. when you increase the number of RIS elements, you'd obtain more transmission power as well as effective performance.
regarding the SNR/SINR, thank you for pointing this out, but I'm not the author of the paper, so I only tried to reproduce the figures more or less the same. authors didn't provide much detail such as which objective (as a reward) they used, hyperparameter settings, etc.
from ris-miso-deep-reinforcement-learning.
Related Issues (20)
- Some doubts about program run time? HOT 5
- Some questions about rewards in the training process HOT 2
- Generating Fig. 4 and 5 HOT 3
- How do I generate comparison data with a trained model? HOT 1
- Questions in 'DDPG.py' and 'environment.py' HOT 1
- Help need for Figure 4 and data you generated HOT 3
- Comparison algorithm problem
- Power conditions used to create figures 8, 11 and 12 HOT 5
- .npy file for reproduction for figure no 4? HOT 1
- Variation in channel at each time step. HOT 5
- Calculation of Unit Modulus constraint.
- Transmitted power handling HOT 4
- Issue in regenerating figure 11
- On the problem of drawing reward to steps diagram HOT 6
- transmit power handing HOT 7
- I have the doubt that i can't re current the results similar to the papers. HOT 3
- Optimization constraints HOT 1
- Unit modulus constraint of RIS HOT 4
- Rewards decrease in late training HOT 11
Recommend Projects
-
React
A declarative, efficient, and flexible JavaScript library for building user interfaces.
-
Vue.js
🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
-
Typescript
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
-
TensorFlow
An Open Source Machine Learning Framework for Everyone
-
Django
The Web framework for perfectionists with deadlines.
-
Laravel
A PHP framework for web artisans
-
D3
Bring data to life with SVG, Canvas and HTML. 📊📈🎉
-
Recommend Topics
-
javascript
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
-
web
Some thing interesting about web. New door for the world.
-
server
A server is a program made to process requests and deliver data to clients.
-
Machine learning
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
-
Visualization
Some thing interesting about visualization, use data art
-
Game
Some thing interesting about game, make everyone happy.
Recommend Org
-
Facebook
We are working to build community through open source technology. NB: members must have two-factor auth.
-
Microsoft
Open source projects and samples from Microsoft.
-
Google
Google ❤️ Open Source for everyone.
-
Alibaba
Alibaba Open Source for everyone
-
D3
Data-Driven Documents codes.
-
Tencent
China tencent open source team.
from ris-miso-deep-reinforcement-learning.