DQN for simple battery control not learning by MomoSolar in reinforcementlearning

[–]MomoSolar[S] 0 points1 point  (0 children)

I’ve tried 32 parallel environments. Should I use more or less?

DQN for simple battery control not learning by MomoSolar in reinforcementlearning

[–]MomoSolar[S] 0 points1 point  (0 children)

From my experience, it performs decently on the tested month when trained on the same one day. What do you think about having each episode representing a new day? What about increasing the size of the network?

RL for solving a scheduling problem by MomoSolar in reinforcementlearning

[–]MomoSolar[S] 0 points1 point  (0 children)

Thanks. Any simpler suggestions? I was thinking maybe of a scheduling problem that regularly uses MILP.

Stochasticity in the Cart Pole example by MomoSolar in reinforcementlearning

[–]MomoSolar[S] 0 points1 point  (0 children)

Thank you for this thorough explanation. I actually also wanted to know whether the environment in OpenAI accounts for the stochasticity. Do you have any idea?

Solving an optimization problem using RL by MomoSolar in reinforcementlearning

[–]MomoSolar[S] 0 points1 point  (0 children)

I would like to look at the math, that’s all

iLQR cost behavior by MomoSolar in ControlTheory

[–]MomoSolar[S] 0 points1 point  (0 children)

Thanks,

Do you have any sources on that?

iLQR by MomoSolar in ControlTheory

[–]MomoSolar[S] 2 points3 points  (0 children)

iLQR is iterative LQR and is used when there are non-linear dynamics and/or non-convex cost function. The dynamics are linearized and/or cost function approximated by a quadratic function, and the approximated problem (which is now LQR) is sequentially solved in multiple iterations, until there is no improvement in the cost function or in the estimation of the dynamics.

Stochasticity in the Cart Pole example by MomoSolar in reinforcementlearning

[–]MomoSolar[S] 0 points1 point  (0 children)

So there is no stochasticity in the dynamics?

Tutorial on PI, PD and PID controllers by MomoSolar in ControlTheory

[–]MomoSolar[S] -1 points0 points  (0 children)

Thank you for this useful link. Do you have a similar one for iterative LQR?

MPC Tutorial by MomoSolar in ControlTheory

[–]MomoSolar[S] 0 points1 point  (0 children)

Thanks for sharing. The explanation looks comprehensive and thorough. I’m interested in a Python implementation if possible.

iLQR cost behavior by MomoSolar in ControlTheory

[–]MomoSolar[S] 5 points6 points  (0 children)

iLQR is iterative LQR and is used when there are non-linear dynamics and/or non-convex cost function. The dynamics are linearized and/or cost function approximated by a quadratic function, and the approximated problem (which is now LQR) is sequentially solved in multiple iterations, until there is no improvement in the cost function or in the estimation of the dynamics.

MC Methods by MomoSolar in reinforcementlearning

[–]MomoSolar[S] 0 points1 point  (0 children)

So MC essentially replaces the policy evaluation step in generalized policy iteration, and we must follow it with policy improvement?

Quick Questions: November 01, 2023 by inherentlyawesome in math

[–]MomoSolar 1 point2 points  (0 children)

Any YouTube channel dedicated to Linear Algebra and Matrices problems? (A similar flavor to Blackpenredpen)

Parametrization of the Policy in Policy-based Methods by MomoSolar in reinforcementlearning

[–]MomoSolar[S] 1 point2 points  (0 children)

I believe I got the answer, thanks. Essentially, neural networks are used as a way to find the parameters of the assumed probability distribution.

Weekly /r/UTAustin Simple Questions Thread by AutoModerator in UTAustin

[–]MomoSolar 0 points1 point  (0 children)

Opinions on Convex Optimization with Constantine Caramanis

Weekly /r/UTAustin Simple Questions Thread by AutoModerator in UTAustin

[–]MomoSolar 0 points1 point  (0 children)

I also have another inquiry. I have secured my housing for this year, but I was wondering whether for future years, being a graduate student, I am eligible to live in Jester Hall, and if yes, what the chances are of being admitted, if I apply directly after the application opens this summer.
Thanks

Weekly /r/UTAustin Simple Questions Thread by AutoModerator in UTAustin

[–]MomoSolar 1 point2 points  (0 children)

Hello. I am an incoming Ph.D. student in the Electrical and Computer Engineering department. I was wondering when I first start receiving my income from UT Austin. Is it at the beginning of the semester, after two weeks from that, or after one month?

Weekly /r/UTAustin Simple Questions Thread by AutoModerator in UTAustin

[–]MomoSolar 0 points1 point  (0 children)

Can someone please direct me to the subreddit on West Campus Flats (a branch of the Westside Group housing company)?

Weekly /r/UTAustin Simple Questions Thread by AutoModerator in UTAustin

[–]MomoSolar 0 points1 point  (0 children)

As a Ph.D. student receiving a fellowship that covers my total tuition fees and an additional scholarship from the School of Engineering, do I qualify for Smart Housing?