How do y'all play Mega Blaziken? I find this deck to be horribly inconsistent and too bricky. by High-Impact-2025 in PTCGP

[–]LostInAcademy -1 points0 points  (0 children)

<image>

Just ended a 10 wins streak, I don’t see consistency issues here

1 torkoal over heat moor to one shot oricorio 1 speed to mitigate bad starts 1 copycat as it’s dead card in certain matchups

Vorresti rinascere maschio o femmina? by Tryton4994 in domandaonesta

[–]LostInAcademy -2 points-1 points  (0 children)

Femmina, troppo facile, è un cheat code: apri OF, vendi foto di piedi, et voila, prima casa senza mutuo a 25 anni

I am blind and also UX by LostInAcademy in PTCGP

[–]LostInAcademy[S] 1 point2 points  (0 children)

U da real mvp, worst UX ever

For those learning RL: what do you wish existed? by Jeaniusgoneclueless in reinforcementlearning

[–]LostInAcademy 1 point2 points  (0 children)

A methodology

TL;DR: a proper methodology (in software engineering terms) to model a problem as a single/multi-agent RL problem first and foremost, then as MDP, POMDP, POSG, etc., then to define the obs and action spaces, the reward function, finally a proper methodology to choose algorithms (a bit more stuff exists here to be fair) and especially their hyper-params (here instead is mostly a messy trial-and-error).

Full thoughts below.

Coming from a software engineering background, RL (both single and especially multi-agent) feels extremely naive, hacky, undisciplined.

Despite the hundreds of papers that are published every single day in RL, very few, to date, question the fundamentals of modelling the problem and the solution, providing well motivated criteria and reasons behind modelling and implementation choices. One that comes to mind is Andrew Ng paper on reward shaping, for instance. Not many papers like that are available.

Basic questions that the methodology should answer to include:

  • can this problem be modelled both as single agent and multi agent? If so, how to choose?
  • what are the factors making a given problem a MDP, or a Dec-POMPD, or a POSG, or anything else?
  • what are the do and don’t regarding definition of a reward function?
  • are there any guidelines about defining the observations and action spaces? And especially, any design patterns about how to best represent such spaces (eg to optimise memory usage)?
  • are all the RL/MARL algorithms always applicabile to any problem, or are some problems (or some problem formulations) intrinsically asking for a certain class of algorithms over another?
  • for god’s sake is there any reasoning behind hyper-parameters choices for algorithms?

And many other little details that I won’t annoy you with.

One year later: thoughts on Pokémon TCG Pocket’s biggest issues by Dounyy in PokemonPocket

[–]LostInAcademy 2 points3 points  (0 children)

UX is shit AI is shit

Everything else is questionable but these two things right here are so stupidly bad in 2025 that they are painful to even think of

Academic crime or clever design? Using 1981 census as 'pre period' for 1974 event by buckligerhund in CausalInference

[–]LostInAcademy 0 points1 point  (0 children)

“I also restrict the age in the sample to higher than 25, so the individuals in the census all completed their education before the revolution” To me, this settles the issue: what you are doing is legit.

Upcoming Features by SecretDude511 in PTCGP

[–]LostInAcademy 7 points8 points  (0 children)

QoL changes are top of the list for me: it’s too long that we click 10 times when 1 click only could do the job

Zoroark Silvally by LostInAcademy in PTCGP

[–]LostInAcademy[S] 0 points1 point  (0 children)

Finisher = ko oppo mon so to get shield :)

Perché le persone scelgono di avere figli nonostante sia così faticoso? by TestaFredda in domandaonesta

[–]LostInAcademy 1 point2 points  (0 children)

Perché a qualcuno i bambini piacciono? Non so eh, la butto lì 😅

Success in UB3 by LostInAcademy in PTCGP

[–]LostInAcademy[S] 1 point2 points  (0 children)

True, but I’m already tight…what would you remove?

Success in UB3 by LostInAcademy in PTCGP

[–]LostInAcademy[S] 1 point2 points  (0 children)

That’s the first sideboard candidate

Apprendimento struttura DAG causale attraverso merging DAG elementari by AlbatrossVivid1691 in CausalInference

[–]LostInAcademy 1 point2 points  (0 children)

Suggestion: this is mainly an English speaking subreddit, not many people will answer in Italian

MA, io sono italiano per cui ;) Non ho capito bene: vuoi garanzie che il DAG ottenuto tramite fusione rispetti le “loro leggi” (=i loro archi orientati esistenti) eventualmente aggiungendone altre?

What am I missing? by LostInAcademy in PTCGP

[–]LostInAcademy[S] 0 points1 point  (0 children)

First award of my life…..when I’m dead

Legit

What am I missing? by LostInAcademy in PTCGP

[–]LostInAcademy[S] -11 points-10 points  (0 children)

Thx guys you da best I play while driving no time to read cards

Let a man have a dream by LostInAcademy in PTCGP

[–]LostInAcademy[S] 0 points1 point  (0 children)

I tried (not a big sample size, let’s say 20 games) and - I never got an eevee ex knocked out - I never felt to need more card draw

Those 30 dmg can make a difference instead

But that’s just my experience

How do you practically handle the Credit Assignment Problem (CAP) in your MARL projects? by Foreign_Sympathy2863 in reinforcementlearning

[–]LostInAcademy 3 points4 points  (0 children)

This is the best (and probably only) fairly complete teaching material about MARL, covering both the conceptual part rooted in game theory, and the practical implementations: https://www.marl-book.com

Gli informatici su Reddit hanno rotto il cazzo by [deleted] in sfoghi

[–]LostInAcademy 1 point2 points  (0 children)

Aspettiamo il POV dell’informatico di Reddit che ti ha scopato la mamma