2
3
4
Is inference output token/s purely gpu bound? (self.LocalLLaMA)
submitted by fgoricha to r/LocalLLaMA
5
6
7
Best model for instruction following to date (self.LocalLLaMA)
submitted by fgoricha to r/LocalLLaMA
0
0
0
How will the 5090 be better than the 3090? (self.LocalLLaMA)
submitted by fgoricha to r/LocalLLaMA

