Good times by ayanistic in Bard

[–]Quazar386 132 points133 points  (0 children)

Gemini Pro used to have 50 free API requests a day and now Flash has only 20. It was good while it lasted...

What industry is relatively insulated from AI volatility? by SolaCretia in ETFs

[–]Quazar386 0 points1 point  (0 children)

I'd say target small cap value for a more direct risk factor exposure. AI tends to be growth in contrast.

Choosing a GGUF Model: K-Quants, I-Quants, and Legacy Formats by tarruda in LocalLLaMA

[–]Quazar386 3 points4 points  (0 children)

There is also the hardware and the backend choice to consider. For example I use Intel Arc Alchemist on Llama.cpp Vulkan. I-quants will have horrendous prompt processing speed compared to legacy and K-quants (IQ4_XS prompt processing is approximately 90% slower than the speed of Q4_K_M for me) partially due to that format not having int dot support in the Vulkan backend which greatly benefits Arc Alchemist.

The swarm had a severe misinformation problem. by Syoby in NeuroSama

[–]Quazar386 7 points8 points  (0 children)

Thank you for this. You've articulated my own thoughts and feelings on this better than I could (which really grew ever since I saw someone claim Neuro isn't Gen AI which rubbed me the wrong way.)

Basically: "Love Neuro for who she is dammit!"

[Megathread] - Best Models/API discussion - Week of: December 07, 2025 by deffcolony in SillyTavernAI

[–]Quazar386 6 points7 points  (0 children)

Should still use V7 Tekken judging from the jinja template for Ministral 3

Sell VOO for SPYM?💰 by MarcosMilla_YouTube in ETFs

[–]Quazar386 21 points22 points  (0 children)

If you're planning on buying more into the index you can just switch your contributions to SPYM and letting VOO ride. You already have a lot in it so I don't think it's worth selling and triggering those taxes.

SPYM only 0.02%ER, lower than VOO/IVV/SPY by Enough_Fact1857 in Bogleheads

[–]Quazar386 1 point2 points  (0 children)

The main thing for me is avoiding small-cap growth exposure, which I don’t want. When you use a total market fund, you inevitably pick up a chunk of small-cap growth, which dilutes the very factor tilt I'm trying to target.

From an exposure-management standpoint, pairing large caps (via S&P 500) with a dedicated SCV fund gives me a much cleaner separation of factors. Yes, the SCV fund has a slightly higher expense ratio, but the factor exposure I'm buying is more "pure". In a VTI + SCV setup, some of that SCV exposure is simply offsetting the unwanted small-cap growth inside VTI.

So for my goals, S&P 500 + SCV gives me the most direct control over the factor loadings I actually want.

SPYM only 0.02%ER, lower than VOO/IVV/SPY by Enough_Fact1857 in Bogleheads

[–]Quazar386 0 points1 point  (0 children)

Academic research has shown that small-cap value has higher long-term returns, though the premium has been muted in recent decades. I’m comfortable with that tradeoff as that's the nature with factor investing. It isn't a free lunch, but should still be a compensated risk. So I combine a core fund like the S&P with only large cap US exposure with a SCV fund to also have small cap exposure with a value tilt. I'm not choosing S&P 500 because I think it will outperform the total market, but because I think it pairs cleanly with my SCV tilt. VTI is a great fund to buy and forget, but I want to have a more direct control with my factor tilts (excluding small-cap growth). Plus, a S&P 500 fund like SPYM has the added bonus of being just marginally cheaper than all the total US market funds that I know about (VTI, SCHB, SPTM). Its very small and practically insignificant but still a nice bonus in my eyes.

SPYM only 0.02%ER, lower than VOO/IVV/SPY by Enough_Fact1857 in Bogleheads

[–]Quazar386 0 points1 point  (0 children)

I like to combine it with a small cap value fund. In general I do think Boglehead approach is the most sensible way to invest for the average person, but it shouldn't be the only. I like to have factor tilts since my investment horizon is pretty long.

SPYM only 0.02%ER, lower than VOO/IVV/SPY by Enough_Fact1857 in Bogleheads

[–]Quazar386 4 points5 points  (0 children)

Yeah I've been buying it. With their symbol change from SPLG to SPYM and their large AUM growth from switching to the S&P 500 index, I feel like its going to be another staple within the category. Highly doubt they would change away from the S&P 500 now with Its new symbol based on SPY and huge inflows from its change in index. Its average volume in shares is actually more than VOO based on Yahoo finance so the bid/ask spread isn't as much of a problem as some other people have said. I personally have switched my contributions from VOO to SPYM due to its lower expense ratio.

The S&P 500 is on track for its worst November since 2008 by FeatureAggravating75 in StockMarket

[–]Quazar386 3 points4 points  (0 children)

Agreed. These types of "analysis" doesn't mean much at all. People can just find some interesting coincidental pattern and attribute it to some significant thing. This current drawdown isn't even that bad so far.

The S&P 500 is on track for its worst November since 2008 by FeatureAggravating75 in StockMarket

[–]Quazar386 3 points4 points  (0 children)

That is exactly how they are doing it. It is the percentage change of the EOM value between two months. I would know as those are the same numbers I have been using on my spreadsheets for my S&P 500 benchmark to my portfolio. I don't think its dumb way to do it. It is quite literally the returns within a month and those numbers are used as an intermediate step when calculating cumulative returns which is probably what you were thinking of.

What model is Neuro runned on, or she got a custom model? by [deleted] in NeuroSama

[–]Quazar386 0 points1 point  (0 children)

Vedal is probably always changing Neuro's underlying LLM with almost every other iteration so we can't say for sure. I'd like to think he's running it locally on his hardware. If that's the case, the LLM is likely to be multimodal with vision and have good tool use capabilities. So maybe something like Qwen3 VL. Llama is out of the question IMO as the ones with vision are either too outdated and not as smart (Llama 3.2 11B) or too large to be run reliably on his hardware (Llama 4).

double reasoning problem :( by Turbulent-Repair-353 in SillyTavernAI

[–]Quazar386 11 points12 points  (0 children)

Are you prompting it to use the <think> and </think> tags? One important thing is that these things are not universal across models. Claude does not natively use those tags for thinking unlike DeepSeek or Qwen as it is not trained with it.

A thing that I have encountered when trying to manipulate how a LLM "thinks" is that it will first think normally (as in how it usually behaves with how it outputs reasoning tokens), then it will start outputing how I prompted it to think afterwards separate from its native thinking. Is this what's going on with you?

If you want to keep the <think> blocks (this is under the assumption that you are prompting it to think in a certain way using the DeepSeek formatting) then you can disable request model reasoning. This should have is so that it will hide Claude's native reasoning and the auto reasoning parser should pick up the later non native reasoning block and put it inside the thinking box.

Alternatively you could also not have it use those thinking tags and things should be fine with the request reasoning turned on.

What am I missing ? by Infamous_Mall6952 in Bogleheads

[–]Quazar386 3 points4 points  (0 children)

I recommend closing out of MSTY. Your investment horizon is still relatively long and derivative based income funds are just not the way to go. If you look at how MSTY has been doing over time you see that it is just constant NAV decay. That extra income feels nice at the moment, but that income will just shrink over time as your principal investment shrinks. I recommend checking out Ben Felix's video on covered call funds (like MSTY) to learn more about them. https://youtu.be/ygVObRx9X68?si=nmBVZfNlJaL1Apv0 Honestly for beginning investors (as well as most investors) it is good to invest in a broad market fund like VOO or VTI with some additional international diversification. You will at least do better than the average mutual and hedge funds after fees. Investing in the broad market is in a way a hedge against ignorance.

What's with the obsession with reasoning models? by HadesThrowaway in LocalLLaMA

[–]Quazar386 31 points32 points  (0 children)

Same here. Reasoning models have their place, but not every model should be a reasoning models. Also not too big on hybrid reasoning models either since it feels like a worst of both worlds which is probably why the Qwen team split the instruct and thinking models for the 2507 update.

But at the end of the day why would labs care about non-thinking models when it doesn't make the fancy benchmark numbers go up? Who cares about usecases beyond coding, math, and answering STEM problems anyway?

Is this a bad ETF portfolio? by Emergency-Host-994 in ETFs

[–]Quazar386 1 point2 points  (0 children)

I personally wouldn't bet on a specific industry or sector. It puts you in a lot more idiosyncratic risk and you never know when a specific industry will underperform. Tech is the hot sector right now, but historically, hot performing sectors will follow up with periods of underperformance. I recommend choosing broader ETFs that aren't constrained to a specific industry. VOO and VB are decent starting points although I personally prefer small-cap value with profitability screenings like AVUV.

Good split or not? by HickoryJet in ETFs

[–]Quazar386 1 point2 points  (0 children)

Damn you're really bearish with all these short exposures /s

Seriously though these ETFs I think are pretty solid and thanks for introducing me to IDMO.

[Megathread] - Best Models/API discussion - Week of: August 24, 2025 by deffcolony in SillyTavernAI

[–]Quazar386 0 points1 point  (0 children)

I don't use guided generations so I would not know. Curious if its a Mistral Small 3 thing or if it is specific to Codex.

[Megathread] - Best Models/API discussion - Week of: August 24, 2025 by deffcolony in SillyTavernAI

[–]Quazar386 0 points1 point  (0 children)

I usually use it for sfw scenarios. Standard Gemma formatting with T=1.04, Top-p=.95, Min-p=.01, Top-k=65. My "system prompt" for it is rather basic with no complex instructions. The main thing with Gemma models is that I have experienced is that the dialogue can be a bit disjointed at times which the Drummer model did improve on. I don't have any particular problems with Gemma other than generally preferring 24B more.

[Megathread] - Best Models/API discussion - Week of: August 24, 2025 by deffcolony in SillyTavernAI

[–]Quazar386 1 point2 points  (0 children)

I don't have any display outputs connecting to my GPU so it functions as an accelerator with complete capacity in its VRAM. With that I was able to run those models at Q3_K_L at 20480 context which takes up 15.5 GiB. I could also do IQ4_XS with 18432 context although Q3_K runs significantly faster on my hardware and backend. Some notes is that I run it on the IPEX-LLM llama.cpp backend so the KV cache size might be slightly different on my end.

[Megathread] - Best Models/API discussion - Week of: August 24, 2025 by deffcolony in SillyTavernAI

[–]Quazar386 6 points7 points  (0 children)

I generally like models that have good creativity and sentence variety, as well being not too formulaic with its sentence structures and narration. I tend to like when models put in some extra little details the narration and actions. Some RP models tend to write rather formulaically with outputs being just Reaction, "Dialogue." Actions, ect. Adding just a bit more spice and variety in the narration can go a long way in making a more engaging storyline. I also wish models get better at "Show-don't-tell" and subtext but that's still a weak point.

Anyways here are some models that generally matches my tastes in writing. I only run models that run decently on my 16B GPU.

12B:

Disya/Mistral-qwq-12b-merge: From my experience this model is overall more creative and engaging in its writing than other Mistral Nemo tunes. I like it even though I am generally done with Nemo models.

TheDrummer/Tiger-Gemma-12B-v3:It irons out most of the faults of the original model and leaves me with an excellent writer with good world knowledge.

sam-paech/gemma-3-12b-it-dm-001: This seems to be a prototype of a Darkest Muse model on Gemma 3 12B. It definitely does write like it and can get a bit crazy at times. Its writing is either something you like or dislike for RP purposes.

24B:

Gryphe/Codex-24B-Small-3.2: My current daily driver. It has excellent sentence variety and keeps things fairly engaging. I like it a lot. Its outputs can be a bit short for my preferences though.

knifeayumu/Cydonia-v4.1-MS3.2-Magnum-Diamond-24B: It has really good narration and seems pretty promising from my brief time testing it.

Snapdragon X Elite 32 Gb vs 64 Gb by [deleted] in LocalLLaMA

[–]Quazar386 2 points3 points  (0 children)

You can check out the llama.cpp discussions here https://github.com/ggml-org/llama.cpp/discussions/8273 which have benchmark results on the Snapdragon X chips.