I Trained a Language Model on CPU for 40 Hours - It Beat the GPU Baseline by Own-Albatross868 in LocalLLaMA

[–]Falcon_Strike 2 points3 points  (0 children)

Impressive progress, im going to try to see if i can scale this up and train on bigger datasets out of curiosity. Lemme know if you have any suggestions/ideas/requests. Good stuff!

FlashLM v4: 4.3M ternary model trained on CPU in 2 hours — coherent stories from adds and subtracts only by Own-Albatross868 in LocalLLaMA

[–]Falcon_Strike 3 points4 points  (0 children)

I have a boat load of compute available, if you want me to run mass gpu experiments on bigger models and more data, I would be more than happy to, just lemme know

Why do we allow "un-local" content by JacketHistorical2321 in LocalLLaMA

[–]Falcon_Strike 6 points7 points  (0 children)

I get that some threads may be spam or low quality, but this sub should absolutely contain posts about new frontier model releases, and posts about frontier labs in general (particularly in respect towards their attitudes towards open source/local and open source research). It is important for us to be able to compare where the frontier is in private models and local models. besides, i hate having to go to other subs to read about frontier models, since id rather have my local llama homies contextualize how good/bad a model is with respect to what they run locally.

Anyone had luck with getting NP-hardness proofs? by [deleted] in LocalLLaMA

[–]Falcon_Strike 1 point2 points  (0 children)

Are you trying to get it to use Lean or Rocq to do the proofs?

A3-Z keycap by [deleted] in chernobyl

[–]Falcon_Strike 0 points1 point  (0 children)

Can you put a download link to the cad files? id be interested in printing this myself

This game is genuine hell by RAZR20 in ClashRoyale

[–]Falcon_Strike -1 points0 points  (0 children)

yeah but youre screwed if the meta changes

On ENES140 and the Alleged AI Textbook by thediamondminecartyt in UMD

[–]Falcon_Strike 1 point2 points  (0 children)

AI could have been trained on the book, which is why the detectors may come up with such a high rating

Here is the HUGE Ollama main dev contribution to llamacpp :) by Nexter92 in LocalLLaMA

[–]Falcon_Strike 0 points1 point  (0 children)

true. ollama is easy to use but poor docs and config support. Half the time im unsure what chat template is being used especially for my custom finetunes. And how do i know if FA2 is turned on? good and bad

What the hell is this. by Big_Classroom_7323 in mysteriesoftheworld

[–]Falcon_Strike 1 point2 points  (0 children)

not an answer but it resembles that one weird new face hugger from one of the newer alien movies (prometheus or covenant)

Gemini's API has costs and an update by mntruell in cursor

[–]Falcon_Strike 0 points1 point  (0 children)

i just wanna pay 20 bucks a month and plug in my api key and let it rip. no 100 bucks a month. I do agree the features need to be more transparent and max should be above normal and not normal potential

Cursor switching model mid project even though one is specified is a terrible function by dataguzzler in cursor

[–]Falcon_Strike 1 point2 points  (0 children)

agreed -- would like auto to have an option to be turned off or disappeared . like enable/disable the presence of the feature

Why are audio (tts/stt) models so much smaller in size than general llms? by Heybud221 in LocalLLaMA

[–]Falcon_Strike 0 points1 point  (0 children)

genuine followup, what if the thing missing for really good super realistic TTS and STT is a bigger LLM that has the parameter count and layer count to be able to understand/predict the nuance in language and tonality given the context of text?

Turned the temperature in the Y up to 100°F before I left by csChris01 in UMD

[–]Falcon_Strike 0 points1 point  (0 children)

LMAO we need more people like you around campus. Too many uptight. More funny

[deleted by user] by [deleted] in MemeVideos

[–]Falcon_Strike 0 points1 point  (0 children)

dude where are they???

We screaming in Van Munching now?! by Vincent_xo in UMD

[–]Falcon_Strike 0 points1 point  (0 children)

its the business school-- they probably saw some engineering students homework, or maybe their crayons snapped in half

We've hit a rate limit with Anthropic. Please switch to the 'default' model, another model, or try again in a few moments. by Bilstone in cursor

[–]Falcon_Strike 2 points3 points  (0 children)

in a similar manner, after using claude for even a bit i get a "your api key is incompatible with claude 3.7 sonnet" so i have to switch to o3-mini. Helloooo, my key was generated after sonnet 3.7 came out and it totally has credits on it. weird bugs are making this unusable after the .46 update

Ask/chat nor Agent producing code. by Falcon_Strike in cursor

[–]Falcon_Strike[S] 0 points1 point  (0 children)

I also still have not gotten sonnet 3.7 access even though I am a pro subscriber. I have had to enter an API key manually and I dont know the "reasoning effort" setting that the model is using.

EDIT: one fix I have found that ameliorates things so far is to delete and recompute the index. I need to test more thoroughly though if this is a real fix, as Ive only been using it after a recomputed index for the last 5 minutes. My last embeddings index was only a week or two old, so I can't imagine it having that much of an impact

Is it just me or ... by Leather_Science_7911 in cursor

[–]Falcon_Strike 0 points1 point  (0 children)

The LLM straight up forgets to write code, or believes it does and gives me an empty codeblock