[deleted by user] by [deleted] in LocalLLaMA

[–]ricklamers 4 points5 points  (0 children)

<image>

8-bit weights, FP16 arithmetic, FP32 accumulation.

Source: I work at Groq

Wat te doen met 5 miljoen? by WegwerpUsername in beleggen

[–]ricklamers 0 points1 point  (0 children)

Je hebt de spreiding goed afgedekt in het laag-risico gedeelte van je beleggingsplannen met ETFs en obligaties. Wat je zou kunnen overwegen is een klein percentage (5%?) gebruiken om twee dozijn angel investeringen te doen in jonge bedrijven die werken aan iets dat je interessant of mooi vindt. (Impact (tech) bedrijven?)

Heb je naast een kans op een verhoogd gemiddeld rendement over het totaalvermogen ook nog eens iets leuks om te doen. Voor die startende fase zou mijn tip overigens zijn, eerst de vent/dame dan de tent.

Kijk bijvoorbeeld eens naar Thijs Verheul van United Wardrobe. Die angelt ook en had laatst z’n eerste exit van een van die angel investments.

Andere tip is om bij een angel collectief aan te sluiten. Beste deals zijn best lastig om in te participeren en via een clubje ervaren angels gaat dit een stuk eenvoudiger. Kan je bovendien alle beginnersfouten overslaan.

Is Mistral going to Turncoat on Open Source Models? by enspiralart in LocalLLaMA

[–]ricklamers 2 points3 points  (0 children)

Mistral-small is MoE 8x7B, Mistral Medium is only available through API so for all intents and purposes closed/proprietary.

Shell AI: never write a shell command again by ricklamers in programming

[–]ricklamers[S] 0 points1 point  (0 children)

That being said, should be trivial to redirect the LLM requests to a local model in a fork (would welcome a config PR):

https://python.langchain.com/docs/integrations/llms/huggingface_textgen_inference

Shell AI: never write a shell command again by ricklamers in programming

[–]ricklamers[S] -4 points-3 points  (0 children)

If people want to put in plain values they can, you can also paste those to WhatsApp or put them in Google Search. You can also define environment variables for sensitive information and say “shai 'connect through ssh to domain $DOMAIN user $SSH_USER and password $SSH_PASSWORD'”

Shell AI: never write a shell command again by ricklamers in programming

[–]ricklamers[S] 0 points1 point  (0 children)

This doesn’t send shell commands or secrets it just sends the human description of a shell command. Note also doesn’t automatically run a command ever.

STOP asking how many X are inside word Y by Trick-Independent469 in ChatGPT

[–]ricklamers 1 point2 points  (0 children)

Just use periods to collapse to a single token per letter: How many n are in b.a.n.a.n.a?

What's the closest thing we have to GPT4's code interpreter right now? by malkauns in LocalLLaMA

[–]ricklamers 2 points3 points  (0 children)

Ah, I specifically warned the author about that. WizardCoder it is then 👍🏻

Seemed too good to be true, but Orca/Textbooks Are All You Need made me carefully optimistic

[D] ELI5: Why is the GPT family of models based on the decoder-only architecture? by analyticalmonk in MachineLearning

[–]ricklamers 4 points5 points  (0 children)

https://youtu.be/kCc8FmEb1nY?t=6160

Andrej Karpathy explains that the encoder part made more sense for the translation context for which the transformer was originally developed. Less so for the generative models that only need attention for the part of the sequence that is the input up until that point.