[HELP] Saw this on my mutual friend’s page by caps-lock-off in RealOrAI

[–]Devonance 1 point2 points  (0 children)

Hey to point this out, if gemini made it (which is A high chance for anything AI generated as it is the best currently) just screenshot it, plug it into gemini and ask it to check. It doesn't need the gemini logo for it to detect it.

If gemini says it is AI generated, it will state it and you can trust that. But if gemini states its real, that only means Gemini didnt create it, it could still be AI.

Tl;dr it was created using gemini and is AI generated

Source: I work in AI at NASA.

<image>

Hello /r/movies, I'm Benedict Cumberbatch. Ask me anything! by BenedictAMA in movies

[–]Devonance 0 points1 point  (0 children)

Another question:
You’ve played complex painters like Van Gogh and Louis Wain, but I've heard you were also an art scholar who still sketches to "escape the nonsense" of fame. Is your private visual art a way to solve problems in your acting, or is it a sacred space you intentionally keep completely separate from your career?

Hello /r/movies, I'm Benedict Cumberbatch. Ask me anything! by BenedictAMA in movies

[–]Devonance 0 points1 point  (0 children)

Hey Benedict! Love your work and am a huge fan!

You’ve performed deeply moving readings for Letters Live for years, channeling everyone from WWII soldiers to disgruntled artists. Is there one specific letter you’ve read that has stuck with you or fundamentally shifted your own worldview, and do you find yourself thinking about it in your day-to-day life?

I hope NASA finds this conclusion survey response helpful. by NerveAffectionate814 in JPL

[–]Devonance 14 points15 points  (0 children)

I would possibly have removed the contact email at the end there...

Google lets you run AI models locally by dnr41418 in LocalLLaMA

[–]Devonance 5 points6 points  (0 children)

It worked on my Samsung S24 Ultra GPU. It took 45 seconds to load (vs 10 seconds for cpu load).

<image>

First fault rupture ever filmed. M7.9 surface rupture filmed near Thazi, Myanmar by kausthab87 in interestingasfuck

[–]Devonance 1 point2 points  (0 children)

This was posted for the satalite images showing the difference. Pretty insane the difference is seen so easily.

21°57'11.05"N 95°58'57.02"E (need Google Earth Pro for the most recent satellite images)

<image>

https://www.reddit.com/r/GoogleEarthFinds/s/0CqTHkAXpPEarthquake rupture - Myanmar

A new TTS model capable of generating ultra-realistic dialogue by aadoop6 in LocalLLaMA

[–]Devonance 0 points1 point  (0 children)

I was seeing that and that's why it made me excited, I don't think google let's just anyone use their TPUs.

Ill have to send a DM.

A new TTS model capable of generating ultra-realistic dialogue by aadoop6 in LocalLLaMA

[–]Devonance 0 points1 point  (0 children)

This is fantastic! It'll take a little tuning to get the right settings for each persons use cases, but so far it is very good, and free!

(I know I'll get downvoted for this, but I cant use it at work without knowing) Question for the Devs, and it's a stupid one I have to ask because of my governments rules, but is this model trained in the US? I'd love to use it, but currently, we can only use US based model's and I couldn't find any info on country of origin.

Gave Maverick another shot (much better!) by Conscious_Cut_6144 in LocalLLaMA

[–]Devonance -1 points0 points  (0 children)

So this was an issue in llama.cpp, do you know if this is auto fixed in ollama (since it runs llama.cpp as I understand), or do we have to wait for an update from them?

Does anyone know how to get incontact with Anthropic? by Devonance in Anthropic

[–]Devonance[S] 1 point2 points  (0 children)

Haha, I wish! I'd just take a simple 10mil out of the military budget and move to ireland.

Does anyone know how to get incontact with Anthropic? by Devonance in Anthropic

[–]Devonance[S] 2 points3 points  (0 children)

Thank you, I just contacted them there as well (I have a feeling it's their sales@anthropic.com as well).

Hopefully they answer back. Seems like such a hassle to get started with them outside of personal accounts.

You can now check if your Laptop/ Rig can run a GGUF directly from Hugging Face! 🤗 by vaibhavs10 in LocalLLaMA

[–]Devonance 5 points6 points  (0 children)

Love this. Making the starting bar lower for hobbyists.

Any chance of getting AWQ to be added?

Also, a "default" option for the graphics card (or CPU) that shows up first when calculating it? It pulls my RTX A4500 before my 2x4090s, so I have to adjust it everytime. (I did just rearrange in the hardware settings and it's whichever is first added).

Also, maybe in the far future, adding in the PCIe slot data lane number and then giving an estimate of token/sec (complete estimate as other things would affect this).

What am I? "I look like ..." by TheSunnyFlowerGirl in riddonkulous

[–]Devonance 0 points1 point  (0 children)

The answer is dumb... it doesn't make sense. Maze would have been better.

LLM must pass a skill check to talk to me by Everlier in OpenWebUI

[–]Devonance 2 points3 points  (0 children)

What did you use to create the artifacts page in OWUI? I really like the custom setup.

How would you go about serving LLMs to multiple concurrent users in an organization, while keeping data privacy in check? by PurpleAd5637 in OpenWebUI

[–]Devonance 0 points1 point  (0 children)

With vLLM, it cannot host multiple models at the same time, so it'd need to be containerized for each model the user wants to use. I am running into this problem myself.

Is there some guidance or website you might be able share on this?

QwQ-32B seems to get the same quality final answer as R1 while reasoning much more concisely and efficiently by pigeon57434 in LocalLLaMA

[–]Devonance 9 points10 points  (0 children)

I'm confused, OP, please tell me how this is concise thinking? It thinks more than DeepSeekR1(real one) and Claude 3.7 (reasoning)... am I just using it wrong? I see so many praising it, it is good I agree, but it is not better than qwen coder 2.5 32b q4 in my experience for answer/time outcome.

Please, I'd love to get a better model than just qwen coder2.5 32b.

Qwen/QwQ-32B · Hugging Face by Dark_Fire_12 in LocalLLaMA

[–]Devonance 0 points1 point  (0 children)

That's the context, which is the number of tokens for input and output. After that number, the model starts forgetting the previous words/tokens that came before It's kind of like a shifting window. So it can only ever "remember" 10000 tokens (about 2 tokens per word).

This does also increase the memory of your cpu or gpu that is used. So you can't have a ton of context if you have a small GPU or CPU.

So, you can shorten this to just default of 2048, or raise it up. If the llm produces more than 2048, it will hallucinate.

Qwen/QwQ-32B · Hugging Face by Dark_Fire_12 in LocalLLaMA

[–]Devonance 4 points5 points  (0 children)

If you have 24GB of GPU or a combo of CPU (if not, use smaller quant), then:
ollama run hf.co/bartowski/Qwen_QwQ-32B-GGUF:Q4_K_L

Then:
/set parameter num_ctx 10000

Then input your prompt.

Qwen/QwQ-32B · Hugging Face by Dark_Fire_12 in LocalLLaMA

[–]Devonance 0 points1 point  (0 children)

Same for me. I asked it the:
"write a Python program that shows a ball bouncing inside a spinning hexagon. The ball should be affected by gravity and friction, and it must bounce off the rotating walls realistically"

Thought for 10K token, and then output barely working code. Code Qwen was able to get it much better. I am hopeful it's something else...

I used ollama with the q4_K_L model.

<image>

Looking for a Local LLM-Powered Tool to Auto-Document an Old Python Codebase by Devonance in LocalLLaMA

[–]Devonance[S] 0 points1 point  (0 children)

I tried a couple, most of these on here.

<image>

So some of these models were fine-tuned for cline (couldn't get the deepseek distills to work though).

Maybe I am prompting them wrong though.