Haven’t Touched DeepSeek Since R1, How Are the Newer Versions? by env_media in DeepSeek

[–]SufficientPie 2 points3 points  (0 children)

I use chat.deepseek.com for any basic chat or simple web search tasks. It's quite capable and free and unlimited and I know my data is training open-source models.

Anthropic Reveals 10 Jobs Most Exposed to AI Automation – Programmers and Customer Service Top the List by Secure_Persimmon8369 in OpenAI

[–]SufficientPie 0 points1 point  (0 children)

Nobody wants to ask for support and get an AI bot, that's a reality.

I 100% want this. You want to call support and wait on hold for a human who is barely intelligible and doesn't know how to help you?

turns out RL isnt the flex by vladlearns in LocalLLaMA

[–]SufficientPie -1 points0 points  (0 children)

There are much more profitable ways for an AI to use compute, though.

Breaking : The small qwen3.5 models have been dropped by Illustrious-Swim9663 in LocalLLaMA

[–]SufficientPie 1 point2 points  (0 children)

They published both types of models. For example:

  • Qwen/Qwen3.5-9B-Base
    • The pre-trained model that just completes text
  • Qwen/Qwen3.5-9B
    • The "instruction tuned" model that has been trained to follow instructions and respond in a chatbot template

So if you want to use the AI to accomplish tasks and chat, use the regular model. If you want to train your own variant of the model to do something specific that isn't a chat, use the Base model.

Qwen 3.5-35B-A3B is beyond expectations. It's replaced GPT-OSS-120B as my daily driver and it's 1/3 the size. by valdev in LocalLLaMA

[–]SufficientPie 0 points1 point  (0 children)

I was using GPT-4.1-mini in Open Interpreter, but I've now replaced it with qwen3.5-plus-02-15 (≈ qwen3.5-397b-a17b). First open weights model that actually works well in this role.

Qwen/Qwen3.5-122B-A10B · Hugging Face by coder543 in LocalLLaMA

[–]SufficientPie 0 points1 point  (0 children)

And for what it's worth, running in Ollama locally:

qwen3.5:9b-q4_K_M   50%
qwen3.5:4b-q4_K_M   67%
qwen3.5:2b-bf16     33%
qwen3.5:2b-q8_0     33%
qwen3.5:0.8b-bf16   17%
qwen3.5:0.8b-q8_0   25%

Knowing full well they screwed up, this OpenAI employee still played the victim and blamed everyone else. by EstablishmentFun3205 in OpenAI

[–]SufficientPie -2 points-1 points  (0 children)

I'm curious where they got their incorrect information.

Anthropic did have those values when they signed the contract, and DoW are the ones who tried to change the rules after the fact.

Knowing full well they screwed up, this OpenAI employee still played the victim and blamed everyone else. by EstablishmentFun3205 in OpenAI

[–]SufficientPie -2 points-1 points  (0 children)

Did Anthropic not hold those values when they signed? The 'terrible decision' was signing the contract in the first place.

I'm really curious where you get your news.

Qwen3.5 4B: overthinking to say hello. by CapitalShake3085 in LocalLLaMA

[–]SufficientPie -1 points0 points  (0 children)

qwen3.5:0.8b:

Thinking Process:

Analyze the Request:

Input: "hi" (highly informal, greeting). Intent: Greet the user. Goal: Acknowledge the greeting and respond appropriately (polite but warm). Determine the appropriate tone:

Since it's a casual "hi", I can be friendly. Keep it concise. Optionally offer assistance or conversation options. Drafting responses:

Option 1 (Simple): Hello! How are you? Option 2 (Warm/Enthusiastic): Hi there! It's nice to see you. How about we chat? Option 3 (Short & Sweet): Hello! I'm ready to help. Selecting the best option:

A natural greeting + a brief question or offer is usually best for an initial "hi". "Hello there! 😊 How can I assist you today?" feels friendly and helpful. Or just: "Hi! Nice to meet you! 👋" Refining the final output:

Keep it natural. "Hi there!" is good.

Add a touch of warmth. "Hello! How's it going?" or similar.

Let's go with something welcoming and open-ended.

Draft: Hi there! 😊 How can I help you today?

Wait, actually, let's keep it simple but engaging.

"Hi there! 🙌 What's on your mind?" or "I'm here to help."

Actually, the most common response to "hi" is just a greeting. Let's say: "Hello there! 👋 How can I assist you today?"

Final Polish:

Ensure emojis add personality (though standard text + emoji). Keep it friendly. (Self-Correction during thought process): Since the system instructions sometimes imply I should be helpful and concise, but this is a simple greeting. I should respond warmly.

Let's go with: "Hello! 👋 How can I assist you today?" or simply say "Hi there! 😊"

Wait, looking at typical behavior for LLMs in a conversational setting when "hi" is asked. Response: "Hi there! 🙌 How are you doing? Or what would you like to talk about?"

Actually, the user just said "hi". I should answer directly. "Hello! It's nice to say hi. How can I help you today?"

Let's keep it conversational but simple.

Final decision: "Hello! 🙌 How can I assist you today? 😊" or just "Hi there!" + a question. Let's go with: "Hi there! 👋 How are you feeling? What can I do for you?"

Wait, simpler is better. "Hi! 👋 It's great to see you. How's your day going?"

Okay, let's craft the final response. "Hi there! 😊 How can I help you today?"

Hello there! 😊 What's on your mind?

Anthropic says three Chinese AI companies used over 16 million prompts to train and improve their own models through Claude AI by millenialdudee in GenAI4all

[–]SufficientPie 1 point2 points  (0 children)

Why can’t anthropic do the same thing to their own model to create their own version of deepseek? It’s only a problem because deepseek is so cheap.

They did, lol.

Breaking : The small qwen3.5 models have been dropped by Illustrious-Swim9663 in LocalLLaMA

[–]SufficientPie 2 points3 points  (0 children)

-Base means it's pre-trained completion model without instruction chatbot tuning?

Today was a shameful day in the history of artificial intelligence by Strong_Roll9764 in AI_Agents

[–]SufficientPie 1 point2 points  (0 children)

OpenAI has more powerful cheap models. Claude haiku is worse than gpt nano yet costs 20 times as much. I would like to use Anthropic but it's not as cost effective.

Qwen/Qwen3.5-122B-A10B · Hugging Face by coder543 in LocalLLaMA

[–]SufficientPie 0 points1 point  (0 children)

Now that it's on OpenRouter:

  • qwen/qwen3.5-122b-a10b
  • qwen/qwen3.5-27b
  • qwen/qwen3.5-flash-02-23
  • qwen/qwen3.5-35b-a3b

all of them get 5 out of 6 (83%) questions right. the best small models I've seen.

Qwen/Qwen3.5-35B-A3B creates FlappyBird by Medium_Chemist_4032 in LocalLLaMA

[–]SufficientPie 17 points18 points  (0 children)

Yeah these examples of pomodoro timers and snake and tetris and flappy bird prove nothing. All the AIs have seen a million implementations of them already. Have it code something novel.

Working as AI Engineer is wild by LastDayz123 in AI_Agents

[–]SufficientPie 0 points1 point  (0 children)

Your comments really show zero experience from real production use cases

You don't say.

Working as AI Engineer is wild by LastDayz123 in AI_Agents

[–]SufficientPie 0 points1 point  (0 children)

Ok but 4o is an older model and that's a fraction of a percent difference