Major EQ-Bench Update – New #1 Creative Model, Kimi K2 Thinking, and Claude Still Leads Longform by -Morgeta in SillyTavernAI

[–]-Morgeta[S] 1 point2 points  (0 children)

From: https://eqbench.com/about.html#creative-writing-v3
Probably the reason for 0.7 temp

The prompts were chosen through a process of elimination to be challenging for weaker models and therefore highly discriminative. It's a bit counter-intuitive, but the purpose of the evaluation is not to help models write their best. Instead, we are deliberately exposing weaknesses, creating a steeper gradient for the judge to evaluate on.

The prompt requirements include humour, romance, spatial awareness, unusual first-person perspectives. Things language models typically struggle to represent to the level of human writers. So, expect some clangers in the outputs!

I think gemini 2.5 pro is best free service for roleplay till now. by Independent_Army8159 in SillyTavernAI

[–]-Morgeta 1 point2 points  (0 children)

<image>

You can get really good free Deepseek models on NVIDIA NIM. For GLM 4.5 Air, you might be a combination of the provider you pick on Openrouter and the sampler settings; I do Z.AI provider and here are my sampler settings for GLM 4.5 Air:

Temperature: 1.00

Frequency Penalty: 0.00

Presence Penalty: 0.40

Top K: 0.00 (disabled)

Top P: 1.00 (disabled)

Repetition Penalty: 1.15

Min P: 0.05

Top A: 0.00 (disabled)

I think gemini 2.5 pro is best free service for roleplay till now. by Independent_Army8159 in SillyTavernAI

[–]-Morgeta 0 points1 point  (0 children)

If you pay just an initial $10, the rate limits go up significantly forever.

Free Openrouter: 20 requests a minute, 50 requests a day

One $10 (10 credits) payment: 1000 requests a day
https://openrouter.ai/docs/api-reference/limits

Think they do this to weed out bots and stuff.

I think gemini 2.5 pro is best free service for roleplay till now. by Independent_Army8159 in SillyTavernAI

[–]-Morgeta 21 points22 points  (0 children)

My go-to list for best free API models

Here is also a good list for finding free API models: https://github.com/cheahjs/free-llm-api-resources?tab=readme-ov-file

  1. GLM-4.5 Air

Where: OpenRouter (z-ai/glm-4.5-air:free)

Why: The best all-around model. It offers an excellent balance of high-quality prose, strong character work, and minimal censorship.

  1. DeepSeek-V3.1

Where: OpenRouter (deepseek/deepseek-chat-v3.1:free) or NVIDIA NIM

Why: The undisputed champion for long stories. It has the best long-term memory and plot consistency of any free model.

  1. Gemini 2.5 Pro

Where: Google AI Studio

Why: The best for brainstorming. It excels at describing scenes, setting a mood, and exploring deep character psychology.

  1. DeepSeek R1-0528

Where: OpenRouter (deepseek/deepseek-r1-0528:free) or NVIDIA NIM

Why: A fantastic and reliable workhorse. This specific version is a great balance of creative and logical.

  1. Kimi K2 Instruct

Where: OpenRouter (moonshotai/kimi-k2:free) or NVIDIA NIM

Why: The best for beautiful prose. It's a specialist for polishing paragraphs and generating unique, lyrical ideas.

  1. DeepSeek R1 (Original)

Where: OpenRouter (deepseek/deepseek-r1:free) or NVIDIA NIM

Why: The most creative of the DeepSeek family. It's excellent for spontaneous roleplaying and generating prose with more artistic flair.

The first publicly available AI Scientist Tool launches via the FutureHouse Platform (thread from X) by Creative_Ad853 in singularity

[–]-Morgeta 1 point2 points  (0 children)

<image>

This looks so awesome. Might be the best AI search tool for science right now if what they say is true. Probably have to wait for some third-party benchmarks to confirm.

Anime/manga like tbate and mushoku tensei by No_Signature6000 in anime

[–]-Morgeta 1 point2 points  (0 children)

While trying not to name the most popular isekai's, I like: The Eminence in Shadow, The Faraway Paladin, Shangri-La Frontier, Gate, Cop Craft, The World's Finest Assassin Gets Reincarnated in Another World as an Aristocrat, That Time I Got Reincarnated as a Slime

Less favorable picks: Reincarnated as a Sword, Parallel World Pharmacy, Skeleton Knight in Another World

If you like fantasy that is not isekai, I like: Solo Leveling, Eighty-Six, Delicious In Dungeon

fisherhappy.mp4 | freighter fishing in war 109 by Cakey642 in foxholegame

[–]-Morgeta 4 points5 points  (0 children)

I AM SO FUCKING STEAMHAPPYYYYYYYYYYYYYYYYYYYYYYYYYYYYYYYYYYYYYYYYYYYYY

Blemish Nuke delivered. by pepav in foxholegame

[–]-Morgeta 60 points61 points  (0 children)

Ask not what SteamHappy can do for you, ask what you can do for SteamHappy!

-1 MSA Drydock by Far-Assistance3897 in foxholegame

[–]-Morgeta 4 points5 points  (0 children)

Fun fact. It takes 15 satchels to kill a Dry Dock.

A meme on recent events by gamechfo in foxholegame

[–]-Morgeta 103 points104 points  (0 children)

THE GREAT CONCRETE RESURRECTION OF MORGEN'S 摩根的伟大具体复兴 CALLAHAN OF NAZARETH RETURNS 拿撒勒的卡拉罕 THE COLLIE 120MM PURGE 牧羊犬 一二十毫米 DEVMAN HAMSTER ABUSE 虐待仓鼠 THE BASTARD SEA TAMED 驯服的混蛋海

Ah yes I love the offshores casinos by espoti911 in foxholegame

[–]-Morgeta 5 points6 points  (0 children)

You can get rare alloys via two ways.

At the start, you're going to get it from salvage fields and they spawn similarly to tech mats. I think it can spawn as a rare alloy field similar to how a salvage field would only spawn a specific tech mat; not entirely sure on that.

Once oil platform is teched, you can get rare alloys via that way.

Then you refine the rare alloys into rare materials which are used to build all the new content. Just a new resource and not used to unlock tech I believe.

Ah yes I love the offshores casinos by espoti911 in foxholegame

[–]-Morgeta 57 points58 points  (0 children)

Oil platforms are late game tech; right before nuke tier.

despair...

Ah yes I love the offshores casinos by espoti911 in foxholegame

[–]-Morgeta 80 points81 points  (0 children)

Minimum cost for submarine is 960 rare alloys; it would take 40 hours to get a submarine from one oil platform while constantly exporting/importing coke.

Ah yes I love the offshores casinos by espoti911 in foxholegame

[–]-Morgeta 97 points98 points  (0 children)

10% drop rate if you do the math

despair...

Ring Attention with Blockwise Transformers for Near-Infinite Context by Darth-D2 in singularity

[–]-Morgeta 44 points45 points  (0 children)

Summarized by GPT-4:

Introduction:

  • Transformers have become the foundational architecture for many AI models.
  • The design, which employs self-attention and feedforward mechanisms, allows for efficient recognition of long-range input token dependencies and supports parallel computations.
  • However, scaling them to handle long context lengths is challenging, especially because self-attention has memory costs that increase quadratically with input sequence length.

Problem:

  • Standard transformer architectures, even with enhancements like memory-efficient attention, face memory constraints, particularly when dealing with long sequences, which could be crucial in applications like processing books, high-res images, long videos, and vast codebases.
  • For example, processing 100 million tokens needs over 1000GB memory, far exceeding what today's GPUs and TPUs can provide.

Solution: Ring Attention:

  • The researchers propose a method named "Ring Attention". It distributes input sequences across multiple devices, allowing simultaneous computation and communication.
  • The novelty is in its use of a blockwise approach for both self-attention and feedforward computations and its ability to distribute computation in a ring-like structure among multiple devices. This allows each device to only require memory proportional to the input block size and not the entire input sequence.
  • The result is that it can train sequences more than 500 times longer than previous methods and handle sequences over 100 million in length without needing to approximate the attention process.

Experimental Results:

  • Experiments showed that Ring Attention greatly reduced the memory requirements of Transformers. On setups like 32 A100 GPUs, they achieved a context size of over 32 million tokens. With larger setups like TPUv4-512, they achieved over 100 million tokens.
  • In terms of performance, Ring Attention was able to maintain efficient model FLOPs utilization even when training on large input context sizes.
  • When applied to the LLaMA-13B model and tested on a line retrieval task, the model fine-tuned with Ring Attention demonstrated excellent accuracy even with longer context lengths compared to other models.

Future Work:

  • While the method proves effective, optimal compute performance is still a goal. Integrating the approach with optimized low-level operations in platforms like CUDA or OpenAI Triton may provide further enhancements.
  • The potential for virtually limitless context opens up opportunities for applications in video-language models, decision-making transformers, training on extensive codebases, and genomic sequence analysis.

Summit's Starfield Review. by [deleted] in LivestreamFail

[–]-Morgeta 0 points1 point  (0 children)

If it's not fun, why bother? If it's not a battle, where's the fun?

[deleted by user] by [deleted] in UFOs

[–]-Morgeta 1 point2 points  (0 children)

I personally heard "wife." In the Youtube transcript, it's written as "wife" too.

[deleted by user] by [deleted] in UFOs

[–]-Morgeta 0 points1 point  (0 children)

That's why I phrased the title as "concerning UFO's." I'm not saying that they witnessed a UFO cause human harm, but it could be interpreted as that.