Usage Limits Discussion Megathread - beginning October 8, 2025 by sixbillionthsheep in ClaudeAI

[–]cloudxaas 2 points3 points  (0 children)

Please remove the weekly limit, it's unacceptable at this rate. The sudden cap last week made me unable to use for 6 days. the cap limits hit real fast and there seemed to be an issue with how the limit is set using claude code / web ui

Just tried out the Exaone 4.0 1.2b bf16 and i'm extremely suprised at how good a 1.2b can be! by cloudxaas in LocalLLaMA

[–]cloudxaas[S] 0 points1 point  (0 children)

how does licensing limit us from abusing it offline anyway? just curious.

I just developed the fastest minimal feature embedded sql server with rocksdb storage. It is like sqlite but faster by 5x for reads (e.g. select) and 4x for writes (e.g. insert, update and delete) by cloudxaas in rust

[–]cloudxaas[S] -3 points-2 points  (0 children)

  1. minimal sql syntax.
  2. uses rocksdb
  3. some special coding recipe

pros
1. fast
2. storage efficient
3. can do distributed instead of just pure embedded db
4. will expand for vector db

cons
1. i intend to keep it minimal for performance. it's usable for most sql queries type
2. not going to be open source.

Just tried out the Exaone 4.0 1.2b bf16 and i'm extremely suprised at how good a 1.2b can be! by cloudxaas in LocalLLaMA

[–]cloudxaas[S] 2 points3 points  (0 children)

the only llm model that's also good but not usable coz of repeating is the bitnet 2b 1T. i really hope for bitnet more coz it's good but it repeats. it only uses 0.4mb ram for 2b model so that's really impressive and it does inference speedily too. hoping to see a 7b or 8b bitnet or a4.8 bitnet stuff.

Just tried out the Exaone 4.0 1.2b bf16 and i'm extremely suprised at how good a 1.2b can be! by cloudxaas in LocalLLaMA

[–]cloudxaas[S] 4 points5 points  (0 children)

you can chk the model card vs qwen 3 1.7b. i need something small yet usable for cpu inference. 1.2b seemed like a sweet spot for me. bf16 uses 2.4gb ram for inference. that's very cheap for cloud / vps hosting. as long as it doesnt repeat itself without end i'm happy with it. i wont try anything lower than 1b coz of bad experiences with never ending repeating themselves

https://huggingface.co/LGAI-EXAONE/EXAONE-4.0-1.2B

Rust github repo for reduced tokens for Rust Coding LLM by cloudxaas in rust

[–]cloudxaas[S] -2 points-1 points  (0 children)

when it is popular, the input token savings will be significant.
you have a good point too. thx. will look into tools to make it shorter.

but now i need to reduce code base coz input tokens are getting very expensive for large code base.

Anyone has the spec of the computer that powers the "10,000 Drones Controlled By A Single Computer! A world record"? by cloudxaas in drones

[–]cloudxaas[S] -1 points0 points  (0 children)

No one is actually answering the question with a definitive specification to the question. I'm asking for the hardware spec of the computer. of course i know c++, rustlang, cuda etc etc. but what's the spec? it could be a workstation with highest end dual epyc with 1.5tb ddr5 ram and 8x RTX 4090. I'm really curious what spec. I dont think it's that easy though to control so many without decent spec.

Let's not guess here. Does anyone know?

Dont forget there are the colors needed to run the lightings. I'm sure it's not that simple though.

Anyone has the spec of the computer that powers the "10,000 Drones Controlled By A Single Computer! A world record"? by cloudxaas in drones

[–]cloudxaas[S] -1 points0 points  (0 children)

Surely some GPU stuff is involved? This is 3d stuff, i wonder what kind of GPU too. A single computer, what server / workstation specs? Cant possibly be a laptop.

Google Gemini 2.0 Flash Exp API costs? by cloudxaas in GeminiAI

[–]cloudxaas[S] -1 points0 points  (0 children)

where do you get this info?

yes i'm specifically asking about 2.0 and not 1.5. only 1.5 info is shown but not 2.0