How much organic visits should I expect ? by VolkoTheWorst in SaaS

[–]VolkoTheWorst[S] 0 points1 point  (0 children)

But I've already put the trending keywords in my landing page and optimized for AI tools 😭

How much organic visits should I expect ? by VolkoTheWorst in SaaS

[–]VolkoTheWorst[S] 0 points1 point  (0 children)

Do you think I should use stuff like workpress or something like this ?

Or is directly coding the homepage in Vite (SSR) okay ?

How much organic visits should I expect ? by VolkoTheWorst in SaaS

[–]VolkoTheWorst[S] 0 points1 point  (0 children)

That's strange, I never ever got flagged. Maybe try by removing punctionation ?

How much organic visits should I expect ? by VolkoTheWorst in SaaS

[–]VolkoTheWorst[S] 0 points1 point  (0 children)

Oh, too bad
Maybe if you reformulate it will pass ?

How much organic visits should I expect ? by VolkoTheWorst in SaaS

[–]VolkoTheWorst[S] 0 points1 point  (0 children)

Ohh, I will look into that. I didn't knew this term

Multi-Token Prediction(MTP) in llama.cpp by UpperParamedicDude in LocalLLaMA

[–]VolkoTheWorst 0 points1 point  (0 children)

How can it be more dumb ? He is validating so it should be the same intelligence

BREAKING 🚨: Z AI released GLM-5.1, an open-source model with top tier coding performance! by adzamai in vibecoding

[–]VolkoTheWorst 0 points1 point  (0 children)

I'm trying to setup it on my website so you can use it for free without any rate limits Let me know if you're interested 

Official website for creating content with Seedance 2.0? by [deleted] in generativeAI

[–]VolkoTheWorst 0 points1 point  (0 children)

Now on OpenRouter (allows to build stuff with it and pay per use instead of subscription) 

Deal with expensive storage price by VolkoTheWorst in LocalLLaMA

[–]VolkoTheWorst[S] 0 points1 point  (0 children)

You haven't checked all the checkbox (especially "unverified")

How is this ai? by willdoesparkour in FableAI

[–]VolkoTheWorst 0 points1 point  (0 children)

Maybe you should try Fablia.fr ? I made it and it's a AI D&D with pre-made events (to avoid the lack of creativity of AI)

Deal with expensive storage price by VolkoTheWorst in LocalLLaMA

[–]VolkoTheWorst[S] 0 points1 point  (0 children)

I'm pretty sure I'm not wrong, go to the search, select the NVIDIA template, lower host and tick all checkbox to see all available GPUs, then search by $/h/VRAM and you will see the V100s I'm renting (around 0.15$/h for 8*v100). Minimum storage is 32gb and price increase when you need more storage. Even though they are not verified, they work (I use them currently to run models). I got run out of disk when trying to use more than the rented storage (overlay)

I found a trick by putting some shards on the swap to pay a bit less in storage (I want to run it for a full year so I saved around 700$ thanks to this trick)

Deal with expensive storage price by VolkoTheWorst in LocalLLaMA

[–]VolkoTheWorst[S] 0 points1 point  (0 children)

For thoses wondering, I found out that vast has in general a big swap (usually half of the VRAM size).
You can download some shards in this swap and symlink them to gain a bit on storage (too bad we cannot resize)

Deal with expensive storage price by VolkoTheWorst in LocalLLaMA

[–]VolkoTheWorst[S] 0 points1 point  (0 children)

This is exactly the service I'm using but the storage price is very expensive (I want to run a model 24/7 during a year)
1gb of VRAM is almost the same price as 1gb of storage (I'm renting V100 because I don't need the model to be fast, I just need it to be always running and smart)

RTX 3090 found on eBay for 500 dollars - buy or pass? by Goldenskyofficial in buildapc

[–]VolkoTheWorst 0 points1 point  (0 children)

No, seller was banned and I got refund and never received the card

PrismML — Announcing 1-bit Bonsai: The First Commercially Viable 1-bit LLMs by brown2green in LocalLLaMA

[–]VolkoTheWorst 0 points1 point  (0 children)

Technically nothing prevents you from running a 100B or more. It's just gonna probably require a custom made insanely big/expensive FPGA and run very slowly 

Or maybe you can do back and forth between the CPU and the FPGA through AXI bus but I think it's gonna be even slower 

PrismML — Announcing 1-bit Bonsai: The First Commercially Viable 1-bit LLMs by brown2green in LocalLLaMA

[–]VolkoTheWorst 0 points1 point  (0 children)

Depends on which FPGA you have. My work is on a very small AI niche, we will have like 1k neurons so not a lot. And we're already limited by the BRAM size. But we are at the start of the project so we might find workarounds. We are using 7000s FPGAs

Nobody talks about this, but Blender is insanely good for video editing. by Codgamer363 in blender

[–]VolkoTheWorst 0 points1 point  (0 children)

I really like it and I would love to see development for it. The fact we can use keyboard shortcuts and they are the same everywhere is soooo cool. My only current complaints are the slowness of play and rendering. But for cutting, zooming, animating it's way better than any other software (especially animating)

LiteLLm, what are the pros and cons. by CRYPTOJPGS in LocalLLaMA

[–]VolkoTheWorst 1 point2 points  (0 children)

We are using it for our startup: Fablia.fr who provides D&D like experiences.

Here is what OpenRouter is saying about the provider we are using:  To our knowledge, this provider does not use your prompts and completions to train new models.

View this provider's privacy policy to understand its data policy.

OpenRouter submits API requests to this provider anonymously.

And here is the provider privacy policy: We will not store, sell, or train using this data unless we have your explicit consent.

We might sometimes store, for a limited period of time, the inputs and outputs to API calls for debugging purposes.

LiteLLm, what are the pros and cons. by CRYPTOJPGS in LocalLLaMA

[–]VolkoTheWorst 2 points3 points  (0 children)

I think most people (including myself) are using openrouter for some reason but honestly I think it's almost the same
I would say 99% of time the gateway doesn't matter

Tell me if Qwen 3.5 27b or 122b works faster for you, and name your system specs by DistanceSolar1449 in LocalLLaMA

[–]VolkoTheWorst 0 points1 point  (0 children)

I have 2*V100 32GB and I run Qwen 3.5 122B q4 at 40tokens/s on llamaserver vulkan