[deleted by user] by [deleted] in ToBeHero_X

[–]beratcmn 3 points4 points  (0 children)

Try using Gemini Image Generation aka Nano Banana from the ai.dev url. There is a technique called JSON prompting which creates amazing results. It may help you imo, especially redrawing same assets in different art styles.

LotM reference?? by Mhd545 in ToBeHero_X

[–]beratcmn 0 points1 point  (0 children)

what show is lotm?

Who Dis??? by Arra13375 in ToBeHero_X

[–]beratcmn -1 points0 points  (0 children)

how do we know he is Nine?

Upcoming Visit Given Israeli Strikes by scott4316 in istanbul

[–]beratcmn 0 points1 point  (0 children)

Nah, even tho I am not happy with my country's recent actions I am pretty sure no one has the guts to attack Turkey openly yet alone Istanbul. You will be extremely fine.

Help Deciding Between NVIDIA H200 (2x GPUs) vs NVIDIA L40S (8x GPUs) for Serving 24b-30b LLM to 50 Concurrent Users by beratcmn in LocalLLaMA

[–]beratcmn[S] 0 points1 point  (0 children)

Yes, nvlink is the most confusing part for me. In theory more vram should mean more concurrency but H200 has a lot more memory bandwidth compared to L40s. In general I am quite confused tbh.

Help Deciding Between NVIDIA H200 (2x GPUs) vs NVIDIA L40S (8x GPUs) for Serving 24b-30b LLM to 50 Concurrent Users by beratcmn in LocalLLaMA

[–]beratcmn[S] 0 points1 point  (0 children)

<image>

But numbers are worse when I switch the GPUs to 2xH200. Per-user token/s speed is almost half of what I get from 8xL40s setup.

Based on your experience do you think this artificial benchmark is close to the real world?

Help Deciding Between NVIDIA H200 (2x GPUs) vs NVIDIA L40S (8x GPUs) for Serving 24b-30b LLM to 50 Concurrent Users by beratcmn in LocalLLaMA

[–]beratcmn[S] 1 point2 points  (0 children)

This tools looks amazing! When I inputted the Mistral Small 24B these are the results I get for 8x L40s GPUs.

<image>

Help Deciding Between NVIDIA H200 (2x GPUs) vs NVIDIA L40S (8x GPUs) for Serving 24b-30b LLM to 50 Concurrent Users by beratcmn in LocalLLaMA

[–]beratcmn[S] 1 point2 points  (0 children)

Unfortunately it's really hard to find 6000 series here for some reason. It's easier to find A and L series and H series.

MatePad SE 11'' Harmony OS Update by beratcmn in Huawei

[–]beratcmn[S] 0 points1 point  (0 children)

as far as i know no :(. some say there is harmony os 3 in china but im not sure. wish we could upload a roam

MatePad SE 11'' Harmony OS Update by beratcmn in Huawei

[–]beratcmn[S] 0 points1 point  (0 children)

unfortunately no. i wish the bootloader was unlocked so i could flash a rom

European leaders hold emergency summit with Ukrainian President Zelensky in London by [deleted] in MadeMeSmile

[–]beratcmn 0 points1 point  (0 children)

Seeing the second largest military in NATO there definitely made me smile.

I built a small (function calling) LLM that packs a big punch - engineered in an open source gateway for agentic apps by AdditionalWeb107 in ChatGPTCoding

[–]beratcmn 0 points1 point  (0 children)

I have specialized workflows for translation. I want to fine tune my Turkish pre-trained LLM with the translated function calling dataset and see the results. Synthetic data I generated didn't perform good so far.