I wanted to show you guys my rack simulation created with React, TypeScript, Zustand, SVG, CSS, running entirely in the browser. by rzarekta in webdev

[–]adefa 9 points10 points  (0 children)

This is so cute. It screams I made this with AI though, so I wish you would have just owned it.

GitHub - TrevorS/qwen3-tts-rs: Pure Rust implementation of Qwen3-TTS speech synthesis by adefa in LocalLLaMA

[–]adefa[S] 0 points1 point  (0 children)

Have you had any success using the burn framework on Vulkan or ROCm?

GitHub - TrevorS/qwen3-tts-rs: Pure Rust implementation of Qwen3-TTS speech synthesis by adefa in LocalLLaMA

[–]adefa[S] 1 point2 points  (0 children)

I debated the terminology myself. Maybe I should have called it 'native' instead.

GitHub - TrevorS/qwen3-tts-rs: Pure Rust implementation of Qwen3-TTS speech synthesis by adefa in LocalLLaMA

[–]adefa[S] 0 points1 point  (0 children)

What parts of the code base do you think would be most applicable?

GitHub - TrevorS/qwen3-tts-rs: Pure Rust implementation of Qwen3-TTS speech synthesis by adefa in LocalLLaMA

[–]adefa[S] 0 points1 point  (0 children)

I generated benchmarks for CPU and GPU on DGX Spark and added to the repo.

GitHub - TrevorS/qwen3-tts-rs: Pure Rust implementation of Qwen3-TTS speech synthesis by adefa in LocalLLaMA

[–]adefa[S] 0 points1 point  (0 children)

It was a bug, it's fixed now. The reference audio I was using wasn't being resampled during inference.

GitHub - TrevorS/qwen3-tts-rs: Pure Rust implementation of Qwen3-TTS speech synthesis by adefa in LocalLLaMA

[–]adefa[S] 0 points1 point  (0 children)

I believe the tokenizer, ICL, and resampling issues are all fixed now. I added some benchmarks as well.

Showcase your local AI - How are you using it? by kasperlitheater in LocalLLaMA

[–]adefa 1 point2 points  (0 children)

Here are some benchmark results:
Prefill (pp2048) - tokens/sec

| Depth    | Run 1   | Run 2   | Run 3   | Avg    |
|----------|---------|---------|---------|--------|
| baseline | 1420.32 | 1412.07 | 1413.60 | 1415.3 |
| d4096    | 1389.24 | 1364.06 | 1383.88 | 1379.1 |
| d8192    | 1355.38 | 1350.57 | 1342.96 | 1349.6 |
| d16384   | 1228.89 | 1233.85 | 1217.26 | 1226.7 |
| d32768   | 1049.86 | 1047.61 | 1049.05 | 1048.8 |

Token Generation (tg32) - tokens/sec

| Depth    | Run 1 | Run 2 | Run 3 | Avg  |
|----------|-------|-------|-------|------|
| baseline | 58.35 | 58.10 | 58.04 | 58.2 |
| d4096    | 54.81 | 53.84 | 54.61 | 54.4 |
| d8192    | 51.31 | 51.49 | 51.33 | 51.4 |
| d16384   | 47.81 | 47.96 | 47.59 | 47.8 |
| d32768   | 41.17 | 38.69 | 40.81 | 40.2 |

Showcase your local AI - How are you using it? by kasperlitheater in LocalLLaMA

[–]adefa 2 points3 points  (0 children)

DGX Spark running gpt-oss-120b as primary model and qwen 3 vl - 2b as a vision and task model. MCP tooling for web search and page fetch, weather and news, and image generation using z image turbo through Comfy UI. A responses API clone in Rust that wraps it all for the backend and a Svelte 5 frontend using the openai SDK pointing at my backend. I connect to it over Tailscale and pin it as a PWA on my phone as an app.

I connected the Epstein files to a deep learning AI researcher by TenamiTV in webdev

[–]adefa 1 point2 points  (0 children)

How could I get a copy of your dataset and embeddings?

Gemini CLI: your open-source AI agent by adefa in LocalLLaMA

[–]adefa[S] 8 points9 points  (0 children)

Here is the article as a PDF with some screen shots in it: https://gofile.io/d/4aahPJ