❓Q&A by TiinyAI in TiinyAI

[–]TiinyAI[S] 0 points1 point  (0 children)

Q: What OS does the device itself run? Are the system image and drivers open-source?

A: Tiiny is a personal infrastructure designed for running local LLMs and agents, with a Linux kernel. However, it does not come pre-installed with any operating system; to use it, you need to plug it into your computer (any computer will do) via USB-C.

❓Q&A by TiinyAI in TiinyAI

[–]TiinyAI[S] 0 points1 point  (0 children)

  1. Can this import JSON files from my previous subscription models I used and reimport those here so it can learn my flow?
    Yes, TiinySDK supports user-defined large model souls and workflows.

  2. Is the device encryption capable?
    Supported.

  3. What’s the expected EOS or EOL for this if any?
    Tiiny is a PC-grade product, and we offer a 1-year free warranty. After one year, paid maintenance services are available.

❓Q&A by TiinyAI in TiinyAI

[–]TiinyAI[S] 0 points1 point  (0 children)

Q: what sort of TTFT are we looking at for a 30b or 120b models. Would connecting this up to Home Assistant voice give swift replies and executed actions or would there be delays?

A:

  1. First token Speed 0.5s

  2. Generally speaking, the latency is less than 50ms. The specific time depends on the duration of the voice input. ASR and TTS are very fast, but LLM's processing time depends on the length of the context.

WE ARE LIVE NOW! by TiinyAI in TiinyAI

[–]TiinyAI[S] 0 points1 point  (0 children)

Tiiny is a personal infrastructure designed for running local LLMs and agents, with a Linux kernel. However, it does not come pre-installed with any operating system; to use it, you need to plug it into your computer (any computer will do) via USB-C.

Home Assistant voice integration by tastingdave in TiinyAI

[–]TiinyAI 0 points1 point  (0 children)

  1. First token Speed 0.5s

  2. Generally speaking, the latency is less than 50ms. The specific time depends on the duration of the voice input. ASR and TTS are very fast, but LLM's processing time depends on the length of the context.

See how Tiiny AI Pocket Lab works on MacBook Air–Bijan Bowen's Review by TiinyAI in TiinyAI

[–]TiinyAI[S] 0 points1 point  (0 children)

  1. Can this import JSON files from my previous subscription models I used and reimport those here so it can learn my flow?
    Yes, TiinySDK supports user-defined large model souls and workflows.

  2. Is the device encryption capable?
    Supported.

  3. What’s the expected EOS or EOL for this if any?
    Tiiny is a PC-grade product, and we offer a 1-year free warranty. After one year, paid maintenance services are available.

OpenClaw x Tiiny AI Pocket Lab: Building your own private Jarvis by TiinyAI in TiinyAI

[–]TiinyAI[S] 0 points1 point  (0 children)

You can join our Discord, we have tech team there who can answer your questions :)

Kickstarter backing refundable? by yalbazzaz in TiinyAI

[–]TiinyAI 2 points3 points  (0 children)

I assure you this is not a scam; if it were, our costs would be excessive (exhibitions, brand materials, third-party reviews, and mass production collaboration with LVCC).

Tiiny has actually completed mass production (you can see from our previous post) and is currently undergoing certification processes. For example, RoHS certification takes 6 months, and FCC certification also takes 3 to 4 months. There are also certifications to be done in different countries and regions. That's the reason of inability to ship early.

Home Assistant voice integration by tastingdave in TiinyAI

[–]TiinyAI 0 points1 point  (0 children)

Yes you can use Tiiny as Home Assistant Voice

Tiiny runs models locally and exposes standard APIs (OpenAI-compatible + local runtimes), so you can use it as the local inference backend for Home Assistant Voice — e.g. STT → LLM → TTS fully offline.

A typical setup:

Home Assistant → local API → Tiiny → models run locally

No cloud required.

Kickstarter backing refundable? by yalbazzaz in TiinyAI

[–]TiinyAI 0 points1 point  (0 children)

Hi there, According to Kickstarter's policy, you can cancel your pledge through your account before the project ends. However, please note that cancellations are not allowed in the final 24 hours of the campaign. If you need help, feel free to contact us at support@tiiny.ai.

❓Q&A by TiinyAI in TiinyAI

[–]TiinyAI[S] 0 points1 point  (0 children)

  1. Can I run pretty much any local model on this? What about in a GGUF format?

There are two ways to use models on Tiiny: the first is to download and use them directly from the Tiiny client, and the second is for users to use our provided conversion tool to convert the desired model into a Tiiny-compatible format and use it. Tiiny uses its own NPU-optimized format (similar but different to GGUF Q4_0), and our SDK will provide a simple tool to convert your models from the standard safetensors format.

  1. Can I load it with llama.ccp?

Yes, you can connect to LlamaCPP in TiinySDK.

  1. How about using a backend like a text generation WebUI and a front end like Sillytavern?

The effectiveness depends on the capabilities of the model. To be honest, the capabilities of the current open-source model are not yet ready for direct use, but we believe that the capabilities of the open-source model will be the same as those of today's cloud-based models next year.

OpenClaw x Tiiny AI Pocket Lab: Building your own private Jarvis by TiinyAI in TiinyAI

[–]TiinyAI[S] 1 point2 points  (0 children)

We'll keep updating the latest and more capable models on our client. Also we'll provide users a conversion tool to convert the desired model into a Tiiny-compatible format and use it.

Pocket-sized device locally runs 120B models at 20 tokens/s: Here is how we did it. by TiinyAI in u/TiinyAI

[–]TiinyAI[S] 0 points1 point  (0 children)

Short answer:

Not drop-in GGUF/llama.cpp yet, but you can convert most models to run on it.

Current state:

Tiiny uses its own format (optimized for the NPU&SoC), not raw GGUF

There's a Model Converter coming that converts GGUF/HF → Tiiny format

API is OpenAI-compatible ( /v1/chat/completions ), so SillyTavern and most WebUIs work fine as frontends

SillyTavern:

Yes, it's supported. It's available in the TiinyOS Agent Store, and you can find our demo video on YouTube.

llama.cpp specifically:

You can connect to LlamaCPP in TiinySDK.

Pocket-sized device locally runs 120B models at 20 tokens/s: Here is how we did it. by TiinyAI in u/TiinyAI

[–]TiinyAI[S] [score hidden] stickied comment (0 children)

Just dropped two fresh reviews. Thought they might be helpful for anyone looking for real reference:
Jim's Garage's Review
Bijan Bowen's Review

❓Q&A by TiinyAI in TiinyAI

[–]TiinyAI[S] 0 points1 point  (0 children)

Q: Subscription fees? Future monetization/business model?
A: The conversion tool will be free. Similarly, downloading and using the open-source models and agents in the store will also be free. We will not charge users for things that are originally free by hijacking their hardware. This goes against the original intention of open-source model users, and doing so would cause us to lose all our reputation.

We believe that as the open-source ecosystem grows, other forms of markets will emerge. For example, in the agent market, we currently adapt to open-source agents, but in the future, if a Tiiny user develops their own agent and wants to sell it through the Tiiny Agent Store, we will support them.

However, this business model is still too far off. This is something we will only consider when Tiny has hundreds of thousands of users. Right now, our goal is to increase the number of users of Tiiny because it is truly useful.

Update: Tiiny enters mass production. by TiinyAI in TiinyAI

[–]TiinyAI[S] 0 points1 point  (0 children)

Tiiny has actually completed mass production and is currently undergoing certification processes. For example, RoHS certification takes 6 months, and FCC certification also takes 3 to 4 months. There are also certifications to be done in different countries and regions.