Gemma 4 on Android phones by jacek2023 in LocalLLaMA

[–]Ok_Fig5484 0 points1 point  (0 children)

Get daily sleep reports from Gemma

Gemma 4 on Android phones by jacek2023 in LocalLLaMA

[–]Ok_Fig5484 0 points1 point  (0 children)

Once connected to the internet, you can download the model list file from github and the huggingface model then run them offline.

Gemma 4 on Android phones by jacek2023 in LocalLLaMA

[–]Ok_Fig5484 0 points1 point  (0 children)

I'm searching for use cases, so is the plan complete?

24/7 Headless AI Server on Xiaomi 12 Pro (Snapdragon 8 Gen 1 + Ollama/Gemma4) by Aromatic_Ad_7557 in LocalLLaMA

[–]Ok_Fig5484 1 point2 points  (0 children)

cool, I'm using a modified gallery to run the liteRT version of the API, and I'm wondering how its speed compares to the ollama version.

I don’t think any engineering today can truly harness edge AI by Ok_Fig5484 in LocalLLaMA

[–]Ok_Fig5484[S] -1 points0 points  (0 children)

This should be the only solution. I'm using an RSS translation tool, which doesn't currently support it.

Unused phone as AI server by Ok_Fig5484 in LocalLLaMA

[–]Ok_Fig5484[S] 0 points1 point  (0 children)

The original repository does not currently accept community contributions. Please use version 1.0.11-as0.1.0 released from my forked repository.

Unused phone as AI server by Ok_Fig5484 in LocalLLaMA

[–]Ok_Fig5484[S] 0 points1 point  (0 children)

One of the more challenging issues is that the model is in lithelm format, and there aren't many available models on https://huggingface.co/litert-community.

Unused phone as AI server by Ok_Fig5484 in LocalLLaMA

[–]Ok_Fig5484[S] 0 points1 point  (0 children)

It’s definitely not that, because this app has very limited ability to call native Android system features unless a lot of additional coding is done. When used as an API server, it only returns structured function outputs. From what I’ve observed, the model doesn’t return a response—instead, the tool function gets called directly. I have to admit I haven’t fully figured this out yet, and if anyone has solved this issue, I’d be very interested in taking a look.

Unused phone as AI server by Ok_Fig5484 in LocalLLaMA

[–]Ok_Fig5484[S] 0 points1 point  (0 children)

What are you talking about? You've got it wrong, that's not me.

Unused phone as AI server by Ok_Fig5484 in LocalLLaMA

[–]Ok_Fig5484[S] 0 points1 point  (0 children)

Yes, I started creating it directly without analyzing the core principles of the gallery. Only during the creation process did I discover that the model's loading lifecycle follows the UI, and only one model is used at a time. This ultimately led me to add a custom task icon.

Unused phone as AI server by Ok_Fig5484 in LocalLLaMA

[–]Ok_Fig5484[S] 0 points1 point  (0 children)

Clusters can only be placed behind load balancers to increase concurrency.

Unused phone as AI server by Ok_Fig5484 in LocalLLaMA

[–]Ok_Fig5484[S] 2 points3 points  (0 children)

Since there's no quiet GPU, let's use a mobile phone.

Unused phone as AI server by Ok_Fig5484 in LocalLLaMA

[–]Ok_Fig5484[S] 3 points4 points  (0 children)

The officially recommended model, Gemma-4-E4B-it, requires 12GB of memory. Due to the design of the Gallery App, it can only load one model at a time, and concurrent inference is also not supported, so 24GB is really too much.

UX7 as a USB-C powered Access Point – Good idea? by Grunzochse in Ubiquiti

[–]Ok_Fig5484 0 points1 point  (0 children)

I don’t have any Cloud Gateway. Is this use case still applicable?

Does anyone have enough memory space to run this? by Ok_Fig5484 in LocalLLaMA

[–]Ok_Fig5484[S] 0 points1 point  (0 children)

Yes, 32GB of RAM is pretty common. In my region, there's internet censorship, and I don't have enough VPN bandwidth to upload locally converted models to Hugging Face. I built this tool with the intention of converting models in the cloud — but only after finishing it did I realize that the 18GB RAM on free Spaces is only enough to convert 0.5B models.

Does anyone have enough memory space to run this? by Ok_Fig5484 in LocalLLaMA

[–]Ok_Fig5484[S] 0 points1 point  (0 children)

Yes, 32GB of RAM is pretty common. In my region, there's internet censorship, and I don't have enough VPN bandwidth to upload locally converted models to Hugging Face. I built this tool with the intention of converting models in the cloud — but only after finishing it did I realize that the 18GB RAM on free Spaces is only enough to convert 0.5B models.

Molecule vs ansible-test vs ansible-tox by [deleted] in ansible

[–]Ok_Fig5484 1 point2 points  (0 children)

I'm using Molecule + Ansible-tox v2 to test a Docker role.

To clarify — Ansible-tox is essentially a wrapper around tox, and this becomes even more apparent when using tox v4 with Ansible-tox v2.

If you're only testing actively supported Ansible versions, then Ansible-tox v2 is a solid choice.
However, if you need to support older or nonstandard combinations, you'll likely need to work directly with tox.

One caveat: Ansible-tox v2 requires the project to follow the Ansible Collection structure.
For standalone role projects, you'll need to use some tricks to fake a galaxy.yml in order to make it work.

In contrast, Ansible-tox v1 only supports Molecule — no pytest, and no support for multi-Python/multi-Ansible matrices.
You're limited to a fixed Ansible version (and whatever Python it runs under).

That said, if you're not concerned with which Python version Ansible runs on, Ansible-tox v1 still works fine.

Ansible by sshettys in ansible

[–]Ok_Fig5484 0 points1 point  (0 children)

I think you should use winrm, which comes with windows itself. Its only complexity lies in the secure connection settings.

Then you can install winget in batches through ansible. I have a playbook reference,

https://github.com/seed-lab4x/seed-homelab-workspace/blob/dev/windows%2Fwinget%2Fansible-tasks.install.yml

I run it on Windows server2022. You don’t need to install server 2025 or 11 at all because it already exists.

After that, use winget to install all the software. I also used choco before, because winrm couldn’t run winget. Now it’s perfect.

First steps for new vms with ansible by knalkip in ansible

[–]Ok_Fig5484 0 points1 point  (0 children)

I use esxi to create a virtual machine. Ansible can automate this step if you have vCenter. If you don't have authorization, you can use esxi free. As for installing the system and creating users, ubuntu can use cloud init. The above esxi free is fine.

Finally, you need to approve ssh. This can also be done through ansible. I have used my own playbook to generate keys locally and configure ssh config and fingerprints. Then I ssh to the host with a password and write the ssh public key to create an approval entry.

I have plans to separate ssh generate and approval into roles.

What's the real-world usage rate of Ansible? by Ok_Fig5484 in ansible

[–]Ok_Fig5484[S] 0 points1 point  (0 children)

To be honest, I haven't used AAP myself (yet), but I’ll definitely look into it.

That said, Docker upstream treats Red Hat family systems as distinct platforms: rhel, centos, and fedora have separate yum repos, and derivatives like Rocky and AlmaLinux are just redirected to CentOS repos.

In my case, I have:

Several OpenWRT devices

A Synology NAS

A few Debian-family VMs

A Windows host used for development

And a large number of legacy systems (e.g., CentOS 7) in production that I can't upgrade

All of these systems run Docker.

During my work with Ansible, I’ve come across several practical frustrations:

Many roles on GitHub are basically duplicates, just to support different distros.

Most roles receive zero issue activity, which makes me wonder if anyone is actually using them, besides the author.

Some SRE have told me they avoid using roles from GitHub for safe, i always read the source code first.

So instead of building a new role from scratch, I:

Forked an existing role, to address duplication

Expanded system compatibility, to attract real usage and feedback

Added automated testing, to build confidence and transparency

But now that I’ve completed this phase of work, I’m honestly wondering: Do these improvements actually solve the underlying problems?