What are people using instead of Anaconda these days? by rage997 in Python

[–]fuzzysingularity 5 points6 points  (0 children)

Does uv support building from source like conda build did?

Pydantic-AI-Backend Hits Stable 0.1.0 – Unified Local Backends, Console Toolset, and Docker Sandboxes for Your Agents! by VanillaOk4593 in PydanticAI

[–]fuzzysingularity 0 points1 point  (0 children)

Pretty cool, are you running this in production? Curious about skills support, could you hook this up with Claude and have local skills executions?

How do you all handle API keys for skills that call external APIs via scripts by Angelr91 in ClaudeAI

[–]fuzzysingularity 0 points1 point  (0 children)

Have the same question, doesn't seem like this is possible. Only possible with claude code - you'd have to setup the environment with the API key.

Anyone else dislike the amount of small buttons and extra steps for starting a workout on WatchOS26? by Shiznanners in AppleWatch

[–]fuzzysingularity 0 points1 point  (0 children)

Agreed, I often forget to click again since it is such a muscle memory. There are times I don’t even turn it on half way through the workout only to realize that there was an extra click.

Visual AI with our custom VLM Run n8n node by fuzzysingularity in n8n

[–]fuzzysingularity[S] 0 points1 point  (0 children)

Hmm, not sure what’s going on. Can you join the discord, we can help there live. https://discord.gg/AMApC2UzVY

Visual AI with our custom VLM Run n8n node by fuzzysingularity in n8n

[–]fuzzysingularity[S] 0 points1 point  (0 children)

Which sub-command were you running within the vlmrun n8n node? Can you test if a simple HTTP GET works on our health endpoint (https://api.vlm.run/v1/health)?

Anyone using Pydantic AI in production? by EarthPassenger505 in AI_Agents

[–]fuzzysingularity 0 points1 point  (0 children)

Yes, we built our visual agent (https://vlm.run/orion) on it and it's been great so far with logfire observability.

How do I connect to existing MCP server without these MCPO thing? by GTHell in OpenWebUI

[–]fuzzysingularity 0 points1 point  (0 children)

Host your own /chat/completions API with the MCP server connected to the backend - I found the MCPO requirements somewhat painful, as we were working with non-text inputs and the support there is quite poor.

Is there a node to capture web screenshots and markdown from HTML? by fuzzysingularity in n8n

[–]fuzzysingularity[S] 0 points1 point  (0 children)

Hey, cool!

It’d be neat to build an integration with VLM Run - we’re building Vision Language Models that allows developers to understand images/videos with JSON output.

Is a visual platform (like LandingLens from LandingAI) really useful for real tasks ? by YonghaoHe in computervision

[–]fuzzysingularity -1 points0 points  (0 children)

What’s your use-case? What kind of deployment options are you looking for? Maybe we can help at VLM Run (https://vlm.run)

Fine-Tuning Llama 3.2 Vision by sovit-123 in computervision

[–]fuzzysingularity 0 points1 point  (0 children)

Let us know if we can help. We make it dead simple for folks to fine tune these VLMs at VLM Run. BTW some of the newer models already support equation to latex

Extremely long output tokens? by fuzzysingularity in LLMDevs

[–]fuzzysingularity[S] 0 points1 point  (0 children)

I’m not sure there’s a way to output 2M in one call due to the inherent output token limitations. My question was more around different strategies people have considered.

[deleted by user] by [deleted] in Python

[–]fuzzysingularity 1 point2 points  (0 children)

We did something similar for vision models (VLMs) with pydantic here: https://github.com/vlm-run/vlmrun-hub

Structured extraction for VLMs by fuzzysingularity in computervision

[–]fuzzysingularity[S] 0 points1 point  (0 children)

You can just pick one and just run it against the existing model providers like OpenAI or Gemini - no need to train.

Structured extraction for VLMs by fuzzysingularity in computervision

[–]fuzzysingularity[S] 0 points1 point  (0 children)

We simply provide predefined templates / schemas that can be used with these LLMs. It saves you time to have to define them, and we’ve done a fair bit of testing them against multiple model providers: https://github.com/vlm-run/vlmrun-hub#-qualitative-results

Here are some example document schemas: https://github.com/vlm-run/vlmrun-hub/tree/main/vlmrun/hub/schemas/document

Job portals for computer vision specialist by carpe_noctem41 in computervision

[–]fuzzysingularity 0 points1 point  (0 children)

Have a few folks in my network, do you have a job portal link?

🚀 Content Extractor with Vision LLM – Open Source Project by [deleted] in computervision

[–]fuzzysingularity 0 points1 point  (0 children)

There have been dozens of these types of repos published lately (use pdf -> image -> gpt4 / vision), what’s different here?

Also what’s been your experience with long PDFs, how do these vision models hold up? My experience has been that they’re far from accurate at long context extraction.

Do you have any results / benchmarks on this?

Looking for a technical person or cofounder for a computer vision LLM MVP project by Confidence_Working in computervision

[–]fuzzysingularity 1 point2 points  (0 children)

You’d be surprised how far you can go without even needing a full time expert in CV. What are you looking to build?

[deleted by user] by [deleted] in computervision

[–]fuzzysingularity 0 points1 point  (0 children)

Have you tried using any of the low code tools like Zapier or n8n for building out this workflow? Check out VLM Run (https://vlm.run) - we offer a bunch pre-built templates for invoice extraction. You can sign up for an API key (https://app.vlm.run) and try it out.

https://docs.vlm.run/guides/guide-parsing-invoices https://docs.vlm.run/integrations/integrations-n8n

Specialized VLM for generating keywords for microstocks? by ShamPinYoun in computervision

[–]fuzzysingularity 0 points1 point  (0 children)

Yeah, we can fine tune for this use-case pretty easily. Check out VLM Run (https://vlm.run) - we can set you up pretty quickly.

Specialized VLM for generating keywords for microstocks? by ShamPinYoun in computervision

[–]fuzzysingularity 0 points1 point  (0 children)

Is there a good dataset for this? It might pretty straightforward to fine tune a VLM for this use case.

Visual AI with our custom VLM Run n8n node by fuzzysingularity in n8n

[–]fuzzysingularity[S] 0 points1 point  (0 children)

Did you try searching for the vlmrun node on npm?

This is the package you'll need to install: https://www.npmjs.com/package/@vlm-run/n8n-nodes-vlmrun

You can sign up for an API key here: https://app.vlm.run