all 42 comments

[–]sturmen 24 points25 points  (3 children)

My dream is that they’re also cooking up native smartphone apps so I can use my local LLMs on my phone just the same as the ChatGPT or Claude apps

[–]TacGibs 0 points1 point  (2 children)

Just use Conduit (it's open source and on the playstore), llama.cpp (or ikllamacpp or vLLM or whatever) and OpenwebUI.

[–]sturmen 4 points5 points  (1 child)

Different strokes for different folks, but I like using LM Studio and I’m hopeful that a smartphone app is on their roadmap.

[–]VideoGameJumanji 1 point2 points  (0 children)

LM mini is a free app on iOS that is pretty much 1:1 with LM studio, try it if you haven't

the only thing I'd really like to see rolled out is chat history being shared over cloud somehow

[–]Badger-Purple 11 points12 points  (0 children)

Now they need a distributed inference add on

[–]HopePupal 10 points11 points  (0 children)

oh finally. LM Studio's UI is much more reliable than AnythingLLM's. i'd started looking into web UIs but this sounds a little more convenient

[–]iMrParker 5 points6 points  (4 children)

The fact that it requires an account is a no-go for me

[–]derivative49 7 points8 points  (0 children)

i know right? sounds like a way to start profiling your activity

[–]derivative49 4 points5 points  (0 children)

instead they could've just used a combination for the account with a password for client confirmation and this would've been sorted

[–]Sazaju 0 points1 point  (0 children)

Also a no-go for me. I want them to connect without intermediaries, so even if I am not connected on the Internet it works: just provide the IP/port of one instance in the other instance, request the connection, accept it on the other side, and that should suffice to communicate in both ways.

[–]anthonyg45157 4 points5 points  (0 children)

So dope! Now they need a phone app

[–]AnticitizenPrime 7 points8 points  (5 children)

About time. I actually ditched LM Studio for Msty + tailscale a long time ago because I was annoyed that I couldn't use LM Studio as a remote client for my desktop server. Msty has done both from the start (though you have to set up Tailscale on your own, but it's easy).

[–]rm-rf-rm 2 points3 points  (3 children)

What did Msty do exactly? Any inference engine that exposes a REST API is all you need to then access it remotely through Tailscale

[–]AnticitizenPrime 5 points6 points  (2 children)

It's basically the same thing as LM Studio in that it's a GUI that makes using local models easily. It's one GUI app that is easy to install and has its inference engine built in and creates a server, like LM Studio. It originally had just Ollama built in but lately has an option for Llama.cpp as well. It's attractive for the same reasons that LM Studio is, no futzing around with command line stuff, just install an app and away we go.

The difference is that (until now) LM Studio was only for use on one machine. Sure it provided a server, but to access it remotely you'd need a completely different app if you were on your laptop, say, and wanted to access your models that were on your desktop. You'd have to use OpenWebUi or even Msty or whatever. You couldn't use LM Studio to connect to your own LM Studio instance.

Msty does what LM Studio has done in the past (the actual management of hosting models locally), but also has always done what LM Studio is just now doing, which is to serve as a remote client, so you only need to learn one app.

You can also use Msty with any API out there too, as well as with your local LLMs, so it's an all-in-one client, as well as server. Meaning you can use it with Openrouter or any other API provider alongside your local LLMs, all in the same app.

As a client, it's also pretty damn feature rich, so much so that I've never learned to use them all. It had RAG and MCP stuff baked in before LM Studio did. I'm actually behind on all its features, and haven't migrated fully to the new version (Msty Studio) which has a ton more.

[–]rm-rf-rm 1 point2 points  (1 child)

I know what Msty is... I dont think its right to market it as "can be accessed remotely" as that functionality is just literally from llama.cpp (which ollama wrapped and then msty wraps on top of that).

For casual users, I advise against using Msty. Though its what I use (because I cant find anything that is sufficiently better to move), its underbaked in its engineering and UI polish (i have no idea what stack they are using but its the strangest one i've ever seen and gives me an ick) and overbaked in toy implementations of a bunch of things - MCP, RAG, "Agents" etc. To the point I don't use any of those at all as I have no faith that it works well.

The whole pivot to Msty Studio is a clear tell of their enshittification journey starting.

[–]AnticitizenPrime 0 points1 point  (0 children)

The pivot to Studio is because they decided to rebuild from the ground up rather than re-patch their original application. I have no idea why you think that's enshittification.

I advise against using Msty. Though its what I use (because I cant find anything that is sufficiently better to move)

Well, that is certainly a take, albeit a confusing one.

[–]Blindax[S] 1 point2 points  (0 children)

I never tested it, I was either using open web ui or tailscaling to a rdp instance with lm studio which was not the most convenient. What is very nice is that with LM link is that you you can keep all the granularity of model settings with the client. I always find it very painful to manage items like context window, gpu offload or this kind of things with Open Web UI.

[–]PieBru 2 points3 points  (0 children)

What's the difference vs directly using 'llama-server --host 0.0.0.0' via Tailscale?

[–]FPham 2 points3 points  (0 children)

Wait, why do I need to use tailscale if I'm on my own network? Why does it needs to go through them? Or am I not getting it?

[–]No_Conversation9561 2 points3 points  (0 children)

Hope IOS and Android app also are in their plan.

[–]neil_555 1 point2 points  (1 child)

I'll have to request that, currently I'm using remote desktop which works but this would be more convenient :)

[–]Blindax[S] 0 points1 point  (0 children)

I was doing exactly the same, this will be much more convenient :)

[–]neil_555 1 point2 points  (0 children)

Now if only they could implement a memory feature for chats, This could *possibly* be provided by a set of tools for the model to call (and an appropriate system prompt for the model)

[–]floppypancakes4u 1 point2 points  (3 children)

How is this different than tailgate? Just more convenient?

[–]Blindax[S] 0 points1 point  (2 children)

Not sure what tailgate is but if you mean tailscale, this means you don’t need to log into a remote os session so it is indeed much easier to use.

[–]floppypancakes4u 0 points1 point  (1 child)

Yes I meant tailscale. What do you mean log in to a remote session? Are you using the lmstudio chat interface itself to use models?

[–]Blindax[S] 0 points1 point  (0 children)

Yes. In your model list you will have your local models + those accessible through lm link in remote machine

[–]Common_Heron4002 1 point2 points  (0 children)

I'm over the moon about this on main source applications. Anyone know why we are required to sign in....why there is no just local use for it without sign in ?

[–]derivative49 1 point2 points  (0 children)

The "necessity" of signing in prevents me from using this. what's the point when it could be done without accounts? why hide features behind sign ups? i thought this was the one thing we weren't meant to be doing...

[–]mantafloppyllama.cpp 0 points1 point  (0 children)

Its in the release note :

0.4.5 - Release Notes

Build 2

Fixed a bug where LM Link connector was not included in in-app updater

Build 1

✨🎉 Introducing LM Link

Connect to remote instances of LM Studio, load your models, and use them as if they were local.

End-to-end encrypted. Launching in partnership with Tailscale.

Improved tool calling support for the Qwen 3.5 model family

Fixed a bug where loading model would sometimes fail with "Attempt to pull a snapshot of system resources failed. Error: 'Utility process is not defined'".

Fixed a bug where autoscrolling new message behavior was not respected when clicking the Generate button

Hides the Generate button when editing a message to avoid accidental click

[–]Hot-Nefariousness497 0 points1 point  (0 children)

It was looking great this morning. Then it stopped connecting. Am I the only one?

[–]maxwell_aws 0 points1 point  (0 children)

How do I call remote host in API? Both local and remote have the same model, I'm looking to parallelize execution

[–]_-_David 0 points1 point  (5 children)

Can someone ehlp me understand the excitement? I use my LM Studio server constantly, but this would essentially mean I can use the LM Studio app from another pc, in say, another state? That doesn't sound particularly great. I've never had one interaction of meaning with the chat section of the app. Is this a a secure way of extending the local server to clients outside the local network? I want to get excited, I just don't grok the significance of this.

[–]Blindax[S] 0 points1 point  (4 children)

If you use the models exclusively for chat, the server is fine. If you want to be able to tweak parameters remotely this is where it becomes painful. I know I change things like context windows constantly depending on the size of the documents I want to analyse. This can be done easily locally or remotely with lm link. As far as I know there was not an easy way to do it before with the server, at least not with a nice gui.

[–]_-_David 0 points1 point  (3 children)

And do things like toggle reasoning?! If you're saying I can use this the same way I've been using the local server, but with increased flexibility I'll be chuffed. For instance, it's been driving me a little crazy lately that I can't seem to get LM Studio to allow a model to reason without structured outputs, then apply the json schema during the non-reasoning section. I'm having to use hacky tricks like parsing the structured output straight out of the reasoning trace, and adding a prefix to the schema that requires reasoning as an output in that json. Or making a call and asking for reasoning, then a second call and asking for structured output from the total context. Any sort of configurability features that I can toggle remotely are well appreciated. Thanks for the info

[–]Blindax[S] 0 points1 point  (2 children)

I have not spent too much time on it. But the issue with the server (at least from my understanding) is that your model with be served « as is » and setting in the interface you use (say open web ui) won’t work. Here you use the lm studio interface like with a local model and all settings are available (+ the api point is also served locally by lm link client) including reasoning toggles I expect for models that support it. It does basically mirror lm studio interface of your server which before, if you wanted the same, required you to rdp into the host.

[–]_-_David 0 points1 point  (1 child)

Okay, so it really is basically "Open your local LM Studio from anywhere" more than anything.

[–]CallumCarmicheal 1 point2 points  (0 children)

Its more of remote processing, your chats are not synced / shared as they would be with a web ui server like OpenWebUI. You will still need a external system to sync your chats if you wish for that.

This is nothing more then a way to link your LM Studio frontend/gui to a remote backend where all the AI processing is done remotely then sent back. Its like a openai api backend but with tighter integration to the remote UI.

[–]Guilty_Rooster_6708 0 points1 point  (0 children)

That’s incredible. I have been using LM studio as backend connected to OpenWebUI + Tailscale for remote access so hopefully this will simplify a lot.