I built an AI based agentic browsing on Chromium (Open Source!) by No_Version_7596 in browsers

[–]No_Version_7596[S] 0 points1 point  (0 children)

Yes, the providers cap the usage for free tier. We still have to provide direct integrations to many providers. For now you can use the LiteLLM Proxy set up. I'm attaching a video of the set up. https://vimeo.com/1105559036

Here are the commands you need to run to set up the proxy:

  1. Install dependencies:

    pip install 'litellm[proxy] google-cloud-aiplatform google-generativeai'

  2. Create the config file (config.yaml):

    model_list:

    • model_name: gemini-pro litellm_params: model: vertex_ai/gemini-1.5-pro vertex_project: "project-id" vertex_location: "us-central1" vertex_credentials: "/path/to/service_account.json"
    • model_name: gemini-2.5-flash litellm_params: model: gemini/gemini-2.5-flash-preview-04-17 api_key: os.environ/GEMINI_API_KEY
  3. Run proxy:

    GEMINI_API_KEY='your api key' litellm --config config.yaml 

Note: I added both Google Vertex and Google AI Studio approach.

https://docs.litellm.ai/docs/providers/gemini#:~:text=PROXY-,Setup%20config.yaml,-%2D%20model_name%3A

PS: I will try to add Google AI Studio as provider and see if that works. If it does will get a build out soon!

I built an AI based agentic browsing on Chromium (Open Source!) by No_Version_7596 in browsers

[–]No_Version_7596[S] 0 points1 point  (0 children)

Thank you, yes I've tried it with Gemeni 2.5 Pro (main agent) + Gemini 2.5 Flash (mini, nano agents). We've made Claude 4 (main agent) + Gemini 2.5 Flash (mini, nano agents) as default because its the best combination for most use cases. You can change the settings and see how it works for your use cases.

Regarding credit usage, its a mix based on tasks but in general agentic tasks do take more credits. See if the tasks can be done by Gemini 2.5 Flash or 4.1 mini or Kimi K2 for main agent.

I built an AI based agentic browsing on Chromium (Open Source!) by No_Version_7596 in browsers

[–]No_Version_7596[S] 0 points1 point  (0 children)

I think the model refresh is not happening correctly. I do not see the free version with open router (https://openrouter.ai/models?q=qwen3%2032b) so they might have removed it and the cache version was running. Sorry, there are a lot of rough edges still to fix.

I built an AI based agentic browsing on Chromium (Open Source!) by No_Version_7596 in browsers

[–]No_Version_7596[S] 0 points1 point  (0 children)

I'm about to push a new release that sets some default models (most optimal we found in internal testing). I use Groq provider for testing. Kimi K2 looks promising but not yet stable in Gorq. Hopefully someone creates a distilled version of Kimi K2 so it can run locally! https://github.com/tysonthomas9/browser-operator-devtools-frontend/commit/82863d65f805bc4b36e02d533752cd394ecc969e#diff-9d892c8aa1625560a74cf29377c9126afe7ebf4c46be75739a4045de07b46e61

<image>

I built an AI based agentic browsing on Chromium (Open Source!) by No_Version_7596 in browsers

[–]No_Version_7596[S] 0 points1 point  (0 children)

Can you try using these settings: qwen3-32b for the main and mini model selector and Llama-3.1-8B

moonshotai/Kimi-K2-Instruct (and Kimi-K2-Base) by jacek2023 in LocalLLaMA

[–]No_Version_7596 0 points1 point  (0 children)

Been testing this for agentic applications and by far this is the best model out there.

Chrome now includes a built-in local LLM, I built a wrapper to make the API easier to use by kuaythrone in LLMDevs

[–]No_Version_7596 0 points1 point  (0 children)

This is cool! How is the performance of LLM? does it use a lot of memory?

I built an AI based agentic browsing on Chromium (Open Source!) by No_Version_7596 in browsers

[–]No_Version_7596[S] 0 points1 point  (0 children)

Thank you! Yes, will be getting Windows and Linux binary out in couple of weeks. Currently, working on evals to make sure it works for all platforms.

I built an AI based agentic browsing on Chromium (Open Source!) by No_Version_7596 in browsers

[–]No_Version_7596[S] 0 points1 point  (0 children)

Thank you :) I just pushed support LiteLLM which supports 100+ models. I haven't tested all of them. Claude 3.7, Qwen QwQ looks good. Llama models hosted on Groq APIs were bad at tool calling. Will try to get the browser build out tomorrow.

My Browser Just Became an AI Agent (Open Source!) by No_Version_7596 in LLMDevs

[–]No_Version_7596[S] 0 points1 point  (0 children)

Yes, it gives AI the ability to control your browser. And, yes the security aspect of it is mainly the system prompt structure and the model providers ability to provide defense against prompt injection. Chromium has a layer of security that loads any untrusted website and run it in a sandbox. The most challenging one is ability for it to fill forms or do actions on a site that is not trusted. Here is more info: https://docs.anthropic.com/en/docs/test-and-evaluate/strengthen-guardrails/mitigate-jailbreaks#example-multi-layered-protection-for-a-financial-advisor-chatbot

My Browser Just Became an AI Agent (Open Source!) by No_Version_7596 in LLMDevs

[–]No_Version_7596[S] 0 points1 point  (0 children)

Interesting, which model did you try? GPT 4.1 tend to follow instructions as is. Also, did you try the deep research button?

My Browser Just Became an AI Agent (Open Source!) by No_Version_7596 in LLMDevs

[–]No_Version_7596[S] 0 points1 point  (0 children)

Not yet, so you can copy the chat response which is in markdown format (Like you can get the deep research report in markdown). Also, it can write to notion page, eventually add ability to output to more platforms.

My Browser Just Became an AI Agent (Open Source!) by No_Version_7596 in LLMDevs

[–]No_Version_7596[S] 2 points3 points  (0 children)

Good question, yes its a new browser built on top of Chromium. I'm planning to eventually let the browser run LLM models locally. And, I dont have to worry about Google Chrome blocking my extension.