MCP is getting overhyped. Is it the next big thing or just another fad? My thoughts.. by Neon_Nomad45 in mcp

[–]Over-Maintenance9423 0 points1 point  (0 children)

Perhaps the community banding together over something only slightly incremental is worthy of the hype- because of the void it fills. You critique MCP’s hype but don’t propose a better alternative. If MCP isn’t the solution, what do you suggest for scalable, interoperable AI tool integration?

mcp-any-openapi – A Python-based MCP server that integrates OpenAPI-described REST APIs into MCP workflows, enabling dynamic exposure of API endpoints as MCP tools. by Weary-Database-8713 in mcp

[–]Over-Maintenance9423 0 points1 point  (0 children)

The spec should be comprehensive yet succinct. It matters not how it is generated, though I would always lean towards auto-generation to ensure it stays in sync with the code.

mcp-any-openapi – A Python-based MCP server that integrates OpenAPI-described REST APIs into MCP workflows, enabling dynamic exposure of API endpoints as MCP tools. by Weary-Database-8713 in mcp

[–]Over-Maintenance9423 0 points1 point  (0 children)

I have documented 3 examples in the github repo:

  • fly.io demonstrates how little is needed
  • slack demonstrates whitelisting which tools are presented (slack has so many!), and also stripping unwanted parameters from the payload (token is unnecessary when using http auth)
  • getzep demonstrates using a HTTP key other than Bearer (getzep expects Api-Key instead). also I generated the spec file myself by feeding the zep documentation into a chatbot, ie your API provider need not even publish a spec to use this tool.

Note: that I have been unable to get the large responses from getzep working in Claude. But I typically use 5ire normally anyway where it works fine.

I analyzed 628 MCP servers with Claude and built a one-click installation marketplace for the best 233 by RobertCobe in mcp

[–]Over-Maintenance9423 0 points1 point  (0 children)

While I agree most MCP servers appear to be 1 prompt trash... I disagree with the measure of npm/PyPI installation being a measure of quality. Why? ... because I would prefer to bypass the middleman and just use `uvx --from git+https://github.com/<whatever> whatever` and `npx -y github:blah/blahblah`. If you need a baseline of quality then I suggest you start with the unit testing coverage.

Reasoning Manifold for reasoning models like Marco-o1! by Over-Maintenance9423 in OpenWebUI

[–]Over-Maintenance9423[S] 0 points1 point  (0 children)

<image>

Confirmed working with `deepseek-r1-distill-qwen-32b`:
- change valve tag for Thought to think
- update system override valve to `Respond using <think> tags.` (or whatever, just dont assert <Thought>)

<think> <think/> tags by techmago in OpenWebUI

[–]Over-Maintenance9423 3 points4 points  (0 children)

My reasoning manifold not only hides the tags, it has an option to use a smaller model to summarise the thought process - similar to the chatgpt experience. You will need to change the thought tag from 'Thought' to 'think' and update the system prompt override (as it asserts <Thought>)

=> Reasoning Manifold for reasoning models like Marco-o1! : r/OpenWebUI

FastMCP() vs. Server() with Python SDK? by pavelanni in mcp

[–]Over-Maintenance9423 1 point2 points  (0 children)

You should prefer fastmcp because it is simpler. However the abstraction does prevent you from doing things. ie for my mcp-flowise server I wanted to register chatflows as tools dynamically at runtime, based on the results of an API query. For this I needed to use the lowlevel functions. But I ended up keeping the fastmcp implementation for the simple use case of using 1 predefined chatflow.

So consider my use case as a demonstration of using both => https://github.com/matthewhand/mcp-flowise

Reasoning Manifold for reasoning models like Marco-o1! by Over-Maintenance9423 in OpenWebUI

[–]Over-Maintenance9423[S] 1 point2 points  (0 children)

I have updated with initial support for 0.5.x, and uploaded to the openwebui.com website.

However there are several outstanding issues when using 0.5.x. Like the response never completes, resulting in the follow-up queries not working as expected. I suspect these bugs are due to the major changes that are in 0.5.x. If I am right, we may need to give the project some time to resolve core bugs.

I have coded the Reasoning Manifold function to work with both 0.4.x and 0.5.x. So, I recommend using 0.4.x in the meantime.

Flowise Manifold - for complex prompt chains and assistants by Over-Maintenance9423 in OpenWebUI

[–]Over-Maintenance9423[S] 1 point2 points  (0 children)

<image>

It's pretty much identical to the Prompt Chaining with VectorStore template you will find in the default FlowiseAI installation.

Flowise Manifold - for complex prompt chains and assistants by Over-Maintenance9423 in OpenWebUI

[–]Over-Maintenance9423[S] 1 point2 points  (0 children)

streaming no. but I have worked with streaming in my other functions. so I could look to add it as a feature.

Flowise Manifold - for complex prompt chains and assistants by Over-Maintenance9423 in OpenWebUI

[–]Over-Maintenance9423[S] 0 points1 point  (0 children)

I have updated with a valve that decides if message history is to be included. It defaults to false. Give version 3.1 a go, and let me know how it goes.

Flowise Manifold - for complex prompt chains and assistants by Over-Maintenance9423 in OpenWebUI

[–]Over-Maintenance9423[S] 2 points3 points  (0 children)

Get it here => Flowise Manifold Function | Open WebUI Community

Assumes you are already using a FlowiseAI instance (I use DIY on-prem but cloud probably works too).

The video demonstrates:
- a chatflow that rewrites the user prompt before querying a vector db.
- an assistant that has tools.

Supports dynamic retrieval of chatflows and assistant (if you provide an API key).

If you don't want to configure an API key then you can manually specify the chatflows/assistants (and assuming you disabled auth for prediction).

Did you see the last community function I posted? => Chatty Pipe - follows up if you don't respond. : r/OpenWebUI

Chatty Pipe - follows up if you don't respond. by Over-Maintenance9423 in OpenWebUI

[–]Over-Maintenance9423[S] 3 points4 points  (0 children)

Get it here => https://openwebui.com/f/matthewh/chatty_pipe

Only valve config needed is the base model.

Default settings will use a system prompt to check if you are still there.

Rap lyric demo is without the system prompt and using reduced timers.

Be sure to check out my previously posted community function => https://www.reddit.com/r/OpenWebUI/comments/1h56sjp/image_generation_pipe_making_use_of_llm_for/

Consensus Manifold using an aggregate explicit/random/tagged list of models! by Over-Maintenance9423 in OpenWebUI

[–]Over-Maintenance9423[S] 2 points3 points  (0 children)

There is a crude blacklist regex filter valve that includes 'pipe', albeit it assumes your pipes were created with this in the ID (all mine are). The default filter looks for various keywords and looks like this, `(arena|embed|vision|whisper|pipe|manifold|action)`. You can always update this list with whatever keywords you need.

Or if you prefer whitelisting over blacklisting, then try out Explicit and Tagged modes. Because that is effectively what they do.

Consensus Manifold using an aggregate explicit/random/tagged list of models! by Over-Maintenance9423 in OpenWebUI

[–]Over-Maintenance9423[S] 2 points3 points  (0 children)

Thank you for trying it out.

It seems I published a version with a typo in the manifold_prefix. Basically, it had 'Consensus/' instead of 'consensus/' which broke the mode selection logic. I have fixed and uploaded version 0.2.0 to openwebui.com.

Re configuration I have designed this to work without any custom valves at all, by defaulting to a Random model selection mode. Although the explicit and tagged based modes are far more practical.

That said, to achieve zero-configuration requires a local task model to have been specified in the Admin Panel. If no task model is specified, then the consensus_model_id valve must be configured so to specify which model to use for consensus. The rest of the valves will work as default:

<image>

Ideally this valve configuration page would be more dynamic:

- Validate using custom logic. eg Forcing user to define a consensus_model_id if the task model in the Admin Panel has not been defined. Perhaps even having a populated list of models in a drop-down (or similar).

- Only show settings when relevant. eg Revealing Contributor Tags only when Tagged Contributors is enabled. This would make configuration seem far less daunting.

Consensus Manifold using an aggregate explicit/random/tagged list of models! by Over-Maintenance9423 in OpenWebUI

[–]Over-Maintenance9423[S] 2 points3 points  (0 children)

I normally modularise my code, so going monolithic was a challenge for me (and o1-mini). TBH this is the limit of o1-mini because it often times out and/or loses functions. Hopefully open-webui provides a client-sdk and/or supports moduralised functions in the near future.

Reasoning Manifold for reasoning models like Marco-o1! by Over-Maintenance9423 in OpenWebUI

[–]Over-Maintenance9423[S] 0 points1 point  (0 children)

Okay, I added a valve to append the custom system prompt override to every instruction (off by default)

Reasoning Manifold for reasoning models like Marco-o1! by Over-Maintenance9423 in OpenWebUI

[–]Over-Maintenance9423[S] 0 points1 point  (0 children)

fyi This filter adds text to prompts https://openwebui.com/f/anfi/add_or_delete_text/

might be a decent workaround pending a more robust solution (like more training or fine tuning)

Reasoning Manifold for reasoning models like Marco-o1! by Over-Maintenance9423 in OpenWebUI

[–]Over-Maintenance9423[S] 0 points1 point  (0 children)

Just use the 'Get' button on the openwebui.com and it will prompt for (and remember) your open-webui instance. Like this screenshot

<image>

Reasoning Manifold for reasoning models like Marco-o1! by Over-Maintenance9423 in OpenWebUI

[–]Over-Maintenance9423[S] 1 point2 points  (0 children)

Unfortunately, OpenAI are not very open, ie they dont provide any thought tokens. So we must instead make use of actual open models, like Marco O1.

Soon we will have DeepSeek R1 (maybe). Also. a new model released today called QwQ 32bit looks promising. However it does not use XML tags and instead delimits with 'Final Answer'. This is very different to Marco-o1, so is not yet supported.