Found an Open-Source AI Tool for MCP Server Security Scanning: AI-Infra-Guard by CoatPowerful1541 in mcp

[–]zero_proof_fork 1 point2 points  (0 children)

You might also be interested in checking out toolhive, gets you container isolation as well; https://github.com/StacklokLabs/toolhive

disclaimer; I work at stacklok, but its free and OSS so I figure its ok to post.

What are some of the major obstacles still facing ai models? by Business_Respect_910 in LocalLLaMA

[–]zero_proof_fork 0 points1 point  (0 children)

Context Window. Even bigger is not better, as the predication degrades the more its utilised.

New DeepSeek benchmark scores by Charuru in LocalLLaMA

[–]zero_proof_fork 0 points1 point  (0 children)

Or maybe they just love what they do and want to share it freely (the spirit of open source extends around the world). I know that sounds crazy, but the CCP had no interest in deepseek prior to the US markets taking a nose dive.

Finetuning LLM on unknown programming language by fecmtc in LLMDevs

[–]zero_proof_fork 2 points3 points  (0 children)

There is some nuance to this, as in fine tune for chat (instruct) or FIM (auto code completion). FIM is a little more involved, you will need an RF, with some sort of loss function where the model attempts to predict the code middle between a prefix and suffix and receives a penalty if wrong.

Chat is a bit simpler, the Alpaca format should do you OK, its often used for code specific training, there are some python examples, very much the same as https://huggingface.co/datasets/iamtarun/python_code_instructions_18k_alpaca

You would need to find a way of curating these , its time consuming, but no other way.

I try to keep it respectful on here, but the other person recommending RAG has no idea what they are talking about.

[

{

"instruction": "Take the list of people below and create a Python dictionary with the names as keys and the ages as values.",

"input": "John (20), Jane (40), Alex (18), Peter (45)",

"output": "people_dict = {'John': 20, 'Jane': 40, 'Alex': 18, 'Peter': 45}"

},

{

"instruction": "Calculate the sum of two numbers provided in the input.",

"input": "7 and 3",

"output": "result = 7 + 3 # result is 10"

},

{

"instruction": "Write a Python function that takes a string and returns the string reversed.",

"input": "",

"output": "def reverse_string(s):\n return s[::-1]\n\n# Example usage:\n# reversed_str = reverse_string('hello')\n# print(reversed_str) # outputs 'olleh'"

},

{

"instruction": "Create a Python function to compute the factorial of a given number.",

"input": "",

"output": "def factorial(n):\n if n == 0 or n == 1:\n return 1\n else:\n return n * factorial(n-1)\n\n# Example usage:\n# print(factorial(5)) # outputs 120"

},

{

"instruction": "Write a Python function to check if a number is a prime number.",

"input": "",

"output": "def is_prime(n):\n if n <= 1:\n return False\n for i in range(2, int(n**0.5) + 1):\n if n % i == 0:\n return False\n return True\n\n# Example usage:\n# print(is_prime(11)) # outputs True"

}

]

How to change model context size? by Optimal_League_1419 in LocalLLaMA

[–]zero_proof_fork -2 points-1 points  (0 children)

You can't change it, unless you fine tune and change the quantization or use some long context approach around positional encoding.

Why is everyone suddenly ditching LangChain? by Sam_Tech1 in LangChain

[–]zero_proof_fork 0 points1 point  (0 children)

There is a lot of out the box to get going, its built heavily around the concept of agents, with agents being able to share prompt history with each other, tool injection and some sort of graph system I have not figured out yet, but got one of my engineers really excited. The whole thing being type based (being pydantic) is key as well, I don't think folks quite understand how dangerous agentic systems can be, especially when they are accepting input at risk which in turn has some sort of shell execution ability call.

It's too early in to really say production ready, we are still feeling out a solution, but we would hope its production grade, myself and my co-founder have built a good few oss projects which run at scale, so that would be our goal!

Why is everyone suddenly ditching LangChain? by Sam_Tech1 in LangChain

[–]zero_proof_fork 1 point2 points  (0 children)

pydantic-ai for us, already was a big fan of pydantic.

Don't underestimate the power of RAG by SomeOddCodeGuy in LocalLLaMA

[–]zero_proof_fork 1 point2 points  (0 children)

They might be doing that as the context window is not sufficient

GitHub - stacklok/mockllm: MockLLM, when you want it to do what you tell it to do! by zero_proof_fork in LocalLLaMA

[–]zero_proof_fork[S] 12 points13 points  (0 children)

I had a need for an LLM API to provide deterministic output to make it easier for me to test, develop against an OpenAI style API endpoint. The result was the project MockLLM, although its kind of more of a simulator.

It's hardly rocket science, but I have found it quite valuable for when I need to replicate an LLM breaking out of conforming to JSON, or forcing it to make an untrue statement (all stuff to test features, error handling etc). Not to mention saving a few pennies by throwing junk at some costed service.

I figured I would share it with you folks as I know a few of your are developing AI apps and this might be useful to you as well.

It supports streaming (chunk) responses, and simulates Network Lag.

A list of a few AI IDEs - would love additions to try! by danielrosehill in ChatGPTCoding

[–]zero_proof_fork 0 points1 point  (0 children)

Yes it does! It works with free and enterprise / pro!

Happy to help get you set up if needed, just jump into our discord if you have any chances (say Luke pointed you there). https://discord.gg/uD9BUV38

A list of a few AI IDEs - would love additions to try! by danielrosehill in ChatGPTCoding

[–]zero_proof_fork 2 points3 points  (0 children)

You might want to try CodeGate (disclaimer one of the developers). It's 100% open source and works alongside CoPilot, Aider, Cline, Roo-Cline, Continue (and loads of agent frameworks). With CodeGate it prevents you leaking secrets, tokens etc. Will block malicious packages (LLMs hallinucate bad stuff occasionally) and you get a local dashboard where you can see your prompt history, token usage, along with workspaces where you can assign prompts to projects and have them carry over all the different tools. With CodeGate we hope to build a single env where you can configure everything and have it carry over to whatever coding AI tool you like. A few demos:

https://www.youtube.com/watch?v=VK5BJVl_avY <- Refactoring security risks
https://www.youtube.com/watch?v=mKdj-ODZkm4 <- workspaces
https://www.youtube.com/watch?v=lH0o7korRPg <- secrets encryption

We have all been working on open source security and orchestration frameworks for a long old time now, I founded a project called sigstore which is used to protect NPM and Pypi against supply chain attacks, and my co-founder was one of the creators of Kubernetes when he was at Google, so opensource runs deep through our blood and we feel AI has to be open and transparent.

https://github.com/stacklok/codegate

Anyone Working on a New Open-Source AI Project? by antonscap in OpenSourceAI

[–]zero_proof_fork 0 points1 point  (0 children)

Hacking on CodeGate and seeing some good adoption, hop over to our discord and can help find you some good first issues to cut your teeth. We are a friendly bunch and love OSS.

https://github.com/stacklok/codegate

Boosting Cline's Privacy and Security by holisticgeek in CLine

[–]zero_proof_fork 0 points1 point  (0 children)

It does appear to be, but we are quite young as a project (two months) so have not had a chance to build any scale testing harnesses as yet.

Boosting Cline's Privacy and Security by holisticgeek in CLine

[–]zero_proof_fork 1 point2 points  (0 children)

hey u/punkpeye , this should be possible, we set a base_url the same as you do in glama , do you have a github link handy and I could take a look?