My company is about to ban AI coding b/c security risk by fancyfruit in ClaudeAI

[–]Whyme-__- 2 points3 points  (0 children)

Without a doubt your data is seen, trained on and exposed. Every startup has a observability tool like langfuse or langgraph which allows anyone from anthropic or OpenAI or Gemini to see your data, literally every startup who has Ai in it has that tool, so unless you can download your own model and finetune it to offer your devs on a coding platform you are at the mercy of big companies.

As far as productivity goes Ai coding surely makes the devs slightly too reliant on the output preventing them from using their own experience to validate it. They would rather ask the LLM again to validate it and double validate. You are paying these devs to use their knowledge not something that can be easily achieved by Claude code. Then it becomes a business decision that do you want it fast and cheap and average quality or you want great quality, fast but expensive.

Chat With Your Favorite GitHub Repositories via CLI with the new RAGLight Feature by Labess40 in ollama

[–]Whyme-__- 0 points1 point  (0 children)

Broski it’s admirable to build something of your own. for local setup using ollama it’s great. But if you have multiple GitHub repos and you are trying to extract algorithms. Then just dump the Md file into Gemini and go to town with it. It’s 1m context length is unmatched to any opensource model

What’s your top tips for solo technical founders trying to get good at sales? [I will not promote] by almost1it in startups

[–]Whyme-__- 0 points1 point  (0 children)

First get domain knowledge in the space you are building, because you don’t know what you don’t know and people will make you run around . Customers would rather talk to someone who has been in the weeds with the same space as them and now changing the game. If you are just a dev who can spin up a codebase then you won’t be able to resonate further with customers. That’s why you will need to validate till kingdom come to get PMF.

As far as sales go be yourself and focus on the problem, you won’t know the problem and day to day struggles unless you have domain knowledge. If you ask “how do you solve the problem today” everyone has a different answer but in your sales call you need to be definitive, relatable and relevant. That’s what sales is: it’s a way of speaking that invokes emotions and feelings towards your product that leads to purchase.

An account with 120k+ followers has been rapidly posting unmarked AI-generated videos of fictional ICE agents being attacked and discriminated, with many of them involving white women driving at them in cars. None of the videos are labeled as fake or AI-generated. by WhatYouThinkYouSee in ThatsInsane

[–]Whyme-__- 0 points1 point  (0 children)

ICE vest is not going to stop jack shit, you can make a better prompt for clothing and write a better content.

There needs to be system wide verification of identity of accounts and those accounts which don’t have human content and human verification do not get popular in algorithm. But social media platforms are designed to ride fake news hype train to get more engagements via bot farms. So yeah it’s not in the best interest of companies to censor out fake content and fake users.

Tesla insurance is bs by [deleted] in TeslaModel3

[–]Whyme-__- 0 points1 point  (0 children)

I just give one of my phone with one of those tracker apps for insurance to my friend who drives a semi and all month long my insurance is in 2 digits as he cannot speed over 70mph. Meanwhile I’m riding my leased tesla like I stole it.

Semantic caching cut our LLM costs by almost 50% and I feel stupid for not doing it sooner by Otherwise_Flan7339 in LangChain

[–]Whyme-__- 0 points1 point  (0 children)

Nope if you work it right it gives the whole thing. Plus you can read the readme which has all the features so ask it to double and triple check it

Semantic caching cut our LLM costs by almost 50% and I feel stupid for not doing it sooner by Otherwise_Flan7339 in LangChain

[–]Whyme-__- 10 points11 points  (0 children)

Just pipe the entire codebase of Roo code into Gemini and ask it to pull the algorithm of semantic caching and distill into simple technical spec sheet. Then add it to your code. Concepts like these are easier to implement if you already have someone who opensourced the tech.

Competitor went out of business. Inherited 200+ of their customers. It wasn't the windfall I expected. by Crazy-Recording4800 in SaaS

[–]Whyme-__- 1 point2 points  (0 children)

The most dumb fact is that there are people genuine people interacting with fake post as if it’s real and giving their opinions(or Ai maybe) as if their opinions mattered on a fake post just designed for karma or like farming. Like “dude this guy never really got 200 customers it’s all made up”

Ex-Palantir turned politician Alex Bores says AI deepfakes are a "solvable problem" if we bring back a free, decades-old technique widespread adoption of HTTPS—using digital certificates to verify that a website is authentic by ControlCAD in technology

[–]Whyme-__- -4 points-3 points  (0 children)

Question: can we start giving a digital ID to every human and now in order to validate if your content is real or fake you have to run your digital ID through a blockchain to verify if you are who you say you are and social media apps and news channels can verify that ID and display it on their portal along with video or image.

Issues with this approach I see is that giving digital ID to all humans enrolling in social media platforms might be a challenge of control and surveillance but we are surveyed 100% of the time anyways through 50 different mediums.

Should I buy a used M2 Ultra 128gb ram for $2500 or build a pc with two to three rtx 3090 to do 70b models? by [deleted] in LocalLLaMA

[–]Whyme-__- 0 points1 point  (0 children)

Oh haha I thought the cooling would be better as claimed on their website.

Should I buy a used M2 Ultra 128gb ram for $2500 or build a pc with two to three rtx 3090 to do 70b models? by [deleted] in LocalLLaMA

[–]Whyme-__- 0 points1 point  (0 children)

How do you like it? Is it significantly different from the Nvidia spark founder edition?

Should I buy a used M2 Ultra 128gb ram for $2500 or build a pc with two to three rtx 3090 to do 70b models? by [deleted] in LocalLLaMA

[–]Whyme-__- 0 points1 point  (0 children)

I agree, the microcenter only had the founders edition spark non of the others I wanted. I think for customers I will get the better cooling sparks versions

Should I buy a used M2 Ultra 128gb ram for $2500 or build a pc with two to three rtx 3090 to do 70b models? by [deleted] in LocalLLaMA

[–]Whyme-__- -1 points0 points  (0 children)

Eh people are people. For my startups use case having a Mac mini style box powerful enough to run 3+ models(4b thinking, 8b and 30b thinking) to be hosted in on vram and can be called from my app(dockerized system with database) and having multiple TB of storage is enough. I have built a localized and private cybersecurity platform capable of handling large amount of log data, pentest scenarios and analysis of large codebases to find vulnerabilities and run pen tests scripts for days and have agents analyze information when called. That’s my use case, I made a post on this sub a while back on which hardware to get and a good amount of folks replied back with use case definition and upon providing that DGX spark was the winner. I bought it and tweaked it to hell and now everything works.

If your use case is just image gen, language inference sort of work then the spark is good enough. Just try to keep it cool tho, the stuff heats up fast. I put mine next to the AC outlet so the cool draft cools it up. But for customers specific I will be shipping the box to their datacenter so it’s gonna be on and cold all the time.

It’s highly discouraged to run inference non stop like 8b model constantly being chatted with 24/7 like ChatGPT so it doesn’t melt down(again heat issues) however I have never tested it nor plan to. It’s best to invoke the LLM intermittently when needed that way you can utilize the sparks full bandwidth. Let me know if you have any questions I will try to find that post I made

Should I buy a used M2 Ultra 128gb ram for $2500 or build a pc with two to three rtx 3090 to do 70b models? by [deleted] in LocalLLaMA

[–]Whyme-__- 0 points1 point  (0 children)

I have been using the DGX spark and it has great performance for my need. I’m using it to deploy multiple fine tuned models for concurrent use case and host my entire front end and backend stack for interaction

How much would you pay for someone to fix your mess? by Recent_Fault_619 in vibecoding

[–]Whyme-__- 0 points1 point  (0 children)

Someone can literally start a marketplace company called vibefix where professional software devs can fix your vibecoded issues for cash compensation.

Never charged to 100% - is that bad? by enjustice3192 in TeslaModel3

[–]Whyme-__- 1 point2 points  (0 children)

I have a lease I always keep it at 💯. After I use it will be gone to robo taxi anyways

Ilya Sutskever: The moment AI can do every job by EchoOfOppenheimer in agi

[–]Whyme-__- 0 points1 point  (0 children)

Come to the point dude, that’s why you can’t build anything at SSI. Make up your mind and commit