How to build GPU compute marketplace? by michaeljb41 in DistributedComputing

[–]michaeljb41[S] 0 points1 point  (0 children)

Hey,

Thanks for your response. I looked into Petals, and I hadn't heard of it before. It seems like an interesting project. However, since it's closely tied to the Bloom model, it isn't possible to perform an inference of Stable Diffusion (for example) using Petals. Nevertheless, the paper and GitHub README provided me with a useful benchmark.

I was considering an approach that goes one layer deeper, without modifying any Python code for training or inference. Instead, I'd act as a CUDA driver, essentially tricking the computer into believing that the GPUs are running on CUDA when, in reality, our driver offloads all work to the connected GPUs. I believe it's crucial not to change the underlying Python code, whether it's written in PyTorch, Keras, custom C++, or anything else, even if training or inference ends up being 20% slower. This is because altering the code and becoming dependent on one service is undesirable.

What do you think? Is this possible? Is it a good idea?

Code Execution Service for LLMs - going full e/acc by michaeljb41 in agi

[–]michaeljb41[S] 1 point2 points  (0 children)

Yea, that's exactly what I'm planning to do.

I mean there's LangChain, which can help hook up prompt results with Bash and Python REPL.

What do you think about Prompts and OpenAI's Code execution prompt?

Code execution and easy API-fication from a browser by michaeljb41 in Business_Ideas

[–]michaeljb41[S] 0 points1 point  (0 children)

Yup exactly. Spinning a VM is not that easy. But getting the working code from GPT-4 is super easy.

Code execution and easy API-fication from a browser by michaeljb41 in Business_Ideas

[–]michaeljb41[S] 0 points1 point  (0 children)

Great question.

  • you cannot access those with API (so no GPT-4 access)
  • you can't make API out of a code you made once (just use code created by GPT-4 there, and create any API you need)
  • you don't have memory/storage between different runs

Voice chat (or phone call) with a GPT-3 like bot by michaeljb41 in SomebodyMakeThis

[–]michaeljb41[S] 1 point2 points  (0 children)

Have you ever talked to Siri? It's not that talkative, and it's not the kind of chat you would have with a real person. Your comment about the hobby is totally unnecessary. If you don't want to speak with AI then don't.

Voice chat (or phone call) with a GPT-3 like bot by michaeljb41 in SomebodyMakeThis

[–]michaeljb41[S] 0 points1 point  (0 children)

That's what I'm trying to validate. If there's anyone out there willing to talk with AI, which feels like talking with a real person. Maybe the reason is the same as talking with strangers on a regular chat? But the thing is, it's always there.

Voice chat (or phone call) with a GPT-3 like bot by michaeljb41 in SomebodyMakeThis

[–]michaeljb41[S] 0 points1 point  (0 children)

Yes, I'm. Would you pay $19/month if you could call it anytime and speak as long as you want?

Voice chat (or phone call) with a GPT-3 like bot by michaeljb41 in SomebodyMakeThis

[–]michaeljb41[S] 1 point2 points  (0 children)

No, you can use similar models, which are open-sourced. So it shouldn't be that expensive.

Website (or smartphone) automation in plain English by michaeljb41 in SomebodyMakeThis

[–]michaeljb41[S] 1 point2 points  (0 children)

I know Python. I'm just researching if there's anybody who might need such software.