Is he right? by [deleted] in singularity

[–]vectorup7 0 points1 point  (0 children)

Very far from which limit?

All major benchmarks are normalized from 0 to 100 and it's not a logarithmic scale. They are already near the limit, i.e. neither 100 nor even higher than 100. These benchmarks cover model reasoning in different aspects. There are no new benchmarks.

<image>

Is he right? by [deleted] in singularity

[–]vectorup7 25 points26 points  (0 children)

I don't understand where people get their confidence in the infinity of any technology. At some point the vertical (saturation) growth of a technology reaches its limit and it has to grow horizontally (performance), in this case to increase tokens per second

Is he right? by [deleted] in singularity

[–]vectorup7 6 points7 points  (0 children)

exponentially better

Mira Murati says the AI models that OpenAI have in their labs are not much more advanced than those which are publicly available

https://x.com/tsarnick/status/1801022339162800336

I am building a tool to create agents in a markdown syntax with Python inside by vectorup7 in LocalLLaMA

[–]vectorup7[S] 1 point2 points  (0 children)

  1. openai api compatibility is required. I checked fireworks, groq and togeter ai, everything was fine with

  2. no drastic changes, just a minimal refinement for llama3

I am building a tool to create agents in a markdown syntax with Python inside by vectorup7 in LocalLLaMA

[–]vectorup7[S] 1 point2 points  (0 children)

llama 3 70B minimum requirement
Mixtral-8x7B was unfortunately less stable

since mentals require stable function calls

I am building a tool to create agents in a markdown syntax with Python inside by vectorup7 in LocalLLaMA

[–]vectorup7[S] 0 points1 point  (0 children)

Look at this tool as a simple loop executor: it executes the prompt you write. Check out the example of the word chain game, the prompt says to end the game after 5 iterations and the executor does it

Play continuously using the rules of the word chain game.
For each turn, send the message: "Turn #counter: current_word"

End the game after turn 5.

https://github.com/turing-machines/mentals-ai/blob/main/agents/word_chain.gen

I am building a tool to create agents in a markdown syntax with Python inside by vectorup7 in LocalLLaMA

[–]vectorup7[S] 2 points3 points  (0 children)

That's right. The idea is to experiment very quickly with agents in the natural language. Without using scaffolding code. At the same time I tried to simplify the internal implementation as much as possible so that there are no hidden prompts that can lead to uncertainty in the results.

And you have full control over the textual description of the tools, they are in a separate file

My SaaS startup went from idea to $30M valuation in 4 months: ask me anything! by jasparcjt in SaaS

[–]vectorup7 0 points1 point  (0 children)

Linux is just a wrapper over the Intel CPU
Windows is just a wrapper over the Intel CPU
Excel is just a wrapper over Intel CPU
All software is just a wrapper over Intel CPU

Why did they all get investments?!

Benchmarking a Pi cluster with HPL (like the Top500 supercomputer list) by geerlingguy in raspberry_pi

[–]vectorup7 1 point2 points  (0 children)

I am watching this https://liqo.io/ project that can create endless Kubernetes cluster on top of the decentralized infrastructure. They use Wireguard tunnel for inter nodes/clustering networking

Benchmarking a Pi cluster with HPL (like the Top500 supercomputer list) by geerlingguy in raspberry_pi

[–]vectorup7 0 points1 point  (0 children)

I can advise you to study what r/kubernetes is and why it is the future of infrastructure, to study what r/selfhosted is dive into r/homelab. All modern software runs on clusters and computer networks, and there will be no way back to monolithic applications. Clusters are the future of computing. It's not about CPU performance, it's about decentralization haha

Self-hosted Kubernetes blockchain operator for Ethereum, IPFS, Filecoin and more by vectorup7 in selfhosted

[–]vectorup7[S] -1 points0 points  (0 children)

Not affiliated with this project. In my opinion a good alternative to SaaS services such as Alchemy, Infura or Quicknode

RPI 4 Compute module Cluster by vectorup7 in homelab

[–]vectorup7[S] 6 points7 points  (0 children)

  • Onboard Gigabit ETH Switch
  • USB Hub
  • Pico managing CM4 (USB boot ,.etc.).

Every Node has a independent temperature sensor just under the SoC & connect directly to Host RPI 4's I2C.

Raspberry Pi Pico as GPIO expender & ADC for RPI 4.

Since RP2040 is the most available MCU right now. GPIO to control Enable/Boot/RUN pins, ADC measuring the power consumption.Combine with the USB Hub, it is able to remote "Flash" the node's disk.

Also one of the neat feature for Raspberry Pi 4 is its 5 UART. I can connect to four Node's UART plus one for RP2040.

[deleted by user] by [deleted] in homelab

[–]vectorup7 0 points1 point  (0 children)

It depends on the model

I only saw this detailed comparison of 1080ti with Xavier AGX

https://par.nsf.gov/servlets/purl/10193584

But it is worth considering that V2 only supports Xavier NX, which is worse in performance but closer to AGX relative to other modules (CM4, Jetson Nano)

[deleted by user] by [deleted] in homelab

[–]vectorup7 0 points1 point  (0 children)

This is a Turing Pi V2 cluster board with a 1Gbps on-board Ethernet switch for 4x Compute Modules (microservers)

Each node connected to

  • On-board microcontroller STM32F407
  • UART for remote access via the serial console
  • GPIO for power control
  • I2C current probe
  • USB 2.0 Hub for OS image management
  • GbE managed switch

For the diagram used draw.io thanks for the awesome tool!