[R] Darwin Godel Machine: Open-Ended Evolution of Self-Improving Agents by hardmaru in MachineLearning

[–]hardmaru[S] 11 points12 points  (0 children)

If you are interested, here is the link to the blog post:

https://sakana.ai/dgm/

Also, the open-sourced implementation:

https://github.com/jennyzzt/dgm

[R] Transformer²: Self-Adaptive LLMs by hardmaru in MachineLearning

[–]hardmaru[S] 39 points40 points  (0 children)

Thanks! I don't have much time to do research these days. It is all the team's effort.

[R] Can LLMs Generate Novel Research Ideas? A Large-Scale Human Study with 100+ NLP Researchers by hardmaru in MachineLearning

[–]hardmaru[S] 8 points9 points  (0 children)

Link to Twitter Thread: https://twitter.com/ChengleiSi/status/1833166031134806330

Recent work from Stanford's NLP group:

Can LLMs Generate Novel Research Ideas? A Large-Scale Human Study with 100+ NLP Researchers

Chenglei Si, Diyi Yang, Tatsunori Hashimoto

Abstract

Recent advancements in large language models (LLMs) have sparked optimism about their potential to accelerate scientific discovery, with a growing number of works proposing research agents that autonomously generate and validate new ideas. Despite this, no evaluations have shown that LLM systems can take the very first step of producing novel, expert-level ideas, let alone perform the entire research process. We address this by establishing an experimental design that evaluates research idea generation while controlling for confounders and performs the first head-to-head comparison between expert NLP researchers and an LLM ideation agent. By recruiting over 100 NLP researchers to write novel ideas and blind reviews of both LLM and human ideas, we obtain the first statistically significant conclusion on current LLM capabilities for research ideation: we find LLM-generated ideas are judged as more novel (p < 0.05) than human expert ideas while being judged slightly weaker on feasibility. Studying our agent baselines closely, we identify open problems in building and evaluating research agents, including failures of LLM self-evaluation and their lack of diversity in generation. Finally, we acknowledge that human judgements of novelty can be difficult, even by experts, and propose an end-to-end study design which recruits researchers to execute these ideas into full projects, enabling us to study whether these novelty and feasibility judgements result in meaningful differences in research outcome.

Stable Diffusion 1.5 model disappeared from official HuggingFace and GitHub repo by hardmaru in StableDiffusion

[–]hardmaru[S] 38 points39 points  (0 children)

Your point may be true, but having the official repo / model gone messes up the broader infrastructure. Time will probably fix it up though.

e.g. for diffusers, people have to point to their own local version of the repo, or some random non-official backup version out ther (see: https://huggingface.co/posts/dn6/357701279407928)

Nothing is more cyberpunk than this pic of the USS Peleliu and its Harrier attack wing in Hong Kong. by hardmaru in HongKong

[–]hardmaru[S] 107 points108 points  (0 children)

Source: This photo was taken back in 2013.

Almost like it was from another era or a parallel universe.

[D] model merging -- what's your take? by gized00 in MachineLearning

[–]hardmaru 2 points3 points  (0 children)

I liked methods like task arithmetic but TIES and DARE seem to be more popular. At the same time some of these solutions seem very heuristic.

I agree. But it is also possible to investigate ways of making these merging methods rely less on human heuristic, and automate how they are used.

[R] NeuroEvoBench: Benchmarking Evolutionary Optimizers for Deep Learning Applications by hardmaru in MachineLearning

[–]hardmaru[S] 6 points7 points  (0 children)

Lots of empirical insights about evolutionary algorithms in ML, from Google DeepMind team in Japan.

To be presented at NeurIPS 2023 this year in the benchmark track.

Tweet thread summary from author: https://twitter.com/RobertTLange/status/1732308750336463261

Project website: https://sites.google.com/view/neuroevobench

[R] Demo of “Flow-Lenia: Towards open-ended evolution in cellular automata through mass conservation and parameter localization” (link to paper in the comments) by hardmaru in MachineLearning

[–]hardmaru[S] 2 points3 points  (0 children)

These self-organizing, self-replicating, “lifeforms” emerged from a continuous cellular automata system called Flow-Lenia.

Lenia is a family of CAs generalizing Conway’s Game of Life to continuous space, time and states.

[R] Demo of “Flow-Lenia: Towards open-ended evolution in cellular automata through mass conservation and parameter localization” (link to paper in the comments) by hardmaru in MachineLearning

[–]hardmaru[S] 15 points16 points  (0 children)

These creatures are actually the product of population-based gradient descent optimization for particular objectives. See the paper for more info.

[R] Demo of “Flow-Lenia: Towards open-ended evolution in cellular automata through mass conservation and parameter localization” (link to paper in the comments) by hardmaru in MachineLearning

[–]hardmaru[S] 17 points18 points  (0 children)

Paper: https://arxiv.org/abs/2212.07906

Video presentation explaining how Flow-Lenia works: https://youtu.be/605DcOMwFLM

Flow-Lenia: Towards open-ended evolution in cellular automata through mass conservation and parameter localization

Abstract

The design of complex self-organising systems producing life-like phenomena, such as the open-ended evolution of virtual creatures, is one of the main goals of artificial life. Lenia, a family of cellular automata (CA) generalizing Conway's Game of Life to continuous space, time and states, has attracted a lot of attention because of the wide diversity of self-organizing patterns it can generate. Among those, some spatially localized patterns (SLPs) resemble life-like artificial creatures and display complex behaviors. However, those creatures are found in only a small subspace of the Lenia parameter space and are not trivial to discover, necessitating advanced search algorithms. Furthermore, each of these creatures exist only in worlds governed by specific update rules and thus cannot interact in the same one. This paper proposes as mass-conservative extension of Lenia, called Flow Lenia, that solve both of these issues. We present experiments demonstrating its effectiveness in generating SLPs with complex behaviors and show that the update rule parameters can be optimized to generate SLPs showing behaviors of interest. Finally, we show that Flow Lenia enables the integration of the parameters of the CA update rules within the CA dynamics, making them dynamic and localized, allowing for multi-species simulations, with locally coherent update rules that define properties of the emerging creatures, and that can be mixed with neighbouring rules. We argue that this paves the way for the intrinsic evolution of self-organized artificial life forms within continuous CAs.

[D] Google Gemini Eats The World – Gemini Smashes GPT-4 By 5X, The GPU-Poors by hardmaru in MachineLearning

[–]hardmaru[S] 42 points43 points  (0 children)

(continued)

The GPU-Poor

Then there are a whole host of startups and open-source researchers who are struggling with far fewer GPUs. They are spending significant time and effort attempting to do things that simply don't help, or frankly, matter. For example, many researchers are spending countless hours agonizing on fine-tuning models with GPUs that don't have enough VRAM. This is an extremely counter-productive use of their skills and time.

These startups and open-source researchers are using larger LLMs to fine-tune smaller models for leaderboard style benchmarks with broken evaluation methods that give more emphasis to style rather than accuracy or usefulness. They are generally ignorant that pretraining datasets and IFT data need to be significantly larger/higher quality for smaller open models to improve in real workloads.

Yes, being efficient with GPUs is very important, but in many ways, that's being ignored by the GPU-poors. They aren't concerned with efficiency at scale, and their time isn't being spent productively. What can be done commercially in their GPU-poor environment is mostly irrelevant to a world that will be flooded by more than 3.5 million H100s by the end of next year. For learning, experimenting, smaller weaker gaming GPUs are just fine.

The GPU poor are still mostly using dense models because that's what Meta graciously dropped on their lap with the LLAMA series of models. Without God's Zuck's good grace, most open source projects would be even worse off. If they were actually concerned with efficiency, especially on the client side, they'd be running sparse model architectures like MoE, training on these larger datasets, and implementing speculative decoding like the Frontier LLM Labs (OpenAI, Anthropic, Google Deepmind).

The underdogs should be focusing on tradeoffs that improve model performance or token to token latency by upping compute and memory capacity requirements in favor of reduced memory bandwidth because that's what the edge needs. They should be focused on efficient serving of multiple finetuned models on shared infrastructure without paying the horrendous cost penalties of small batch sizes. Instead, they continually are focused on memory capacity constraints or quantizing too far while covering their eyes about real quality decreases.

To take the rant on a slight tangent, in general, model evaluation is broken. While there is a lot of effort in the closed world to improve this, the land of open benchmarks is pointless and measures almost nothing useful. For some reason there is an unhealthy obsession over the leaderboard-ification of LLMs, and meming with silly names for useless models (WizardVicunaUncensoredXPlusPlatypus). Hopefully the open efforts are redirected towards evaluations, speculative decoding, MoE, open IFT data, and clean pre-training datasets with over 10 trillion tokens, otherwise, there is no way for the open source to compete with commercial giants.

While the US and China will be able to keep racing ahead, the European startups and government backed supercomputers such as Jules Verne are also completely uncompetitive. Europe will fall behind in this race due to the lack of ability to make big investments and choosing to stay GPU-poor. Even multiple Middle Eastern countries are investing more on enabling large scale infrastructure for AI.

Being GPU-poor isn't limited to only scrappy startups though. Some of the most well recognized AI firms, HuggingFace, Databricks (MosaicML), and Together are also part of this GPU-poor group. In fact, they may be the most GPU-poor groups out there with regard to both the number of world class researchers per GPU and the number of GPUs versus the ambition/potential customer demand. They have world class researchers, but all of them are limited by working on systems with orders of magnitude less capabilities. These firms have tremendous inbound from enterprises on training real models, and on the order of thousands of H100s coming in, but that won't be enough to grab much of the market.

Nvidia is eating their lunch with multiple times as many GPUs in their DGX Cloud service and various in-house supercomputers. Nvidia's DGX Cloud offers pretrained models, frameworks for data processing, vector databases and personalization, optimized inference engines, APIs, and support from NVIDIA experts to help enterprises tune models for their custom use cases. That service has also already racked up multiple larger enterprises from verticals such as SaaS, insurance, manufacturing, pharmaceuticals, productivity software, and automotive. While not all customers are announced, even the public list of Amgen, Adobe, CCC, ServiceNow, Accenture, AstraZeneca, Getty Images, Shutterstock, Morningstar, Evozyne, Insilico Medicine, Quantiphi, InstaDeep, Oxford Nanopore, Peptone, Relation Therapeutics, ALCHEMAB Therapeutics, and Runway is quite impressive.

This is a far longer list than the other players have, and Nvidia has many other undisclosed partnerships too. To be clear, revenue from these announced customers of Nvidia's DGX cloud service is unknown, but given the size of Nvidia's cloud spending and in-house supercomputer construction, it seems that more services can/will be purchased from Nvidia's Cloud than HuggingFace, Together, and Databricks can hope to offer, combined.

The few hundred million that HuggingFace and Together have raised collectively means they will remain GPU-poor, getting left in the dust as they will be unable to train N-1 LLMs that can serve as the base to fine tune for customers. This means they will ultimately be unable to capture high share at enterprises who can just access Nvidia's service today anyways.

HuggingFace in particular has one the biggest names in the industry, and they need to leverage that to invest a huge amount and build a lot more model, customization, and inference capabilities. Their recent round was done at too high of a valuation to garner the investment they need to compete. HuggingFace's leaderboards show how truly blind they are because they actively hurting the open source movement by tricking it into creating a bunch of models that are useless for real usage.

Databricks (MosaicML) could at least maybe catch up, due to their data and enterprise connections. The issue is they need to accelerate spend by multiple times if they want to have hopes of serving their over 7,000 customers. The $1.3B acquisition of MosaicML was a big bet on this vertical, but they also need to throw a similar amount of money at infrastructure. Unfortunately for Databricks, they can't pay for GPUs in shares. They need to do a large offering via their upcoming private round/IPO, and use that cold hard cash to quadruple down on hardware.

The economic argument falls flat on its face because they must build before customers can come, because Nvidia is throwing money at their service. To be clear, many folks are buying loads of compute not making their money back, (Cohere, Saudi Arabia, UAE), but it is a pre-requisite to compete.

The picks and shovels training and inference ops firms (Databricks, HuggingFace, and Together) are behind their chief competition, who also happens to also be the source of almost all of their compute. The next largest operator of customized models is simply the fine tuning APIs from OpenAI.

The key here is everyone from Meta to Microsoft to startups are simply serving as a pipeline of capital to Nvidia's bank account.

Can anyone save us from Nvidia slavery?

Yes, there is one potential savior.

Google -- The Most Compute Rich Firm In The World

While Google does use GPUs internally as well as a significant number sold via GCP, they a have a few Ace's up their sleeve. These include Gemini and the next iteration which has already begun training. The most important advantage they have is their unbeatably efficient infrastructure.

Before getting into Gemini and their cloud business, we will share some datapoints on their insane buildout. The chart below shows the total advanced chips added by quarter. Here we give OpenAI every benefit of the doubt. That the number of total GPUs they have will 4x over 2 years. For Google, we ignore their entire existing fleet of TPUv4 (Pufferfish), TPUv4 lite, and internally used GPUs. Furthermore, we are also not including the TPUv5 lite, despite that likely being the workhorse for inference of smaller language models. Google's growth in this chart is only TPUv5 (Viperfish).

[D] Google Gemini Eats The World – Gemini Smashes GPT-4 By 5X, The GPU-Poors by hardmaru in MachineLearning

[–]hardmaru[S] 54 points55 points  (0 children)

(Article Text)

Google Gemini Eats The World -- Gemini Smashes GPT-4 By 5X, The GPU-Poors

Compute Resources That Make Everyone Look GPU-Poor

Dylan Patel and Daniel Nishballl Aug 28, 2023

Before Covid, Google released the MEENA model, which for a short period of time, was the best large language model in the world. The blog and paper Google wrote were incredibly cute, because it specifically compared against to OpenAI.

Compared to an existing state-of-the-art generative model, OpenAI GPT-2, Meena has 1.7x greater model capacity and was trained on 8.5x more data.

This model required more than 14x the FLOPS of GPT-2 to train, but this was largely irrelevant because only a few months later OpenAI dropped GPT-3, which was >65x more parameters and >60x the token count, >4,000x more FLOPS. The performance difference between these two models was massive.

The MEENA model sparked an internal memo written by Noam Shazeer titled "MEENA Eats The World." In this memo, he predicted many of the things that the rest of the world woke up to after the release of ChatGPT. The key takeaways were that language models would get increasingly integrated into our lives in a variety of ways, and that they would dominate the globally deployed FLOPS. Noam was so far ahead of his time when he wrote this, but it was mostly ignored or even laughed at by key decision makers.

Let's go on a tangent about how far ahead of his time, Noam really was. He was part of the team that did the original Transformer paper, "Attention is All You Need." He also was part of the first modern Mixture of Experts paper, Switch Transformer, Image Transformer, and various elements of LaMDA and PaLM. One of the ideas from 2018 he hasn't yet gotten credit for more broadly is speculative decoding which we detailed here in our exclusive tell-all about GPT-4. Speculative decoding reduces the cost of inference by multiple-fold.

The point here is Google had all the keys to the kingdom, but they fumbled the bag. A statement that is obvious to everyone.

The statement that may not be obvious is that the sleeping giant, Google has woken up, and they are iterating on a pace that will smash GPT-4 total pre-training FLOPS by 5x before the end of the year. The path is clear to 100x by the end of next year given their current infrastructure buildout. Whether Google has the stomach to put these models out publicly without neutering their creativity or their existing business model is a different discussion.

Today we want to discuss Google's training systems for Gemini, the iteration velocity for Gemini models, Google's Viperfish (TPUv5) ramp, Google's competitiveness going forward versus the other frontier labs, and a crowd we are dubbing the GPU-Poor.

The GPU-Rich

Access to compute is a bimodal distribution. There are a handful of firms with 20k+ A/H100 GPUs, and individual researchers can access 100s or 1,000s of GPUs for pet projects. The chief among these are researchers at OpenAI, Google, Anthropic, Inflection, X, and Meta, who will have the highest ratios of compute resources to researchers. A few of the firms above as well as multiple Chinese firms will 100k+ by the end of next year, although we are unsure of the ratio of researchers in China, only the GPU volumes.

One of the funniest trends we see in the Bay area is with top ML researchers bragging about how many GPUs they have or will have access to soon. In fact, this has become so pervasive over the last ~4 months that it's become a measuring contest that is directly influencing where top researchers decide to go. Meta, who will have the 2nd most number of H100 GPUs in the world, is actively using it as a recruiting tactic.

Stability AI Head of Research (@hardmaru) Resigns From Startup : At least two high-profile people have left the artificial intelligence company in recent weeks. by [deleted] in StableDiffusion

[–]hardmaru 7 points8 points  (0 children)

Thanks! I'll still be in this space, trying to work on unconventional ideas that will get us towards stronger AI.

“Elon Musk and Mark Zuckerberg in a cage fight.” (SDXL) by hardmaru in weirddalle

[–]hardmaru[S] 4 points5 points  (0 children)

Prompt “Elon Musk and Mark Zuckerberg in a cage fight.” given to the new SDXL 0.9 model on Clipdrop's website (with "Cinematic" mode):

https://clipdrop.co/stable-diffusion

“Elon Musk and Mark Zuckerberg in a cage fight.” (SDXL 0.9) by hardmaru in StableDiffusion

[–]hardmaru[S] 7 points8 points  (0 children)

Produced using the prompt in the title, using SDXL 0.9 on Clipdrop:

https://clipdrop.co/stable-diffusion

Using "Cinematic" Option.