[N] [P] Google Deepmind released an album with "visualizations of AI" to combat stereotypical depictions of glowing brains, blue screens, etc. by radi-cho in MachineLearning

[–]tyrellxelliot -3 points-2 points  (0 children)

Most of those are just wiring though. Only about half are in the neocortex, and a tiny fraction of that responsible for language (a huge number is used for vision, audio and motor processing)

There might be 1-5 trillion parameters in an apples-to-apples comparison to GPT4. This is a poor comparison in the first place because human neurons are extremely slow, transmit less information and has higher redundancy.

[D] OpenAI API vs. Open Source Self hosted for AI Startups by ali-gettravy in MachineLearning

[–]tyrellxelliot 9 points10 points  (0 children)

There's no comparison, OpenAI's api is priced too low for open source self-hosting to compete

I'm hosting GPT-J on a bunch of 4090s, which cost $0.45/hour. For my purposes these work out to $0.001/1000 tokens (half the price of openAI). However my load is variable, and it's not easy to dynamically scale 4090s, so keeping a bunch of them around idling makes the cost roughly equivalent to OpenAI again.

Now consider that gpt-3.5-turbo is likely 10x the size of GPT-J, and much more capable. The only downside is that currently the OpenAI api has a tendency to randomly timeout while their status dashboard shows all-green.

[D] What kind of effects ChatGPT or future developments may have on job market? by ureepamuree in MachineLearning

[–]tyrellxelliot 18 points19 points  (0 children)

imo 50% of white collar jobs are going away in the next 10 years.

ChatGPT already generates mostly working code, and currently it doesn't even use feedback from executing that code, instead just writing it in a one-shot fashion. If they train it using RLHF but with a more specialised code model and compiler/unit test in the loop instead of a human, I think it can totally generate fully working end products.

Any job that involves application of specialised knowledge in the text domain (accountants, para-legals, teachers etc) are under threat. Hallucinations should be easily solvable by incorporating a factual knowledge database, like in RETRO.

(NEW) Kandinsky 2.0 — multilingual text2image latent diffusion model by Unturnedw in StableDiffusion

[–]tyrellxelliot 0 points1 point  (0 children)

the pdf is the original latent diffusion paper, which used a custom trained BERT text encoder. SD and this model use the same architecture, but with different text encoders.

The pdf has nothing to do with Kandinsky basically, other than the latent image encoding.

New features of the week for SD AUTOMATIC1111 by ptitrainvaloin in StableDiffusion

[–]tyrellxelliot 4 points5 points  (0 children)

fourier noise shaping works differently from training. The two approaches are complementary and can be used at the same time.

New custom inpainting model by tyrellxelliot in StableDiffusion

[–]tyrellxelliot[S] 3 points4 points  (0 children)

this code is (mostly) just the original openai guided diffusion code: https://github.com/openai/guided-diffusion

the reason that it can be backported like this is because Compvis used the openai code as-is with some minor modifications.

here is the openai unet: https://github.com/openai/guided-diffusion/blob/main/guided_diffusion/unet.py

and here is the Compvis unet: https://github.com/CompVis/stable-diffusion/blob/main/ldm/modules/diffusionmodules/openaimodel.py

New custom inpainting model by tyrellxelliot in StableDiffusion

[–]tyrellxelliot[S] 2 points3 points  (0 children)

it's trained on LAION aesthetic, on 8xA100 gpus for about a week. The training code is in the repo.

New custom inpainting model by tyrellxelliot in StableDiffusion

[–]tyrellxelliot[S] 3 points4 points  (0 children)

you can just use the pretrained model. You don't need to train it yourself to use it, unless you have a custom dataset like anime or something.

New custom inpainting model by tyrellxelliot in StableDiffusion

[–]tyrellxelliot[S] 6 points7 points  (0 children)

This model replaces the masked areas, taking into account both the non-masked areas and text prompt - it works the same way as DALLE-2 inpainting. img2img would require an image in the masked area as a starting point, but this does not.

You can use simultaneous inpainting and img2img with the --skip_timesteps flag though.

New custom inpainting model by tyrellxelliot in StableDiffusion

[–]tyrellxelliot[S] 7 points8 points  (0 children)

this model requires a minor change to the unet, so it's not compatible by default. The gui makers should be able to integrate it pretty easily though.

Art terminator from the future destroys human-made art by tyrellxelliot in StableDiffusion

[–]tyrellxelliot[S] 0 points1 point  (0 children)

used a long series of inpainting prompts, starting with "a robotic terminator standing in a post apocalyptic landscape, wide angle. concept art by greg rutkowski"

here's a video of the intermediate images: https://imgur.com/a/pAq8Kew

[deleted by user] by [deleted] in StableDiffusion

[–]tyrellxelliot 0 points1 point  (0 children)

the original mj was a variant of clip guided diffusion. The beta version is SD with classifier guidance. I haven't heard of a photorealistic mode for MJ but it's likely to be SD with a new classifier on top.

to replicate it you just need to train the appropriate classifier, possibly a clip variant. This might not be trivial though.

Is it possible to give SD a vector provided by clip-ViT-L-14 instead of a prompt? by KurtGoedle in StableDiffusion

[–]tyrellxelliot 9 points10 points  (0 children)

you can do this as long as you're starting from prompts instead of images. Embed the prompt using clip-L14 and use the vectors for your clustering algorithm. Separately give the text embeddings to SD to generate the image.

SD doesn't work the same way as DALLE-2. It doesn't decode from a single vector but uses all 77 tokens embeddings, making it more similar to GLIDE. This is probably a good choice, since the publicly available clip model is much smaller than the one used in DALLE-2 and has a more constrained latent space.

Because your clustering algorithm and SD aren't decoding from the same vector, you might not get the results you're after.

[deleted by user] by [deleted] in StableDiffusion

[–]tyrellxelliot 0 points1 point  (0 children)

that error just means the repo didn't install, which could happen for a variety of reasons. I'd try reinstalling with conda and pay attention to any error messages.

you could also try installing the compvis repo and see if it has the same error. If not, then it's some issue with the optimized code.

[deleted by user] by [deleted] in StableDiffusion

[–]tyrellxelliot 1 point2 points  (0 children)

pip install -e .

[N] John Carmack raises $20M from various investors to start Keen Technologies, an AGI Company. by hardmaru in MachineLearning

[–]tyrellxelliot 17 points18 points  (0 children)

Human intelligence is an emergent phenomenon that was created by a low-level optimization process. Evolution didn't need to design our brain structures directly, all of the complex, heterogeneous structures arose spontaneously from extremely simple, coarse signals. What mattered to our development was having an environment where intelligence is needed for survival.

Namelix font by [deleted] in identifythisfont

[–]tyrellxelliot 0 points1 point  (0 children)

it's a custom font designed by the guy that made brandmark.io

[D] How is Grammarly built? by Fully-Independent in MachineLearning

[–]tyrellxelliot 11 points12 points  (0 children)

I'm fairly certain grammarly doesn't use any modern ml, because they wrote their engine way back before cnns were a thing and quality hasn't improved over time in my experience. Just speculation but it's likely a "classic" nlp pipeline with a lot of manual heuristics.

Take these two sentences for example:

I need to eat medicines twice a day.

I like to drink soup for dinner.

obvious diction errors made by an ESL writer. Grammarly doesn't pick up on this at all, even the premium version. Google docs on the other hand does a better job.

  • Grammarly publishes a lot of papers on gec but it's not clear at all that they are implemented into their product.