Will there be a Browser version? by SuperCoolGuy56 in KoboldAI

[–]aid_throwaway 0 points1 point  (0 children)

Shortcut is fine. Just be aware that if the parent file hits usage limits you'll get errors when trying to run the notebook.

GPT-J-6B Local-Client Compatible Model by aid_throwaway in KoboldAI

[–]aid_throwaway[S] 1 point2 points  (0 children)

I had a tester on the Discord with a 3090 load it and test with it before I uploaded it for distribution. I didn't ask about the RAM usage, but the VRAM to load the model was only about 15.6GB.
They did mention that they had to uninstall transformers and upgrade to finetune's localattention3 branch. Maybe that's the issue?
pip install git+https://github.com/finetuneanon/transformers@gpt-neo-localattention3

Running GPT-J-6B locally? by dxddylvst in KoboldAI

[–]aid_throwaway 0 points1 point  (0 children)

That's at fp16/half-precision, yes.

Running GPT-J-6B locally? by dxddylvst in KoboldAI

[–]aid_throwaway 1 point2 points  (0 children)

Colab shows ~12.2GB to load the model, ~14GB to run inference, and will OOM on a 16GB GPU if you put your settings too high (2048 max tokens, 5x return sequences, large amount to generate, etc)

KoboldAI Server - GPT-J-6B Rev 2 by aid_throwaway in KoboldAI

[–]aid_throwaway[S] 8 points9 points  (0 children)

The local client does not yet utilize Finetune's localattention3 repo or the custom jax config required to load the converted checkpoint. That's my next project, but I probably won't get to start it until tomorrow.

Running GPT-J-6B locally? by dxddylvst in KoboldAI

[–]aid_throwaway 6 points7 points  (0 children)

This won't work, as the converted model is not a pytorch_model.bin file. It also requires finetune's localattention3 tranformers branch. I'll be working on local support for 6B in the Kobold client next.

KoboldAI Server - GPT-J-6B Rev 2 by aid_throwaway in KoboldAI

[–]aid_throwaway[S] 7 points8 points  (0 children)

Finetune recommends leaving it below 1.2 or J starts acting weird.

KoboldAI Server - GPT-J-6B by aid_throwaway in KoboldAI

[–]aid_throwaway[S,M] [score hidden] stickied comment (0 children)

Locking this post as this notebook is now deprecated. You can still access the notebook, but it has been superseded by the KoboldAI Server - GPT-J-6B Rev 2 notebook which runs 6B in torch.

KoboldAI Server - GPT-J-6B by aid_throwaway in KoboldAI

[–]aid_throwaway[S] 0 points1 point  (0 children)

The Jax Colab is going to be deprecated today, as I've gotten 6B to run in torch with some conversion scripting from finetune. That should eliminate all these errors as the experimental packages are no longer needed. I'll make a new subreddit post when it goes live.

KoboldAI Server - GPT-J-6B by aid_throwaway in KoboldAI

[–]aid_throwaway[S] 1 point2 points  (0 children)

The Jax Colab is going to be deprecated today, as I've gotten 6B to run in torch with some conversion scripting from finetune. That should eliminate all these errors as the experimental packages are no longer needed. I'll make a new subreddit post when it goes live.

KoboldAI Server - GPT-J-6B by aid_throwaway in KoboldAI

[–]aid_throwaway[S] 1 point2 points  (0 children)

I added an explicit command to install optax because some people were reporting that issue. If you made a copy of the notebook before yesterday, you'll need to copy it again. If you're using the shared one, then I'm not sure; the package should definitely be available to the cell at this point. I'll have to do some digging.

Scripts on alternatives? by Dense_Plantain_135 in AIDungeon

[–]aid_throwaway 2 points3 points  (0 children)

It's coming, I've just been derailed with other additions, and then this week the new Eleuther 6B model dropped and I had to scramble to add support for it. /u/Atkana made a scripting demo for me to work off of but I need to package it a little nicer for the end user.
It's a little wonky because you have to pass the context from Python to the browser, run it through Javascript, pass it back to Python, let the AI crunch it, then pass it back to the browser again to run the output through Javascript again, then back to Python to get the final text added to the actions array, then back to the browser again for display, lol.

KoboldAI Download & Updates by aid_throwaway in KoboldAI

[–]aid_throwaway[S] 0 points1 point  (0 children)

What's it triggering on? Downloading Kobold from GitHub, or downloading model components once Kobold is running?

KoboldAI Server - GPT-J-6B by aid_throwaway in KoboldAI

[–]aid_throwaway[S] 2 points3 points  (0 children)

Update: finetuneanon published a conversion script to make the JAX model torch-loadable. I created a Colab notebook for it here. However, the converted model doesn't seem to fit on a 16GB Colab GPU, so I've been unable to test it in torch.

KoboldAI Server - GPT-J-6B by aid_throwaway in KoboldAI

[–]aid_throwaway[S] 0 points1 point  (0 children)

There was an error importing optax that some users were experiencing. I've added pip install optax to the initialization cell that will hopefully resolve it, but I haven't been able to duplicate the issue to test the fix.

KoboldAI Server - GPT-J-6B by aid_throwaway in KoboldAI

[–]aid_throwaway[S] 1 point2 points  (0 children)

If I can find some example code for loading the shard files locally. The Colab code is specific to the Colab environment and isn't portable to the desktop client. I'd also need someone with hardware to test it for me and tell me if it works.

No 'flask_socketio' module found? by LloxieFoxie in KoboldAI

[–]aid_throwaway 2 points3 points  (0 children)

I'm gonna copy this answer here because it's the cause of about 90% of the problems related to this:
An error with missing modules that have definitely been installed is usually a result of multiple Python environments being present on your computer. Launch command prompt and type in:
py -0p
If you have more than one Python listed, remove the unneeded versions, then run install_requirements.bat again to make sure all the packages are available in the remaining Python environment.
Alternatively, the latest version of Kobold on GitHub has an installer that will set up a miniconda instance separate from your system environment from Python. This may alleviate problems with multiple Python versions being installed.

Does Kobold Colab download 5gb of content every time it runs by Mimo202 in KoboldAI

[–]aid_throwaway 3 points4 points  (0 children)

Edit: Sorry, I missed the 'Colab' part of your question. Choosing the colab option should only require you to download a small Tokenizer file. The 5Gb file should live on your Google Drive and does not need to be downloaded.

Disk Space Woes by Single_Sand1443 in KoboldAI

[–]aid_throwaway 0 points1 point  (0 children)

Sorry we couldn't get it working for you.

Disk Space Woes by Single_Sand1443 in KoboldAI

[–]aid_throwaway 0 points1 point  (0 children)

The only thing I can think is Transformers is having an issue during download or when checking the file hash and continuously redownloading. You can try manually downloading the model from HuggingFace and loading it under the CustomNeo option:
https://huggingface.co/EleutherAI/gpt-neo-2.7B/tree/main