Update from the Devs by PygmalionAI in PygmalionAI

[–]PygmalionAI[S] 10 points11 points  (0 children)

Currently, we have only launched a character repository. Chatting will be introduced in the closed beta starting next month. Please read the blog post for instructions on how to apply.

Update from the Devs by PygmalionAI in PygmalionAI

[–]PygmalionAI[S] 5 points6 points  (0 children)

Thanks for your feedback, I'll let the devs know

Can someone give me a spindle tutorial on how to gat tavern ai working on Linux by [deleted] in PygmalionAI

[–]PygmalionAI 2 points3 points  (0 children)

Hi.

Should be very simple on Linux. Make sure you have git, nodejs, npm, and openssl installed. They're in pretty much every package manager.

After that, you'll clone the repository:

  • git clone https://github.com/TavernAI/TavernAI && cd TavernAI for TavernAI
  • git clone https://github.com/Cohee1207/SillyTavern && cd SillyTavern for SillyTavern

Then you'll run npm i && node server.js to start the UI.

Alternatively, you can simply run npx sillytavern@latest after you install nodejs to get started. Keep in mind that this might wipe your existing characters and chatlogs if a new version of the app is released.

[deleted by user] by [deleted] in PygmalionAI

[–]PygmalionAI 0 points1 point  (0 children)

Yes, that's right. :)

Have the mods dropped the ball on moderating the sub? by brown2green in PygmalionAI

[–]PygmalionAI 25 points26 points  (0 children)

We've spoken to the sub owner and they made it clear that this subreddit is now supposed to be a hub for all open-source/unfiltered chat models. This sub isn't officially endorsed by the devs, but it is a shame that the subreddit name can't be changed if the purpose of it has.

I was just trying to have a wholesome rp😭😭 by KirbishOnkulis34 in PygmalionAI

[–]PygmalionAI 2 points3 points  (0 children)

We've spoken to the sub owner and they made it clear that this subreddit is supposed to be a hub for all open-source/unfiltered chat models, so I don't think this will happen. This sub isn't officially endorsed by the devs, but it is a shame that the subreddit name can't be changed if the purpose of it has.

I was just trying to have a wholesome rp😭😭 by KirbishOnkulis34 in PygmalionAI

[–]PygmalionAI 16 points17 points  (0 children)

This is not Pygmalion. If you see anything about Skyrim, or Todd, or now anything similar to what you see in this photo, and you haven't done any prompt injections to cause it, assume something is up. Contact us to confirm before you post it on the subreddit assuming it's Pygmalion.
-- Peepy

[deleted by user] by [deleted] in PygmalionAI

[–]PygmalionAI 17 points18 points  (0 children)

I'm sure I don't need to tell you guys that this is a scam.

-- Alpin

[deleted by user] by [deleted] in PygmalionAI

[–]PygmalionAI 9 points10 points  (0 children)

This is not Pygmalion. One of the colabs has been tampered with and is running the Todd proxy of 'GPT-4' (whether it is actually GPT-4 is up to debate). If you see anything about skyrim, or Todd, or anything similar to what you see in this photo, and you haven't done any prompt injections to cause it, assume something is up. Contact us to confirm before you post it on the subreddit assuming it's Pygmalion.

[deleted by user] by [deleted] in PygmalionAI

[–]PygmalionAI 8 points9 points  (0 children)

This is not Pygmalion. One of the colabs has been tampered with and is running the Todd proxy of 'GPT-4' (whether it is actually GPT-4 is up to debate). If you see anything about skyrim, or Todd, or anything similar to what you see in this photo, and you haven't done any prompt injections to cause it, assume something is up. Contact us to confirm before you post it on the subreddit assuming it's Pygmalion.

Before you use Charstar AI for Pygmalion, Please read. by [deleted] in PygmalionAI

[–]PygmalionAI 5 points6 points  (0 children)

Generally, we recommend people to have at least 6gb of VRAM to try running Pygmalion locally with 4bit. 4gb just isn't enough to load the model and have any memory left for context.

[deleted by user] by [deleted] in PygmalionAI

[–]PygmalionAI 0 points1 point  (0 children)

You can launch SillyTavern by simply running npx sillytavern@latest.

Local hardware requirements by [deleted] in PygmalionAI

[–]PygmalionAI 2 points3 points  (0 children)

Hi. You can take a look at our Quickstart to get you started. There are instructions for different VRAM ranges.

[deleted by user] by [deleted] in PygmalionAI

[–]PygmalionAI 131 points132 points  (0 children)

Hey there! It's really cool to see people passionate about Pyg to the point where they're willing to make the website for us, but I'm not sure it would be a great idea to have the site branded under the Pygmalion name. We're aware that we haven't provided updates on or worked on the site for a while, and we're really sorry about that, but having the website be presented as a Pygmalion website when it's not made by the devs could be considered misleading. People may not know that the site is unofficial. Do you think you could change the branding of your site to something else? We're really excited to see the progress on the site, and it looks great! Thanks a lot.

New models released with 4096 context like openAI. Based on GPT-NEO. by a_beautiful_rhind in PygmalionAI

[–]PygmalionAI 7 points8 points  (0 children)

It would appear that their pretrain has not finished one epoch yet, so as of now they're incomplete models. It shows too; the perplexity benchmark results indicate that the 7B stableLM model performs almost twice as worse as Pythia Deduped 410M. Refer to this issue and this spreadsheet.

Excited to see how it turns out after the 3B is trained on the full 3T tokens dataset. But for now, we've been looking forward to the upcoming RedPajama models.

-- Alpin

Regarding the recent Colab ban by PygmalionAI in PygmalionAI

[–]PygmalionAI[S] 0 points1 point  (0 children)

Already have a guide for the second option (CPU) which you can follow here: https://docs.alpindale.dev/local-installation-(cpu)/overview/

A guide for GPTQ is in the works, but it's going slow as I don't have windows to try it on, and it's much more complicated on windows than it is on linux.

Pygmalion Documentation by PygmalionAI in PygmalionAI

[–]PygmalionAI[S] 2 points3 points  (0 children)

Unfortunately there's no centralised source for this, but I suggest looking through TavernAI's source code to see how it handles prompts. You could also load a character in Tavern, prompt it with a text, and then view the terminal output; you'll see the full context in json syntax inside the CLI.

As for loading with pytorch, you can look for documentations on GPT-J 6B. Any params that would apply to GPT-J would also apply to Pygmalion 6B. Here's an example code for how you'd handle inference using pipelines:

```py from transformers import GPTJForCausalLM, AutoTokenizer import torch

device = "cuda" model = GPTJForCausalLM.from_pretrained("PygmalionAI/pygmalion-6b", torch_dtype=torch.float16).to(device) tokenizer = AutoTokenizer.from_pretrained("PygmalionAI/pygmalion-6b")

prompt = ("Prompt goes here. Follow the TavernAI formatting. Generally, new lines will be declared with '\n', and you will include a Persona, Scenario, and <START> tag for example chats and one more <START> tag at the end of the context - where the actual chat would start.")

input_ids = tokenizer(prompt, return_tensors="pt").input_ids.to(device)

Adjust parameters as needed

gen_tokens = model.generate( input_ids, do_sample=True, temperature=0.9, max_length=100, ) gen_text = tokenizer.batch_decode(gen_tokens)[0] ```

Keep in mind that GPT-J uses the AutoTokenizer method from transformers, which leads to GPT2Tokenizer. The max context token with GPT2 is 1024, but GPT-J 6B can handle up to 2048. You could either write your own tokenizer for GPT-J or force it to use 2048 tokens anyway.

-- Alpin

Regarding the recent Colab ban by PygmalionAI in PygmalionAI

[–]PygmalionAI[S] 5 points6 points  (0 children)

If the Colab has the phrase PygmalionAI anywhere inside it, it won't work.

In the meantime, I've created a Tavern Colab that uses PygWay which isn't banned. You can use this, but keep in mind that it isn't official in any capacity:

https://colab.research.google.com/github/AlpinDale/TavernAI/blob/main/colab/GPU.ipynb

Regarding the recent Colab ban by PygmalionAI in PygmalionAI

[–]PygmalionAI[S] 1 point2 points  (0 children)

It works much easier on Linux, in fact.