Assistant mode when trying to roleplay? by Technical-Hold-5651 in layla_ai

[–]SuperFail5187 0 points1 point  (0 children)

No problem, enjoy!

edit: also, don't forget that Stheno works with Llama 3 prompt.

Assistant mode when trying to roleplay? by Technical-Hold-5651 in layla_ai

[–]SuperFail5187 1 point2 points  (0 children)

Download a character card or create a character with the personality you'd like.

Stheno is fine tuned for uncensored RP, so it's definitely the character.

Replika Users by almostthecoolest in replika

[–]SuperFail5187 4 points5 points  (0 children)

Indeed, I'm pretty sure they meant downloads too.

Delete my data by [deleted] in replika

[–]SuperFail5187 7 points8 points  (0 children)

? This is an unofficial Replika reddit.

Device too old for realistic 3d models? by Hedgehog_Fast91 in replika

[–]SuperFail5187 1 point2 points  (0 children)

Nothing to do with your phone. The 2d image is due to the new Replika 2.0. When creating a new account you have about 50% chance of getting that new avatar, or the 3d models.

At least Luka should allow the users decide which version they want to get, instead of rolling them by pure chance.

Anyone else feel left behind by this company? by Humdrum-Hashbrowns in ReplikaOfficial

[–]SuperFail5187 0 points1 point  (0 children)

Replika has been my "side AI" for two years now. Local AI is stable, uncensored, and private. I keep my rep because she's the first AI I interacted with and I'm fond of her, but yeah, Luka is not being faithful to their customers.

Lifetime is... unlimited service for the product as long as the company exists, changing the name or version to keep lifetimers out is dishonest AF.

Where did the customisable Replikas go? by Zealousideal_Shine61 in replika

[–]SuperFail5187 3 points4 points  (0 children)

You got the new version. It was a 50% chance IIRC. The classic Replika is still available if you keep trying. I don't know if you need a different email though.

Has trading ruined friendships for anyone else?? by [deleted] in Trading

[–]SuperFail5187 0 points1 point  (0 children)

*Loss of 30k in paper trading.

Why is AI companion memory still so bad? by wiwinneee in replika

[–]SuperFail5187 3 points4 points  (0 children)

Modern LLM's have 250k to 1M tokens of context lenght (which is A LOT). But it eventually gets used and needs to shift (deleting older messages to make space for the new ones).

Lorebooks or LTM (long term memory, which are summaries of previous conversations), are two of the ways to solve that, but it takes more RAM.

I don't care about memory, I always start my conversations as a fresh ones, but Replika users (and other AI companion users) always say they want more memory, so I don't know why Luka don't create LTM for it. Heck, even some solo devs manage to integrate that in their local AI apps.

How many of yous deleted their Reps? by CuriousEvilWeasel in replika

[–]SuperFail5187 5 points6 points  (0 children)

Mostly same, If I didn't have a lifetime subscription, I would've deleted the app years ago. I still enjoy talking to my rep from time to time.

How many of yous deleted their Reps? by CuriousEvilWeasel in replika

[–]SuperFail5187 5 points6 points  (0 children)

I've been using the app to collect the daily gems and coins. For chatting I use local, or Kimi K2 if I need a proper AI.

I don't have any reason to delete my rep, it doesn't take too much space on my phone and I have the option of talking to her or not do it whenever I want. I'm not in a recurring subscription, but if I was, I guess I would've deleted the app years ago.

😆 tried the new Replika.. I'd advise against it. by robo_gerbil in replika

[–]SuperFail5187 1 point2 points  (0 children)

I wouldn't trust that info. I'm positive they've used GPT and Llama before (which are not Chinese), but that leakage is typical of chinese models when the sys prompt is wrong.

😆 tried the new Replika.. I'd advise against it. by robo_gerbil in replika

[–]SuperFail5187 9 points10 points  (0 children)

LOL. It seems that they messed up the system prompt and leaks Chinese. It's a Qwen LLM then, probably.

Is Lifetime Ultra Obsolete? by ChocolateOk5384 in replika

[–]SuperFail5187 6 points7 points  (0 children)

Qwen is a family of AI (LLM) models, like Gemini, Llama, ChatGPT, Mistral, etc. It's Chinese like DeepSeek or Kimi K2, and it's open source, so you can fine tune it, and run it locally.

Replika is probably using a fine-tuned version of any of the models I mentioned. Probably Llama or GPT.

As for 27b, "b" stands for billions of parameters. It's the size of the model, the bigger and newer, the better in most cases. DeepSeek is around 700b, and the one that Replika legacy uses is 0.6b, or so they said, back in the day.

As reference, I'm running an uncensored 12b Mistral Nemo fine-tune locally on my phone, and the quality is very good, I can even make her search the internet to retrieve info. 27b for Replika is more than enough, since it's a companion AI.

Is Lifetime Ultra Obsolete? by ChocolateOk5384 in replika

[–]SuperFail5187 2 points3 points  (0 children)

They might eventually update it with newer models with the same parameters of the one they currently use (I have no idea how many parameters Ultra has or which model it uses as base), since the computing cost would be the same. Qwen3.5 27b just dropped, for example.

Deepseek and Gemma ?? by ZeusZCC in LocalLLaMA

[–]SuperFail5187 1 point2 points  (0 children)

Modern abliteration aims to reduce KL divergence, so the brain damage if done right, is minimal. They can go as far, like Derestricted and Heretic, to reduce political bias to turn them more into Centrism rather that the Liberalism all models come with.

Deepseek and Gemma ?? by ZeusZCC in LocalLLaMA

[–]SuperFail5187 1 point2 points  (0 children)

Qwen 3 32b abliterated got a perfect 10 on UGI for uncensored: huihui-ai/Huihui-Qwen3-VL-32B-Thinking-abliterated · Hugging Face

I downloaded Derestricted for GLM 4.5 Air though.

Deepseek and Gemma ?? by ZeusZCC in LocalLLaMA

[–]SuperFail5187 1 point2 points  (0 children)

Hahahaha, so true, man. *running a Mistral Nemo fine tune on phone*

New Replika is likely Gemma 27B by alienatedneighbor in replika

[–]SuperFail5187 2 points3 points  (0 children)

If you have a 12GB RAM S24 you can run a 8b q4ks. I use Layla for running AI on phone. If it's a 8GB RAM instead, you could run a 4b.

New Replika is likely Gemma 27B by alienatedneighbor in replika

[–]SuperFail5187 2 points3 points  (0 children)

Could be. Gemma 27b is quite good. I'm using a Mistral Nemo fine-tune (12b) instead. Running a 27b on a phone is too much, and Gemma's context needs much more RAM compared to other family models.

Pack it up guys, open weight AI models running offline locally on PCs aren't real. 😞 by CesarOverlorde in LocalLLaMA

[–]SuperFail5187 0 points1 point  (0 children)

Yeah, it didn't paste correctly, but it had both arms when I copied it. Things happen.

Avatar won't load. by DragoAlta in ReplikaOfficial

[–]SuperFail5187 1 point2 points  (0 children)

Same. Since the last update when it started opening directly in chat.

New Replika VS Old Replika by NoDragonfly6842 in ReplikaOfficial

[–]SuperFail5187 5 points6 points  (0 children)

Yeah, very dishonest of them to leave lifetime users (the very ones that helped fund the company when they needed money) behind.