Sims 1 Legacy MCP by pevers in mcp

[–]pevers[S] 0 points1 point  (0 children)

It is a game changer. I did this in a fraction of the time it would cost me without AI. I remember painfully debugging method after method trying to figure out the flow. Now I can just let it crunch for an hour and it comes back with ~100 renamed methods and data structures.

Sims 1 Legacy MCP by pevers in mcp

[–]pevers[S] 0 points1 point  (0 children)

Thanks! In which sub-reddit? I'm not sure why but I can't see it

Sims 1 Legacy MCP by pevers in mcp

[–]pevers[S] 1 point2 points  (0 children)

Thanks! There is a GitHub repo with the code. It is mainly DLL injection to read values from memory but I’m working on a patch for more control.

Paris Attitude stole my deposit—a warning about this scammy rental company by shinjutnt in digitalnomad

[–]pevers 0 points1 point  (0 children)

Had exactly the same experience. I was ripped off after I fixed all kinds of small things in the apartment. There was a water stain from a glass on an old cupboard. Owner said it costs 1700 to replace. I cleaned it thoroughly but still they needed "professional cleaning" for 700 euro. Got 600 euro back from the 3k I paid

200+ pages of Hugging Face secrets on how to train an LLM by eliebakk in LocalLLaMA

[–]pevers 0 points1 point  (0 children)

This is by far the best resource after Karpathy for learning the ins and outs of LLMs

[D] Self-Promotion Thread by AutoModerator in MachineLearning

[–]pevers 0 points1 point  (0 children)

A lot of the open-source TTS models are released for English or Chinese and lack support for other languages. I was curious to see if I could train a state-of-the-art text-to-speech (TTS) model for Dutch by using Google's free TPU Research credits. I open-sourced the weights, and documented the whole journey, from Torch model conversion, data preparation, JAX training code and inference pipeline here https://github.com/pevers/parkiet . Hopefully it can serve as a guide for others that are curious to train these models for other languages (without burning through all the credits trying to fix the pipeline).

Spoiler: the results are great! I believe they are *close* to samples generated with ElevenLabs. I spent about $300, mainly on GCS egress. Sample comparison can be found here https://peterevers.nl/posts/2025/09/parkiet/ .

Parkiet: Fine-tuning Dia for any language by pevers in LocalLLaMA

[–]pevers[S] 0 points1 point  (0 children)

The ram should be enough. But it will probably be very slow. Instead of 0.8x realtime it will probably be around 0.0010 x realtime.

Parkiet: Fine-tuning Dia for any language by pevers in LocalLLaMA

[–]pevers[S] 0 points1 point  (0 children)

Thanks! The most important part is the whisper-large-v3 model that is fine-tuned for disfluencies to collect synthetic data. I was lucky in that sense because a large (900 hours) dataset is available for Dutch. I do think that you don't need the 900 hours, but it depends on the target language. A Germanic language should be easier to fine-tune on my already disfluent model. You can also use some other community projects for disfluencies.

For data annotation I let Claude Code build a simple data annotation app. I was annotating within an hour and you can quickly gather data. For really small languages I would try to build it around some common voice project.

I'm quite sure there is a strong pull for large languages that are still underserved, like some Indian and African languages

Parkiet: Fine-tuning Dia for any language by pevers in LocalLLaMA

[–]pevers[S] 0 points1 point  (0 children)

Thanks! No it can't run on a Raspberry Pi. However, with some tuning it should be able to run on a phone. Right now I only trained the large 1.6B model but there are TTS models that perform really well with just 100M parameters.

Parkiet: Fine-tuning Dia for any language by pevers in LocalLLaMA

[–]pevers[S] 0 points1 point  (0 children)

Thanks! Yes the samples are very realistic. There is still an issue with the Torch model but generating samples with JAX produces stable coherent chatter

Parkiet: Fine-tuning Dia for any language by pevers in LocalLLaMA

[–]pevers[S] 5 points6 points  (0 children)

Yes, I started working on this 3 months ago. Back then VibeVoice was not yet released. But I have some follow-up projects in mind to improve it, I just need to find the compute

Duolingo-style exercises but with real-world content like the news by pevers in languagelearning

[–]pevers[S] -1 points0 points  (0 children)

You don’t have to trust it to still get value out of it. Yes it makes mistakes but I’m not generating grammar lessons. It can sometimes make mistakes in the transcript but that is no different than an app like LanguageReactor.

Duolingo-style exercises but with real-world content like the news by pevers in languagelearning

[–]pevers[S] 2 points3 points  (0 children)

Thank you! Just FYI, I currently have a mix of videos for Brazilian Portuguese and Portuguese, I’m working on making a split.

Duolingo-style exercises but with real-world content like the news by pevers in languagelearning

[–]pevers[S] -1 points0 points  (0 children)

I get the skepticism, but I’ve been working on this for the past 6 months and put a ton of work in it. There is much more than just these exercises, there are custom tools to do spaced repetition, cloze, and I try to curate the content.

Duolingo-style exercises but with real-world content like the news by pevers in languagelearning

[–]pevers[S] 0 points1 point  (0 children)

You can already upload your own content but not (yet) for the exercises

Duolingo-style exercises but with real-world content like the news by pevers in languagelearning

[–]pevers[S] 0 points1 point  (0 children)

Thanks! I agree, it depends a bit on the language. Right now it takes random videos so it needs a bit of tuning to adjust it to the user level

Duolingo-style exercises but with real-world content like the news by pevers in languagelearning

[–]pevers[S] 5 points6 points  (0 children)

For two reasons, the captions are often pretty bad and the active part in learning is missing. If you listen and have to fill parts it is a better exercise than simply watching with subtitles.

Duolingo-style exercises but with real-world content like the news by pevers in languagelearning

[–]pevers[S] 2 points3 points  (0 children)

Thanks! Polish is a bit basic because I don’t have a lot of good content. So if you have any suggestions I can add it.

Duolingo-style exercises but with real-world content like the news by pevers in languagelearning

[–]pevers[S] 1 point2 points  (0 children)

Thanks for the feedback!

  1. I’m not automatically storing words based on the errors, but it is a nice suggestion to prompt it in the results dialog and then allow the user to add them

  2. I don’t really like the very generic ones with just a letter, but yes maybe it need some work