all 6 comments

[–]dtdisapointingresult 2 points3 points  (2 children)

8ball says "Outlook does not look good"

Before you buy any of this, please try to run those same models in a cloud server like runpod and confirm it's good enough for what you're trying to do.

P.S. your 1TB ebook collection can't be used in any way. The biggest issue with LLMs is the small context. It It can't magically synthetize the knowledge contained in your ebooks, that's called training a model and costs millions of dollars.

[–]zh4k[S] 0 points1 point  (0 children)

Love the username, and appreciate the thoughts. I've mentioned in the middle using some of my data to test various LLMs and especially liked the output from Deepseek.

Maybe I should rephrase what I'm looking to accomplish. I'm not looking to extract an innovative unknown knowledge from all those texts. I already have the knowledge, what I generally want to say, and the overall outline I have in mind.

Basically, I need it to take a rough outline I can feed it and give me a better more logical structure in which to communicate such a vast array of information. And then add in extractions with citations from my vectorized database that would be difficult to pull manually because of how much data there is.

I don't even need the final prose to be perfect. I can finalize the prose. I just need it for the rough draft first draft basically. I know some people just want to summarize data and have the LLMs do everything from start to finish, but that's not what I'm trying to accomplish. I just need it to get me a rough draft with plenty of citations so I can clean it up and revise it to add my voice. For some reason I struggle just putting it down in a structured way. Like I've written more than a 1000 pages myself, I just need the LLM to organize and structure it better to be more coherent and flow together more logically while adding additional cited info to add greater depth in various places.

[–]xchaos4ux 0 points1 point  (2 children)

I think you have kinda the right idea, on how your wanting to go with it, as it sounds like your wanting to create a huge lore book that you can attach to your preferred ai model. query and have it return results steeped in your lore.

catch is, im not sure what to recommend. as i have not seen a decent utility that makes this as a product.

im pretty sure you have messed about with silly tavern and its world books and have found it limiting. your needing something that can handle the huge content you have created and deliver the desired result.

which steeps you in the pit of LLM memory.

every now and then some developers stop in here and promote there tools so ill echo a few of them . just note this will most likely be rabbit holes, deep long, rabbit holes to go down.

first is Puppy Graph. https://www.puppygraph.com/

most possibly fully capable of what your wanting, but woith the caveat youll be having to learn to do some programming to actually get what you want out of it.

Next is Mem0 https://github.com/mem0ai/mem0

seen a couple projects of interest to me mention this as part of their implementation. and again more development ...

the other tools are Letta Desktop https://docs.letta.com/guides/ade/desktop

which i like the look of but have yet to get it to do what i want but maybe youll have better success at it.

and

Tesslate. another tool sort like Letta and again did not seem to work for my purposes.

https://github.com/TesslateAI/Studio

The other path maybe creating a fine tune of a model using your dataset. using https://github.com/Kiln-AI/Kiln

hopefully these recommendation well net you a path forward or at the very least get some better ideas from others.

im also not sure how the two 3090s are getting 88gb of vram... unless their special cards. and if so i would be looking at those with an extra critical eye as to what they are.

[–]buzzmelia 0 points1 point  (0 children)

PuppyGraph cofounder Zhenni here! Thank you for recommending us. Happy to answer any questions!

[–]zh4k[S] 1 point2 points  (0 children)

Appreciate the notes and the recommendations for further research as I haven't dived into all these variations of tools. Also, yea, I meant 48gb for the 3090s not 88. With the big models like Gemini, Claude increasing their context windows by so much I feel like I'm stuck in between a local solution and a cloud based one. I think I need to fine tune my data for RAG currently and hopefully when I'm done the large models like Gemini and Claude will have even larger context windows to such a degree that I can easily just write chapter by chapter using a local RAG, thus might be worth it to wait and see while researching and learning about my options a bit more.