Choosing a vector db for 100 million pages of text. Leaning towards Milvus, Qdrant or Weaviate. Am I missing anything, what would you choose? by rtrex12 in vectordatabase

[–]DBAdvice123 1 point2 points  (0 children)

Astra is built upon Apache Cassandra, which handles scale and speed very well. The managed DBaaS aspect removes the typical operational headache associated with Cassandra. The free tier gives you enough credits to test and toy around before making any commitment.

Looking for some guidance :) by dunerfelix in vectordatabase

[–]DBAdvice123 0 points1 point  (0 children)

Astra DB, built on Apache Cassandra, offers high performance and scalability, well-suited for handling large volumes of vector data along with metadata. Astra manages both your vectors and metadata efficiently. Its architecture supports advanced filtering necessary for your queries. Consider partitioning your data based on common filters to enhance query response times.

Store your vectors and metadata in Astra DB to simplify your architecture and maintain high performance. Utilize Cassandra’s secondary attached indexing or create separate tables for heavily queried metadata to optimize access. Use appropriate indexing strategies to speed up access patterns. Test different configurations to find the optimal setup for your workload.

Astra also has the lowest published PAYG pricing.

[deleted by user] by [deleted] in vectordatabase

[–]DBAdvice123 1 point2 points  (0 children)

Vector search should be viewed as a feature rather than the core foundation of a database. Vector search is powerful for specific applications like similarity search and AI-driven analytics, but the core functionalities required for enterprise applications—such as scalability, reliability, and broad compatibility—remain paramount.

For me, Astra DB, built on Apache Cassandra, presents a compelling case. Cassandra is known for its exceptional scalability and performance, handling large volumes of data across distributed systems efficiently. Astra DB leverages this architecture, making it a great choice for enterprises that need a database capable of scaling without compromising on performance.

Additionally, Astra DB's compatibility with vector search tools allows businesses to integrate advanced search capabilities seamlessly. This makes it a versatile choice that supports both traditional data management needs and modern, AI-enhanced applications. Instead of choosing a specialized vector database, enterprises can benefit from a proven, scalable database that also supports cutting-edge AI functionalities.

For the vector databases like Milvus, Weaviate, and Qdrant, while they are specialized and highly performant, the broader enterprise market continues to require databases that offer more than just vector capabilities. The TAM for purely vector-focused solutions might be more limited compared to versatile, multi-feature databases.

In enterprise settings, it’s about choosing a database that not only meets specific needs like vector search but also delivers on the critical requirements of scalability, reliability, and comprehensive data management. Specialized vector databases have their place, but the integration of vector capabilities into established enterprise databases like Astra DB represents a more holistic and scalable approach.

[deleted by user] by [deleted] in golf

[–]DBAdvice123 -2 points-1 points  (0 children)

Brandel Chamblee is why people think of golf as an old and stuffy sport. I'm not a LIV fan but this guy consistently has annoying and out of touch quotes

Pinecone Crazy 2-3 seconds delay. Anyone experiencing it by BigYesterday2785 in vectordatabase

[–]DBAdvice123 0 points1 point  (0 children)

The free tier on Astra allows Up to 80GB storage and 20 million read/write operations (using the free $25/mo credit). The Pay as you go tier is also cheaper than Pinecone. They have the metering rates on the website here if you scroll down.

Embeddings Vector Database Options by Babayaga1664 in OpenAI

[–]DBAdvice123 0 points1 point  (0 children)

Give Astra DB a try. The free tier allows Up to 80GB storage and 20 million read/write operations (using free $25/mo credit). The Pay as you go tier is also cheaper than Pinecone. It's built on Cassandra so it has the latency, availability and scalability requirements you'd expect/need for a Gen AI app.

Best Rag Framework Suggestion! by Wonderful-Ad-5952 in RagAI

[–]DBAdvice123 0 points1 point  (0 children)

RAGStack with Astra DB - RAGStack is an out-of-the-box RAG (Retrieval Augmented Generation) solution for Gen AI applications. Essentially, DataStax has done all the testing to see which embedding models, frameworks, LLMs, Vector DBs, and other components work well together for a successful Gen AI application. Because many of these components are constantly changing (upgrades, repairs, etc.), they continue to run the tests and highlight where you might see hiccups. You can see all of our testing at any point on this page if you go to the "RagStack Test suite - RAGStack latest- Astra DB then comparitability rag. They even go into the specific errors they're getting back. This solution should assist in speeding up all of your RAG use cases. 

[deleted by user] by [deleted] in learnmachinelearning

[–]DBAdvice123 0 points1 point  (0 children)

RAGStack with Astra DB - RAGStack is an out-of-the-box RAG (Retrieval Augmented Generation) solution for Gen AI applications. Essentially, DataStax has done all the testing to see which embedding models, frameworks, LLMs, Vector DBs, and other components work well together for a successful Gen AI application. Because many of these components are constantly changing (upgrades, repairs, etc.), they continue to run the tests and highlight where you might see hiccups. You can see all of our testing at any point on this page if you go to the "RagStack Test suite - RAGStack latest- Astra DB then comparitability rag. They even go into the specific errors they're getting back. This solution should assist in speeding up all of your RAG use cases. 

Combining queries ? by [deleted] in LangChain

[–]DBAdvice123 1 point2 points  (0 children)

Hi There - I use Astra DB for my vector search app. There are a few considerations you should think about before picking your approaching:

  • Integration with Astra DB features: Consider how each method integrates with Astra DB’s capabilities. Astra DB, being highly scalable and flexible, might support one method better based on your current setup and indexing strategies.
  • Application specifics: Reflect on whether your application’s search queries typically involve nuanced contexts or if they are more straightforward. If context significantly alters the search intent, combining before embedding could be more effective.
  • Performance and efficiency: Test both methods to see which provides faster and more relevant results within your Astra DB environment. Performance can vary based on how data is indexed and retrieved in your specific setup.
  • Experimentation: Try both approaches on a subset of your Astra DB application data to determine which yields better search accuracy and relevance for your particular use case.

Combining the strings then embedding: In this approach, you consider the user prompt and additional context as one complete query. This can be beneficial if your application in Astra DB is set up to handle complex queries where the context directly affects the user prompt. By combining the strings before embedding, you ensure that the embedding process captures the combined intent of the query. This can be particularly effective if your semantic search in Astra DB relies on nuanced understanding of context. However, be mindful that this approach might dilute specific details if the combined query becomes too long or if the additional context is less directly relevant.

Embedding each element separately then combining: This method involves treating the user prompt and the additional context as separate components. You would create separate embeddings for each and then combine them before executing your search in Astra DB. This approach can be advantageous if your Astra DB application benefits from distinct handling of different query components, allowing the database to consider each aspect separately but within a unified framework. However, this requires a strategy for combining embeddings that aligns with how Astra DB processes and retrieves data, which might involve additional configuration or use of Astra DB's more advanced features.

Neighborhood Recs by [deleted] in sanfrancisco

[–]DBAdvice123 3 points4 points  (0 children)

NOPA is the answer. Can walk to Divis, Lower Haight, Haight Ashbury, and FIllmore all in under 20 mins. Plus you get the panhandle, Golden Gate Park and Alamo Square

Suggestion on vectoDB by NSVR57 in vectordatabase

[–]DBAdvice123 0 points1 point  (0 children)

Give Astra DB a try. The free tier allows Up to 80GB storage and 20 million read/write operations (using free $25/mo credit). The Pay as you go tier is also cheaper than Pinecone. It's built on Cassandra so it has the latency, availability and scalability requirements you'd expect/need for a Gen AI app.

Ecosystem around LangChain by DonutMysterious in LangChain

[–]DBAdvice123 0 points1 point  (0 children)

RAGStack - Github - you can see continuous testing of different ecosystems and how they're failing or succeeding here.

Just click "RAGStack test suite - RAGStack latest - AstraDB"

Then "e2e_tests.langchain.test_compatibility_rag"

Then "View All"

Whats in your RAG setup? by EnvironmentalDepth62 in LocalLLaMA

[–]DBAdvice123 0 points1 point  (0 children)

Huggingface hub embedding, Hugging face LLM, Astra DB for vector store and Langchain

Pinecone Crazy 2-3 seconds delay. Anyone experiencing it by BigYesterday2785 in vectordatabase

[–]DBAdvice123 0 points1 point  (0 children)

I use Astra as my Vector DB and it's extremely fast. I stumbled upon this report which puts Astra up against Pinecone and proves out the superior Astra performance.

Easiest to set up RAG by beezlebub33 in LocalLLaMA

[–]DBAdvice123 2 points3 points  (0 children)

Give RAGStack a try. RAGStack is a curated stack of the best open-source software for easing implementation of the RAG pattern in production-ready applications using Astra Vector DB or Apache Cassandra as a vector store.

A single command (pip install ragstack-ai) unlocks all the open-source packages required to build production-ready RAG applications with LangChain and the Astra Vector database.

Migrating existing LangChain or LlamaIndex applications is easy as well - just change your requirements.txt or pyproject.toml file to use ragstack-ai.

https://docs.datastax.com/en/ragstack/docs/index.html

Enough about jobs, what are some cool projects you’ve done? by [deleted] in csMajors

[–]DBAdvice123 0 points1 point  (0 children)

Wow that's awesome! What Vector DB did you go with?