-
Notifications
You must be signed in to change notification settings - Fork 16.4k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Implementation for Matching Engine Vectorstore #3104
Implementation for Matching Engine Vectorstore #3104
Conversation
I just pushed a new updates addressing the comments. However, we were trying to add |
|
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
my main comment, inline with some of the others - is it simpler to just do the client creation OUTSIDE of the class, and then pass in an already initialized client? would cut back on a lot of the ags being passed around
@hwchase17 I thought that was addressed with the |
…ments in from_texts to make them required.
think he means to make |
@dev2049 I already added the |
i just meant you should update |
Great @dev2049 !! So do you need me to do anything else for the merge? |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
looks great! thanks
@hwchase17 any chance to get this into release anytime soon? |
@hwchase17 Same question here: Would be nice to see this released |
One concern is that the docs are stored/retrieve from GCS which is slow (and somewhat defeats the purpose of using a Vector DB) |
@tomaspiaggio should you create a PR your branch to master? |
@hwchase17 Any updates on this one? Would be a cool feature! |
Will this be merged to master? @hwchase17 |
Keen to get this merged into master @hwchase17 |
Once we have Matching engine index is deployed, What is the best retriever on langchain to get the query results ? @tomaspiaggio |
Have been using the Vector Search (Matching Engine) with langchain for a couple of days now and I've been hitting my head against a wall to solve a problem. I notice that when embeddings are sent to Vector Search they get stored and a file is also created and stored within a separate GCS bucket that is referenced when queried. I am looking for a way to remove the embeddings from the Vector Search but it seems I can only do it with gcloud commands but I need to know the datapoint_ids. What would be the best way to store the datapoint_ids that are related to the documents that are being embedded? |
We just finished the implementation for the vector store using the GCP Matching Engine.
We'll be contributing the implementation.
Related to #2892
If you have any questions or suggestions please contact me (@tomaspiaggio) or @scafati98.