Skip to content
This repository has been archived by the owner on Nov 13, 2024. It is now read-only.

Add Anyscale Embedding model support #198

Merged
merged 5 commits into from
Dec 13, 2023

Conversation

kylehh
Copy link
Contributor

@kylehh kylehh commented Dec 5, 2023

Problem

No support for Open Source embedding models

Solution

Added Anyscale Embedding model support for

Type of Change

  • Bug fix (non-breaking change which fixes an issue)
  • [ x] New feature (non-breaking change which adds functionality)
  • Breaking change (fix or feature that would cause existing functionality to not work as expected)
  • This change requires a documentation update
  • Infrastructure change (CI configs, etc)
  • Non-code change (docs, etc)
  • None of the above: (explain here)

Test Plan

Describe specific steps for validating this change.
unit test tests/system/record_encoder/test_anyscale_record_encoder.py
canopy test config/anyscale.yaml

Copy link
Contributor

@igiloh-pinecone igiloh-pinecone left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

@kylehh LGTM!

Please see one clarification question


chat_engine:
llm: &llm
type: AnyscaleLLM # Options: [OpenAILLM, AnyscaleLLM]
params:
model_name: meta-llama/Llama-2-7b-chat-hf # The name of the model to use.
model_name: HuggingFaceH4/zephyr-7b-beta # The name of the model to use.
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

@kylehh why do you want to change the llama2 version?

@igiloh-pinecone igiloh-pinecone added this pull request to the merge queue Dec 10, 2023
@igiloh-pinecone igiloh-pinecone removed this pull request from the merge queue due to a manual request Dec 10, 2023
We have a hacky workaround in CLI to select the batch_size of documents by the expected batch_size of underlying chunks to encode.
However this mechanism used a hardcoded number, instead of using the actual Encoder's batch_size
Copy link
Contributor

@igiloh-pinecone igiloh-pinecone left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

LGTM

@igiloh-pinecone igiloh-pinecone added this pull request to the merge queue Dec 12, 2023
@github-merge-queue github-merge-queue bot removed this pull request from the merge queue due to failed status checks Dec 12, 2023
@igiloh-pinecone igiloh-pinecone added this pull request to the merge queue Dec 13, 2023
Merged via the queue into pinecone-io:main with commit 7844f31 Dec 13, 2023
10 checks passed
@kylehh kylehh deleted the ae-embedding branch December 13, 2023 15:24
@grpinto
Copy link

grpinto commented Jan 21, 2024

I don't understand how is this going to change the initial variable definition, how should it be provided, I already provided anyscale api key and still cannot use the command :

canonopy new test

This is the error that I get :

(canopy-env) kingsize@Goncalos-MacBook-Pro-3 EdGenAI % canopy new test
Canopy is going to create a new index named canopy--test with the following initialization parameters:
{}

Do you want to continue? [y/N]: y
Error: Failed to create a new index. Reason:
Canopy has failed to infer vectors' dimensionality using the selected encoder: OpenAIRecordEncoder. You can provide the dimension manually, try using a different encoder, or fix the underlying error:
Failed to enconde documents using OpenAIRecordEncoder. Error: Your OpenAI account seem to have reached the rate limit. Details: You exceeded your current quota, please check your plan and billing details. For more information on this error, read the docs: https://platform.openai.com/docs/guides/error-codes/api-errors.

Sign up for free to subscribe to this conversation on GitHub. Already have an account? Sign in.
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

3 participants