gpt4all.py 855 B

123456789101112131415161718192021
  1. from typing import Optional
  2. from chromadb.utils import embedding_functions
  3. from embedchain.config import BaseEmbedderConfig
  4. from embedchain.embedder.base import BaseEmbedder
  5. from embedchain.models import VectorDimensions
  6. class GPT4AllEmbedder(BaseEmbedder):
  7. def __init__(self, config: Optional[BaseEmbedderConfig] = None):
  8. # Note: We could use langchains GPT4ALL embedding, but it's not available in all versions.
  9. super().__init__(config=config)
  10. if self.config.model is None:
  11. self.config.model = "all-MiniLM-L6-v2"
  12. embedding_fn = embedding_functions.SentenceTransformerEmbeddingFunction(model_name=self.config.model)
  13. self.set_embedding_fn(embedding_fn=embedding_fn)
  14. vector_dimension = VectorDimensions.GPT4ALL.value
  15. self.set_vector_dimension(vector_dimension=vector_dimension)