1234567891011121314151617181920212223242526272829303132333435363738394041424344454647484950515253545556575859606162636465666768 |
- ---
- title: ❓ FAQs
- description: 'Collections of all the frequently asked questions'
- ---
- #### How to use GPT-4 as the LLM model?
- <CodeGroup>
- ```python main.py
- import os
- from embedchain import Pipeline as App
- os.environ['OPENAI_API_KEY'] = 'xxx'
- # load llm configuration from gpt4.yaml file
- app = App.from_config(yaml_path="gpt4.yaml")
- ```
- ```yaml gpt4.yaml
- llm:
- provider: openai
- config:
- model: 'gpt-4'
- temperature: 0.5
- max_tokens: 1000
- top_p: 1
- stream: false
- ```
- </CodeGroup>
- #### I don't have OpenAI credits. How can I use some open source model?
- <CodeGroup>
- ```python main.py
- import os
- from embedchain import Pipeline as App
- os.environ['OPENAI_API_KEY'] = 'xxx'
- # load llm configuration from opensource.yaml file
- app = App.from_config(yaml_path="opensource.yaml")
- ```
- ```yaml opensource.yaml
- llm:
- provider: gpt4all
- config:
- model: 'orca-mini-3b.ggmlv3.q4_0.bin'
- temperature: 0.5
- max_tokens: 1000
- top_p: 1
- stream: false
- embedder:
- provider: gpt4all
- config:
- model: 'all-MiniLM-L6-v2'
- ```
- </CodeGroup>
- #### How to contact support?
- If docs aren't sufficient, please feel free to reach out to us using one of the following methods:
- <Snippet file="get-help.mdx" />
|