| from langchain_openai import ChatOpenAI | |
| import os | |
| try: | |
| from dotenv import load_dotenv | |
| load_dotenv() | |
| except Exception: | |
| pass | |
| # gpt-3.5-turbo-0125 | |
| def get_llm(model="gpt-3.5-turbo", max_tokens=1024, temperature=0.0, | |
| streaming=True, timeout=30, **kwargs): | |
| llm = ChatOpenAI( | |
| model=model, | |
| api_key=os.environ.get("OPENAI_API_KEY", None), | |
| max_tokens=max_tokens, | |
| streaming=streaming, | |
| temperature=temperature, | |
| timeout=timeout, | |
| **kwargs, | |
| ) | |
| return llm | |