|
22 | 22 | from langchain.tools import BaseTool, StructuredTool |
23 | 23 |
|
24 | 24 | from codeinterpreterapi.agents import OpenAIFunctionsAgent |
25 | | -from codeinterpreterapi.chains import (get_file_modifications, |
26 | | - remove_download_link) |
| 25 | +from codeinterpreterapi.chains import get_file_modifications, remove_download_link |
27 | 26 | from codeinterpreterapi.config import settings |
28 | 27 | from codeinterpreterapi.prompts import code_interpreter_system_message |
29 | | -from codeinterpreterapi.schema import (CodeInput, CodeInterpreterResponse, |
30 | | - File, UserRequest) |
31 | | -from codeinterpreterapi.utils import (CodeAgentOutputParser, |
32 | | - CodeCallbackHandler, |
33 | | - CodeChatAgentOutputParser) |
| 28 | +from codeinterpreterapi.schema import ( |
| 29 | + CodeInput, |
| 30 | + CodeInterpreterResponse, |
| 31 | + File, |
| 32 | + UserRequest, |
| 33 | +) |
| 34 | +from codeinterpreterapi.utils import ( |
| 35 | + CodeAgentOutputParser, |
| 36 | + CodeCallbackHandler, |
| 37 | + CodeChatAgentOutputParser, |
| 38 | +) |
34 | 39 |
|
35 | 40 |
|
36 | 41 | class CodeInterpreterSession: |
@@ -86,15 +91,20 @@ def _choose_llm( |
86 | 91 | openai_api_base = getenv("OPENAI_API_BASE") |
87 | 92 | deployment_name = getenv("DEPLOYMENT_NAME") |
88 | 93 | openapi_type = getenv("OPENAI_API_TYPE") |
89 | | - if openapi_type == "azure" and openai_api_version and openai_api_base and deployment_name: |
| 94 | + if ( |
| 95 | + openapi_type == "azure" |
| 96 | + and openai_api_version |
| 97 | + and openai_api_base |
| 98 | + and deployment_name |
| 99 | + ): |
90 | 100 | return AzureChatOpenAI( |
91 | 101 | temperature=0.03, |
92 | 102 | openai_api_base=openai_api_base, |
93 | 103 | openai_api_version=openai_api_version, |
94 | 104 | deployment_name=deployment_name, |
95 | 105 | openai_api_key=openai_api_key, |
96 | 106 | max_retries=3, |
97 | | - request_timeout=60*3 |
| 107 | + request_timeout=60 * 3, |
98 | 108 | ) |
99 | 109 | else: |
100 | 110 | return ChatOpenAI( |
|
0 commit comments