Skip to content
Projects
Groups
Snippets
Help
Loading...
Help
Submit feedback
Contribute to GitLab
Sign in
Toggle navigation
D
dify
Project
Project
Details
Activity
Releases
Cycle Analytics
Repository
Repository
Files
Commits
Branches
Tags
Contributors
Graph
Compare
Charts
Issues
0
Issues
0
List
Board
Labels
Milestones
Merge Requests
0
Merge Requests
0
CI / CD
CI / CD
Pipelines
Jobs
Schedules
Charts
Wiki
Wiki
Snippets
Snippets
Members
Members
Collapse sidebar
Close sidebar
Activity
Graph
Charts
Create a new issue
Jobs
Commits
Issue Boards
Open sidebar
ai-tech
dify
Commits
77f9e8ce
Unverified
Commit
77f9e8ce
authored
Jan 03, 2024
by
Chenhe Gu
Committed by
GitHub
Jan 03, 2024
Browse files
Options
Browse Files
Download
Email Patches
Plain Diff
add example api url endpoint in placeholder (#1887)
Co-authored-by:
takatost
<
takatost@gmail.com
>
parent
5ca4c4a4
Changes
3
Show whitespace changes
Inline
Side-by-side
Showing
3 changed files
with
121 additions
and
59 deletions
+121
-59
llm.py
..._runtime/model_providers/openai_api_compatible/llm/llm.py
+96
-52
openai_api_compatible.yaml
...roviders/openai_api_compatible/openai_api_compatible.yaml
+2
-2
text_embedding.py
...rs/openai_api_compatible/text_embedding/text_embedding.py
+23
-5
No files found.
api/core/model_runtime/model_providers/openai_api_compatible/llm/llm.py
View file @
77f9e8ce
import
logging
from
decimal
import
Decimal
from
urllib.parse
import
urljoin
import
requests
import
json
...
...
@@ -9,9 +10,12 @@ from typing import Optional, Generator, Union, List, cast
from
core.model_runtime.entities.common_entities
import
I18nObject
from
core.model_runtime.utils
import
helper
from
core.model_runtime.entities.message_entities
import
ImagePromptMessageContent
,
PromptMessage
,
AssistantPromptMessage
,
PromptMessageContent
,
\
PromptMessageContentType
,
PromptMessageFunction
,
PromptMessageTool
,
UserPromptMessage
,
SystemPromptMessage
,
ToolPromptMessage
from
core.model_runtime.entities.model_entities
import
ModelPropertyKey
,
ModelType
,
PriceConfig
,
ParameterRule
,
DefaultParameterName
,
\
from
core.model_runtime.entities.message_entities
import
ImagePromptMessageContent
,
PromptMessage
,
\
AssistantPromptMessage
,
PromptMessageContent
,
\
PromptMessageContentType
,
PromptMessageFunction
,
PromptMessageTool
,
UserPromptMessage
,
SystemPromptMessage
,
\
ToolPromptMessage
from
core.model_runtime.entities.model_entities
import
ModelPropertyKey
,
ModelType
,
PriceConfig
,
ParameterRule
,
\
DefaultParameterName
,
\
ParameterType
,
ModelPropertyKey
,
FetchFrom
,
AIModelEntity
from
core.model_runtime.entities.llm_entities
import
LLMMode
,
LLMResult
,
LLMResultChunk
,
LLMResultChunkDelta
from
core.model_runtime.errors.invoke
import
InvokeError
...
...
@@ -89,6 +93,8 @@ class OAIAPICompatLargeLanguageModel(_CommonOAI_API_Compat, LargeLanguageModel):
headers
[
"Authorization"
]
=
f
"Bearer {api_key}"
endpoint_url
=
credentials
[
'endpoint_url'
]
if
not
endpoint_url
.
endswith
(
'/'
):
endpoint_url
+=
'/'
# prepare the payload for a simple ping to the model
data
=
{
...
...
@@ -105,8 +111,10 @@ class OAIAPICompatLargeLanguageModel(_CommonOAI_API_Compat, LargeLanguageModel):
"content"
:
"ping"
},
]
endpoint_url
=
urljoin
(
endpoint_url
,
'chat/completions'
)
elif
completion_type
is
LLMMode
.
COMPLETION
:
data
[
'prompt'
]
=
'ping'
endpoint_url
=
urljoin
(
endpoint_url
,
'completions'
)
else
:
raise
ValueError
(
"Unsupported completion type for model configuration."
)
...
...
@@ -119,8 +127,24 @@ class OAIAPICompatLargeLanguageModel(_CommonOAI_API_Compat, LargeLanguageModel):
)
if
response
.
status_code
!=
200
:
raise
CredentialsValidateFailedError
(
f
'Credentials validation failed with status code {response.status_code}: {response.text}'
)
raise
CredentialsValidateFailedError
(
f
'Credentials validation failed with status code {response.status_code}'
)
try
:
json_result
=
response
.
json
()
except
json
.
JSONDecodeError
as
e
:
raise
CredentialsValidateFailedError
(
f
'Credentials validation failed: JSON decode error'
)
if
(
completion_type
is
LLMMode
.
CHAT
and
(
'object'
not
in
json_result
or
json_result
[
'object'
]
!=
'chat.completion'
)):
raise
CredentialsValidateFailedError
(
f
'Credentials validation failed: invalid response object, must be
\'
chat.completion
\'
'
)
elif
(
completion_type
is
LLMMode
.
COMPLETION
and
(
'object'
not
in
json_result
or
json_result
[
'object'
]
!=
'text_completion'
)):
raise
CredentialsValidateFailedError
(
f
'Credentials validation failed: invalid response object, must be
\'
text_completion
\'
'
)
except
CredentialsValidateFailedError
:
raise
except
Exception
as
ex
:
raise
CredentialsValidateFailedError
(
f
'An error occurred during credentials validation: {str(ex)}'
)
...
...
@@ -134,8 +158,8 @@ class OAIAPICompatLargeLanguageModel(_CommonOAI_API_Compat, LargeLanguageModel):
model_type
=
ModelType
.
LLM
,
fetch_from
=
FetchFrom
.
CUSTOMIZABLE_MODEL
,
model_properties
=
{
ModelPropertyKey
.
CONTEXT_SIZE
:
credentials
.
get
(
'context_size'
),
ModelPropertyKey
.
MODE
:
'chat'
ModelPropertyKey
.
CONTEXT_SIZE
:
int
(
credentials
.
get
(
'context_size'
)
),
ModelPropertyKey
.
MODE
:
credentials
.
get
(
'mode'
),
},
parameter_rules
=
[
ParameterRule
(
...
...
@@ -197,10 +221,10 @@ class OAIAPICompatLargeLanguageModel(_CommonOAI_API_Compat, LargeLanguageModel):
return
entity
# validate_credentials method has been rewritten to use the requests library for compatibility with all providers following OpenAI's API standard.
def
_generate
(
self
,
model
:
str
,
credentials
:
dict
,
prompt_messages
:
list
[
PromptMessage
],
model_parameters
:
dict
,
tools
:
Optional
[
list
[
PromptMessageTool
]]
=
None
,
stop
:
Optional
[
List
[
str
]]
=
None
,
stream
:
bool
=
True
,
\
tools
:
Optional
[
list
[
PromptMessageTool
]]
=
None
,
stop
:
Optional
[
List
[
str
]]
=
None
,
stream
:
bool
=
True
,
\
user
:
Optional
[
str
]
=
None
)
->
Union
[
LLMResult
,
Generator
]:
"""
Invoke llm completion model
...
...
@@ -223,6 +247,8 @@ class OAIAPICompatLargeLanguageModel(_CommonOAI_API_Compat, LargeLanguageModel):
headers
[
"Authorization"
]
=
f
"Bearer {api_key}"
endpoint_url
=
credentials
[
"endpoint_url"
]
if
not
endpoint_url
.
endswith
(
'/'
):
endpoint_url
+=
'/'
data
=
{
"model"
:
model
,
...
...
@@ -233,8 +259,10 @@ class OAIAPICompatLargeLanguageModel(_CommonOAI_API_Compat, LargeLanguageModel):
completion_type
=
LLMMode
.
value_of
(
credentials
[
'mode'
])
if
completion_type
is
LLMMode
.
CHAT
:
endpoint_url
=
urljoin
(
endpoint_url
,
'chat/completions'
)
data
[
'messages'
]
=
[
self
.
_convert_prompt_message_to_dict
(
m
)
for
m
in
prompt_messages
]
elif
completion_type
==
LLMMode
.
COMPLETION
:
endpoint_url
=
urljoin
(
endpoint_url
,
'completions'
)
data
[
'prompt'
]
=
prompt_messages
[
0
]
.
content
else
:
raise
ValueError
(
"Unsupported completion type for model configuration."
)
...
...
@@ -245,7 +273,7 @@ class OAIAPICompatLargeLanguageModel(_CommonOAI_API_Compat, LargeLanguageModel):
data
[
"tool_choice"
]
=
"auto"
for
tool
in
tools
:
formatted_tools
.
append
(
helper
.
dump_model
(
PromptMessageFunction
(
function
=
tool
)))
formatted_tools
.
append
(
helper
.
dump_model
(
PromptMessageFunction
(
function
=
tool
)))
data
[
"tools"
]
=
formatted_tools
...
...
@@ -313,6 +341,7 @@ class OAIAPICompatLargeLanguageModel(_CommonOAI_API_Compat, LargeLanguageModel):
if
chunk
:
decoded_chunk
=
chunk
.
decode
(
'utf-8'
)
.
strip
()
.
lstrip
(
'data: '
)
.
lstrip
()
chunk_json
=
None
try
:
chunk_json
=
json
.
loads
(
decoded_chunk
)
# stream ended
...
...
@@ -323,13 +352,15 @@ class OAIAPICompatLargeLanguageModel(_CommonOAI_API_Compat, LargeLanguageModel):
finish_reason
=
"Non-JSON encountered."
)
if
len
(
chunk_json
[
'choices'
])
==
0
:
if
not
chunk_json
or
len
(
chunk_json
[
'choices'
])
==
0
:
continue
delta
=
chunk_json
[
'choices'
][
0
][
'delta'
]
chunk_index
=
ch
unk_json
[
'choices'
][
0
][
'index'
]
choice
=
chunk_json
[
'choices'
][
0
]
chunk_index
=
ch
oice
[
'index'
]
if
'index'
in
choice
else
chunk_index
if
delta
.
get
(
'finish_reason'
)
is
None
and
(
delta
.
get
(
'content'
)
is
None
or
delta
.
get
(
'content'
)
==
''
):
if
'delta'
in
choice
:
delta
=
choice
[
'delta'
]
if
delta
.
get
(
'content'
)
is
None
or
delta
.
get
(
'content'
)
==
''
:
continue
assistant_message_tool_calls
=
delta
.
get
(
'tool_calls'
,
None
)
...
...
@@ -348,16 +379,26 @@ class OAIAPICompatLargeLanguageModel(_CommonOAI_API_Compat, LargeLanguageModel):
)
full_assistant_content
+=
delta
.
get
(
'content'
,
''
)
elif
'text'
in
choice
:
if
choice
.
get
(
'text'
)
is
None
or
choice
.
get
(
'text'
)
==
''
:
continue
# transform assistant message to prompt message
assistant_prompt_message
=
AssistantPromptMessage
(
content
=
choice
.
get
(
'text'
,
''
)
)
full_assistant_content
+=
choice
.
get
(
'text'
,
''
)
else
:
continue
# check payload indicator for completion
if
chunk_json
[
'choices'
][
0
]
.
get
(
'finish_reason'
)
is
not
None
:
yield
create_final_llm_result_chunk
(
index
=
chunk_index
,
message
=
assistant_prompt_message
,
finish_reason
=
chunk_json
[
'choices'
][
0
][
'finish_reason'
]
)
else
:
yield
LLMResultChunk
(
model
=
model
,
...
...
@@ -374,6 +415,8 @@ class OAIAPICompatLargeLanguageModel(_CommonOAI_API_Compat, LargeLanguageModel):
finish_reason
=
"End of stream."
)
chunk_index
+=
1
def
_handle_generate_response
(
self
,
model
:
str
,
credentials
:
dict
,
response
:
requests
.
Response
,
prompt_messages
:
list
[
PromptMessage
])
->
LLMResult
:
...
...
@@ -455,7 +498,8 @@ class OAIAPICompatLargeLanguageModel(_CommonOAI_API_Compat, LargeLanguageModel):
message
=
cast
(
AssistantPromptMessage
,
message
)
message_dict
=
{
"role"
:
"assistant"
,
"content"
:
message
.
content
}
if
message
.
tool_calls
:
message_dict
[
"tool_calls"
]
=
[
helper
.
dump_model
(
PromptMessageFunction
(
function
=
tool_call
))
for
tool_call
in
message_dict
[
"tool_calls"
]
=
[
helper
.
dump_model
(
PromptMessageFunction
(
function
=
tool_call
))
for
tool_call
in
message
.
tool_calls
]
# function_call = message.tool_calls[0]
# message_dict["function_call"] = {
...
...
api/core/model_runtime/model_providers/openai_api_compatible/openai_api_compatible.yaml
View file @
77f9e8ce
...
...
@@ -33,8 +33,8 @@ model_credential_schema:
type
:
text-input
required
:
true
placeholder
:
zh_Hans
:
在此输入您的 API endpoint URL
en_US
:
Enter your API endpoint URL
zh_Hans
:
Base URL, eg. https://api.openai.com/v1
en_US
:
Base URL, eg. https://api.openai.com/v1
-
variable
:
mode
show_on
:
-
variable
:
__model_type
...
...
api/core/model_runtime/model_providers/openai_api_compatible/text_embedding/text_embedding.py
View file @
77f9e8ce
import
time
from
decimal
import
Decimal
from
typing
import
Optional
from
urllib.parse
import
urljoin
import
requests
import
json
...
...
@@ -42,8 +43,11 @@ class OAICompatEmbeddingModel(_CommonOAI_API_Compat, TextEmbeddingModel):
if
api_key
:
headers
[
"Authorization"
]
=
f
"Bearer {api_key}"
endpoint_url
=
credentials
.
get
(
'endpoint_url'
)
if
not
endpoint_url
.
endswith
(
'/'
):
endpoint_url
+=
'/'
endpoint_url
=
credentials
[
'endpoint_url'
]
endpoint_url
=
urljoin
(
endpoint_url
,
'embeddings'
)
extra_model_kwargs
=
{}
if
user
:
...
...
@@ -144,8 +148,11 @@ class OAICompatEmbeddingModel(_CommonOAI_API_Compat, TextEmbeddingModel):
if
api_key
:
headers
[
"Authorization"
]
=
f
"Bearer {api_key}"
endpoint_url
=
credentials
.
get
(
'endpoint_url'
)
if
not
endpoint_url
.
endswith
(
'/'
):
endpoint_url
+=
'/'
endpoint_url
=
credentials
[
'endpoint_url'
]
endpoint_url
=
urljoin
(
endpoint_url
,
'embeddings'
)
payload
=
{
'input'
:
'ping'
,
...
...
@@ -160,8 +167,19 @@ class OAICompatEmbeddingModel(_CommonOAI_API_Compat, TextEmbeddingModel):
)
if
response
.
status_code
!=
200
:
raise
CredentialsValidateFailedError
(
f
"Invalid response status: {response.status_code}"
)
raise
CredentialsValidateFailedError
(
f
'Credentials validation failed with status code {response.status_code}'
)
try
:
json_result
=
response
.
json
()
except
json
.
JSONDecodeError
as
e
:
raise
CredentialsValidateFailedError
(
f
'Credentials validation failed: JSON decode error'
)
if
'model'
not
in
json_result
:
raise
CredentialsValidateFailedError
(
f
'Credentials validation failed: invalid response'
)
except
CredentialsValidateFailedError
:
raise
except
Exception
as
ex
:
raise
CredentialsValidateFailedError
(
str
(
ex
))
...
...
@@ -175,7 +193,7 @@ class OAICompatEmbeddingModel(_CommonOAI_API_Compat, TextEmbeddingModel):
model_type
=
ModelType
.
TEXT_EMBEDDING
,
fetch_from
=
FetchFrom
.
CUSTOMIZABLE_MODEL
,
model_properties
=
{
ModelPropertyKey
.
CONTEXT_SIZE
:
credentials
.
get
(
'context_size'
),
ModelPropertyKey
.
CONTEXT_SIZE
:
int
(
credentials
.
get
(
'context_size'
)
),
ModelPropertyKey
.
MAX_CHUNKS
:
1
,
},
parameter_rules
=
[],
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment