Skip to content
Projects
Groups
Snippets
Help
Loading...
Help
Submit feedback
Contribute to GitLab
Sign in
Toggle navigation
D
dify
Project
Project
Details
Activity
Releases
Cycle Analytics
Repository
Repository
Files
Commits
Branches
Tags
Contributors
Graph
Compare
Charts
Issues
0
Issues
0
List
Board
Labels
Milestones
Merge Requests
0
Merge Requests
0
CI / CD
CI / CD
Pipelines
Jobs
Schedules
Charts
Wiki
Wiki
Snippets
Snippets
Members
Members
Collapse sidebar
Close sidebar
Activity
Graph
Charts
Create a new issue
Jobs
Commits
Issue Boards
Open sidebar
ai-tech
dify
Commits
c2f1be9d
Commit
c2f1be9d
authored
Jun 19, 2023
by
John Wang
Browse files
Options
Browse Files
Download
Email Patches
Plain Diff
fix: model kwargs
parent
bcaf2274
Changes
1
Show whitespace changes
Inline
Side-by-side
Showing
1 changed file
with
11 additions
and
4 deletions
+11
-4
llm_builder.py
api/core/llm/llm_builder.py
+11
-4
No files found.
api/core/llm/llm_builder.py
View file @
c2f1be9d
...
@@ -34,6 +34,8 @@ class LLMBuilder:
...
@@ -34,6 +34,8 @@ class LLMBuilder:
def
to_llm
(
cls
,
tenant_id
:
str
,
model_name
:
str
,
**
kwargs
)
->
Union
[
StreamableOpenAI
,
StreamableChatOpenAI
]:
def
to_llm
(
cls
,
tenant_id
:
str
,
model_name
:
str
,
**
kwargs
)
->
Union
[
StreamableOpenAI
,
StreamableChatOpenAI
]:
provider
=
cls
.
get_default_provider
(
tenant_id
)
provider
=
cls
.
get_default_provider
(
tenant_id
)
model_credentials
=
cls
.
get_model_credentials
(
tenant_id
,
provider
,
model_name
)
mode
=
cls
.
get_mode_by_model
(
model_name
)
mode
=
cls
.
get_mode_by_model
(
model_name
)
if
mode
==
'chat'
:
if
mode
==
'chat'
:
if
provider
==
'openai'
:
if
provider
==
'openai'
:
...
@@ -48,15 +50,20 @@ class LLMBuilder:
...
@@ -48,15 +50,20 @@ class LLMBuilder:
else
:
else
:
raise
ValueError
(
f
"model name {model_name} is not supported."
)
raise
ValueError
(
f
"model name {model_name} is not supported."
)
model_credentials
=
cls
.
get_model_credentials
(
tenant_id
,
provider
,
model_name
)
model_kwargs
=
{
'top_p'
:
kwargs
.
get
(
'top_p'
,
1
),
'frequency_penalty'
:
kwargs
.
get
(
'frequency_penalty'
,
0
),
'presence_penalty'
:
kwargs
.
get
(
'presence_penalty'
,
0
),
}
model_extras_kwargs
=
model_kwargs
if
mode
==
'completion'
else
{
'model_kwargs'
:
model_kwargs
}
return
llm_cls
(
return
llm_cls
(
model_name
=
model_name
,
model_name
=
model_name
,
temperature
=
kwargs
.
get
(
'temperature'
,
0
),
temperature
=
kwargs
.
get
(
'temperature'
,
0
),
max_tokens
=
kwargs
.
get
(
'max_tokens'
,
256
),
max_tokens
=
kwargs
.
get
(
'max_tokens'
,
256
),
top_p
=
kwargs
.
get
(
'top_p'
,
1
),
**
model_extras_kwargs
,
frequency_penalty
=
kwargs
.
get
(
'frequency_penalty'
,
0
),
presence_penalty
=
kwargs
.
get
(
'presence_penalty'
,
0
),
callbacks
=
kwargs
.
get
(
'callbacks'
,
None
),
callbacks
=
kwargs
.
get
(
'callbacks'
,
None
),
streaming
=
kwargs
.
get
(
'streaming'
,
False
),
streaming
=
kwargs
.
get
(
'streaming'
,
False
),
# request_timeout=None
# request_timeout=None
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment