Unverified Commit 33901384 authored by takatost's avatar takatost Committed by GitHub

fix: httpx socks package missing (#1977)

parent 7a221d08
...@@ -260,7 +260,7 @@ provider_credential_schema: ...@@ -260,7 +260,7 @@ provider_credential_schema:
fetch_from=FetchFrom.CUSTOMIZABLE_MODEL, fetch_from=FetchFrom.CUSTOMIZABLE_MODEL,
model_type=model_type, model_type=model_type,
model_properties={ model_properties={
'mode': ModelType.LLM, ModelPropertyKey.MODE: ModelType.LLM,
}, },
parameter_rules=rules parameter_rules=rules
) )
......
...@@ -40,8 +40,8 @@ LLM_BASE_MODELS = [ ...@@ -40,8 +40,8 @@ LLM_BASE_MODELS = [
], ],
fetch_from=FetchFrom.CUSTOMIZABLE_MODEL, fetch_from=FetchFrom.CUSTOMIZABLE_MODEL,
model_properties={ model_properties={
'mode': LLMMode.CHAT.value, ModelPropertyKey.MODE: LLMMode.CHAT.value,
'context_size': 4096, ModelPropertyKey.CONTEXT_SIZE: 4096,
}, },
parameter_rules=[ parameter_rules=[
ParameterRule( ParameterRule(
...@@ -84,8 +84,8 @@ LLM_BASE_MODELS = [ ...@@ -84,8 +84,8 @@ LLM_BASE_MODELS = [
], ],
fetch_from=FetchFrom.CUSTOMIZABLE_MODEL, fetch_from=FetchFrom.CUSTOMIZABLE_MODEL,
model_properties={ model_properties={
'mode': LLMMode.CHAT.value, ModelPropertyKey.MODE: LLMMode.CHAT.value,
'context_size': 16385, ModelPropertyKey.CONTEXT_SIZE: 16385,
}, },
parameter_rules=[ parameter_rules=[
ParameterRule( ParameterRule(
...@@ -128,8 +128,8 @@ LLM_BASE_MODELS = [ ...@@ -128,8 +128,8 @@ LLM_BASE_MODELS = [
], ],
fetch_from=FetchFrom.CUSTOMIZABLE_MODEL, fetch_from=FetchFrom.CUSTOMIZABLE_MODEL,
model_properties={ model_properties={
'mode': LLMMode.CHAT.value, ModelPropertyKey.MODE: LLMMode.CHAT.value,
'context_size': 8192, ModelPropertyKey.CONTEXT_SIZE: 8192,
}, },
parameter_rules=[ parameter_rules=[
ParameterRule( ParameterRule(
...@@ -202,8 +202,8 @@ LLM_BASE_MODELS = [ ...@@ -202,8 +202,8 @@ LLM_BASE_MODELS = [
], ],
fetch_from=FetchFrom.CUSTOMIZABLE_MODEL, fetch_from=FetchFrom.CUSTOMIZABLE_MODEL,
model_properties={ model_properties={
'mode': LLMMode.CHAT.value, ModelPropertyKey.MODE: LLMMode.CHAT.value,
'context_size': 32768, ModelPropertyKey.CONTEXT_SIZE: 32768,
}, },
parameter_rules=[ parameter_rules=[
ParameterRule( ParameterRule(
...@@ -276,8 +276,8 @@ LLM_BASE_MODELS = [ ...@@ -276,8 +276,8 @@ LLM_BASE_MODELS = [
], ],
fetch_from=FetchFrom.CUSTOMIZABLE_MODEL, fetch_from=FetchFrom.CUSTOMIZABLE_MODEL,
model_properties={ model_properties={
'mode': LLMMode.CHAT.value, ModelPropertyKey.MODE: LLMMode.CHAT.value,
'context_size': 128000, ModelPropertyKey.CONTEXT_SIZE: 128000,
}, },
parameter_rules=[ parameter_rules=[
ParameterRule( ParameterRule(
...@@ -349,8 +349,8 @@ LLM_BASE_MODELS = [ ...@@ -349,8 +349,8 @@ LLM_BASE_MODELS = [
], ],
fetch_from=FetchFrom.CUSTOMIZABLE_MODEL, fetch_from=FetchFrom.CUSTOMIZABLE_MODEL,
model_properties={ model_properties={
'mode': LLMMode.CHAT.value, ModelPropertyKey.MODE: LLMMode.CHAT.value,
'context_size': 128000, ModelPropertyKey.CONTEXT_SIZE: 128000,
}, },
parameter_rules=[ parameter_rules=[
ParameterRule( ParameterRule(
...@@ -419,8 +419,8 @@ LLM_BASE_MODELS = [ ...@@ -419,8 +419,8 @@ LLM_BASE_MODELS = [
model_type=ModelType.LLM, model_type=ModelType.LLM,
fetch_from=FetchFrom.CUSTOMIZABLE_MODEL, fetch_from=FetchFrom.CUSTOMIZABLE_MODEL,
model_properties={ model_properties={
'mode': LLMMode.COMPLETION.value, ModelPropertyKey.MODE: LLMMode.COMPLETION.value,
'context_size': 4096, ModelPropertyKey.CONTEXT_SIZE: 4096,
}, },
parameter_rules=[ parameter_rules=[
ParameterRule( ParameterRule(
...@@ -459,8 +459,8 @@ LLM_BASE_MODELS = [ ...@@ -459,8 +459,8 @@ LLM_BASE_MODELS = [
model_type=ModelType.LLM, model_type=ModelType.LLM,
fetch_from=FetchFrom.CUSTOMIZABLE_MODEL, fetch_from=FetchFrom.CUSTOMIZABLE_MODEL,
model_properties={ model_properties={
'mode': LLMMode.COMPLETION.value, ModelPropertyKey.MODE: LLMMode.COMPLETION.value,
'context_size': 4096, ModelPropertyKey.CONTEXT_SIZE: 4096,
}, },
parameter_rules=[ parameter_rules=[
ParameterRule( ParameterRule(
......
...@@ -10,7 +10,7 @@ from core.model_runtime.entities.llm_entities import LLMResult, LLMResultChunk, ...@@ -10,7 +10,7 @@ from core.model_runtime.entities.llm_entities import LLMResult, LLMResultChunk,
from core.model_runtime.entities.message_entities import PromptMessage, PromptMessageTool, AssistantPromptMessage, \ from core.model_runtime.entities.message_entities import PromptMessage, PromptMessageTool, AssistantPromptMessage, \
UserPromptMessage, SystemPromptMessage UserPromptMessage, SystemPromptMessage
from core.model_runtime.entities.model_entities import ParameterRule, DefaultParameterName, AIModelEntity, ModelType, \ from core.model_runtime.entities.model_entities import ParameterRule, DefaultParameterName, AIModelEntity, ModelType, \
FetchFrom FetchFrom, ModelPropertyKey
from core.model_runtime.errors.validate import CredentialsValidateFailedError from core.model_runtime.errors.validate import CredentialsValidateFailedError
from core.model_runtime.model_providers.__base.large_language_model import LargeLanguageModel from core.model_runtime.model_providers.__base.large_language_model import LargeLanguageModel
from core.model_runtime.model_providers.huggingface_hub._common import _CommonHuggingfaceHub from core.model_runtime.model_providers.huggingface_hub._common import _CommonHuggingfaceHub
...@@ -97,7 +97,7 @@ class HuggingfaceHubLargeLanguageModel(_CommonHuggingfaceHub, LargeLanguageModel ...@@ -97,7 +97,7 @@ class HuggingfaceHubLargeLanguageModel(_CommonHuggingfaceHub, LargeLanguageModel
fetch_from=FetchFrom.CUSTOMIZABLE_MODEL, fetch_from=FetchFrom.CUSTOMIZABLE_MODEL,
model_type=ModelType.LLM, model_type=ModelType.LLM,
model_properties={ model_properties={
'mode': LLMMode.COMPLETION.value ModelPropertyKey.MODE: LLMMode.COMPLETION.value
}, },
parameter_rules=self._get_customizable_model_parameter_rules() parameter_rules=self._get_customizable_model_parameter_rules()
) )
......
...@@ -57,8 +57,8 @@ qdrant-client==1.6.4 ...@@ -57,8 +57,8 @@ qdrant-client==1.6.4
cohere~=4.32 cohere~=4.32
pyyaml~=6.0.1 pyyaml~=6.0.1
numpy~=1.25.2 numpy~=1.25.2
unstructured~=0.10.27
unstructured[docx,pptx,msg,md,ppt]~=0.10.27 unstructured[docx,pptx,msg,md,ppt]~=0.10.27
bs4~=0.0.1 bs4~=0.0.1
markdown~=3.5.1 markdown~=3.5.1
google-generativeai~=0.3.2 google-generativeai~=0.3.2
httpx[socks]~=0.24.1
\ No newline at end of file
Markdown is supported
0% or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment