Skip to content

Commit d2d0312

Browse files
committed
fix: try to infer possible provider from the model name, resolves ScrapeGraphAI#805
1 parent 777a685 commit d2d0312

File tree

1 file changed

+12
-3
lines changed

1 file changed

+12
-3
lines changed

scrapegraphai/graphs/abstract_graph.py

+12-3
Original file line numberDiff line numberDiff line change
@@ -144,9 +144,18 @@ def _create_llm(self, llm_config: dict) -> object:
144144
"ollama", "oneapi", "nvidia", "groq", "anthropic", "bedrock", "mistralai",
145145
"hugging_face", "deepseek", "ernie", "fireworks", "togetherai"}
146146

147-
split_model_provider = llm_params["model"].split("/", 1)
148-
llm_params["model_provider"] = split_model_provider[0]
149-
llm_params["model"] = split_model_provider[1]
147+
if '/' in llm_params["model"]:
148+
split_model_provider = llm_params["model"].split("/", 1)
149+
llm_params["model_provider"] = split_model_provider[0]
150+
llm_params["model"] = split_model_provider[1]
151+
else:
152+
possible_providers = [provider for provider, models_d in models_tokens.items() if llm_params["model"] in models_d]
153+
if len(possible_providers) <= 0:
154+
raise ValueError(f"""Provider {llm_params['model_provider']} is not supported.
155+
If possible, try to use a model instance instead.""")
156+
llm_params["model_provider"] = possible_providers[0]
157+
print((f"Found providers {possible_providers} for model {llm_params['model']}, using {llm_params['model_provider']}.\n"
158+
"If it was not intended please specify the model provider in the graph configuration"))
150159

151160
if llm_params["model_provider"] not in known_providers:
152161
raise ValueError(f"""Provider {llm_params['model_provider']} is not supported.

0 commit comments

Comments
 (0)