Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

fixed bind_tools #223

Merged
merged 1 commit into from
May 15, 2024
Merged
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
31 changes: 19 additions & 12 deletions libs/vertexai/langchain_google_vertexai/chat_models.py
Original file line number Diff line number Diff line change
Expand Up @@ -126,7 +126,7 @@
"frequency_penalty",
"candidate_count",
]
_args_not_pass_to_prediction_service = _allowed_params + ["stream", "streaming"]
_allowed_params_prediction_service = ["request", "timeout", "metadata"]


@dataclass
Expand Down Expand Up @@ -490,13 +490,14 @@ def _completion_with_retry(
def _completion_with_retry_inner(generation_method: Callable, **kwargs: Any) -> Any:
return generation_method(**kwargs)

params = (
{k: v for k, v in kwargs.items() if k in _allowed_params_prediction_service}
if kwargs.get("is_gemini")
else kwargs
)
return _completion_with_retry_inner(
generation_method,
**{
k: v
for k, v in kwargs.items()
if v not in _args_not_pass_to_prediction_service
},
**params,
)


Expand All @@ -518,13 +519,14 @@ async def _completion_with_retry_inner(
) -> Any:
return await generation_method(**kwargs)

params = (
{k: v for k, v in kwargs.items() if k in _allowed_params_prediction_service}
if kwargs.get("is_gemini")
else kwargs
)
return await _completion_with_retry_inner(
generation_method,
**{
k: v
for k, v in kwargs.items()
if v not in _args_not_pass_to_prediction_service
},
**params,
)


Expand Down Expand Up @@ -648,7 +650,7 @@ def _generate(
messages=messages,
stop=stop,
run_manager=run_manager,
stream=stream,
is_gemini=True,
**kwargs,
)

Expand Down Expand Up @@ -736,6 +738,7 @@ def _generate_gemini(
self.prediction_client.generate_content,
max_retries=self.max_retries,
request=request,
**kwargs,
)
return self._gemini_response_to_chat_result(response)

Expand All @@ -752,6 +755,8 @@ async def _agenerate_gemini(
request=self._prepare_request_gemini(
messages=messages, stop=stop, **kwargs
),
is_gemini=True,
**kwargs,
)
return self._gemini_response_to_chat_result(response)

Expand Down Expand Up @@ -927,6 +932,7 @@ def _stream_gemini(
self.prediction_client.stream_generate_content,
max_retries=self.max_retries,
request=request,
is_gemini=True,
**kwargs,
)
for response_chunk in response_iter:
Expand Down Expand Up @@ -978,6 +984,7 @@ async def _astream(
self.async_prediction_client.stream_generate_content,
max_retries=self.max_retries,
request=request,
is_gemini=True,
**kwargs,
)
async for response_chunk in await response_iter:
Expand Down
Loading