From 467e5dd8c37481b359788a7a0f5d131c0f1355c5 Mon Sep 17 00:00:00 2001 From: riteshghorse Date: Tue, 18 Jul 2023 14:46:30 -0400 Subject: [PATCH] pydoc --- .../ml/inference/huggingface_inference.py | 15 ++++++++++----- 1 file changed, 10 insertions(+), 5 deletions(-) diff --git a/sdks/python/apache_beam/ml/inference/huggingface_inference.py b/sdks/python/apache_beam/ml/inference/huggingface_inference.py index 8c49b37b0e73..a7530a942dfc 100644 --- a/sdks/python/apache_beam/ml/inference/huggingface_inference.py +++ b/sdks/python/apache_beam/ml/inference/huggingface_inference.py @@ -171,7 +171,8 @@ def __init__( max_batch_size: Optional[int] = None, large_model: bool = False, **kwargs): - """Implementation of the abstract base class of ModelHandler interface + """ + Implementation of the abstract base class of ModelHandler interface for Hugging Face. This class shouldn't be instantiated directly. Use HuggingFaceModelHandlerKeyedTensor or HuggingFaceModelHandlerTensor. @@ -257,7 +258,8 @@ class HuggingFaceModelHandlerKeyedTensor( HuggingFaceModelHandler[Dict[str, Union[tf.Tensor, torch.Tensor]], PredictionResult, Union[AutoModel, TFAutoModel]]): - """Implementation of the ModelHandler interface for HuggingFace with + """ + Implementation of the ModelHandler interface for HuggingFace with Keyed Tensors for PyTorch/Tensorflow backend. Depending on the type of tensors, the model framework is determined @@ -275,7 +277,8 @@ def run_inference( model: Union[AutoModel, TFAutoModel], inference_args: Optional[Dict[str, Any]] = None ) -> Iterable[PredictionResult]: - """Runs inferences on a batch of Keyed Tensors and returns an Iterable of + """ + Runs inferences on a batch of Keyed Tensors and returns an Iterable of Tensors Predictions. This method stacks the list of Tensors in a vectorized format to optimize @@ -356,7 +359,8 @@ class HuggingFaceModelHandlerTensor(HuggingFaceModelHandler[Union[tf.Tensor, Union[AutoModel, TFAutoModel]] ): - """Implementation of the ModelHandler interface for HuggingFace with + """ + Implementation of the ModelHandler interface for HuggingFace with Tensors for PyTorch/Tensorflow backend. Depending on the type of tensors, the model framework is determined @@ -374,7 +378,8 @@ def run_inference( model: Union[AutoModel, TFAutoModel], inference_args: Optional[Dict[str, Any]] = None ) -> Iterable[PredictionResult]: - """Runs inferences on a batch of Tensors and returns an Iterable of + """ + Runs inferences on a batch of Tensors and returns an Iterable of Tensors Predictions. This method stacks the list of Tensors in a vectorized format to optimize