From e0d7c52703306411575956e124a32678f14ca5f7 Mon Sep 17 00:00:00 2001 From: Azure SDK for Python bot Date: Fri, 15 Jun 2018 10:39:36 -0700 Subject: [PATCH] [AutoPR cognitiveservices/data-plane/ComputerVision] Update CognitiveService ComputerVision API to V2 (#2747) * Generated from c2567da015994dac3301f495b6c5011d28ceaffe Add post-processing directive to swap argument order for RecognizeText to position the url argument ahead of the mode argument. autorest does not generate the intended order, so we add a post-processing directive, for each target language, * Generated from 973322cfd6ec2ced60a732f53b5318ceded5de7b Fix validation errors * Generated from 70efe042d607cbd973734e432da395ed35191a03 Add Spanish support for ServiceLanguage This is for parity with V1. --- .../computervision/computer_vision_api.py | 36 +++++++++---------- .../vision/computervision/models/__init__.py | 14 ++++---- .../computervision/models/category_detail.py | 5 +++ .../models/category_detail_py3.py | 7 +++- .../models/computer_vision_api_enums.py | 6 ++++ .../models/image_description.py | 4 +-- .../models/image_description_details.py | 9 ----- .../models/image_description_details_py3.py | 11 +----- .../models/image_description_py3.py | 4 +-- .../computervision/models/landmark_results.py | 4 +-- .../models/landmark_results_py3.py | 4 +-- ...s_landmarks_item.py => landmarks_model.py} | 4 +-- ...rks_item_py3.py => landmarks_model_py3.py} | 4 +-- .../vision/computervision/version.py | 2 +- 14 files changed, 56 insertions(+), 58 deletions(-) rename azure-cognitiveservices-vision-computervision/azure/cognitiveservices/vision/computervision/models/{landmark_results_landmarks_item.py => landmarks_model.py} (89%) rename azure-cognitiveservices-vision-computervision/azure/cognitiveservices/vision/computervision/models/{landmark_results_landmarks_item_py3.py => landmarks_model_py3.py} (90%) diff --git a/azure-cognitiveservices-vision-computervision/azure/cognitiveservices/vision/computervision/computer_vision_api.py b/azure-cognitiveservices-vision-computervision/azure/cognitiveservices/vision/computervision/computer_vision_api.py index 3d79fee9607d..e6999202c26b 100644 --- a/azure-cognitiveservices-vision-computervision/azure/cognitiveservices/vision/computervision/computer_vision_api.py +++ b/azure-cognitiveservices-vision-computervision/azure/cognitiveservices/vision/computervision/computer_vision_api.py @@ -41,7 +41,7 @@ def __init__( raise ValueError("Parameter 'azure_region' must not be None.") if credentials is None: raise ValueError("Parameter 'credentials' must not be None.") - base_url = 'https://{AzureRegion}.api.cognitive.microsoft.com/vision/v1.0' + base_url = 'https://{AzureRegion}.api.cognitive.microsoft.com/vision/v2.0' super(ComputerVisionAPIConfiguration, self).__init__(base_url) @@ -76,7 +76,7 @@ def __init__( super(ComputerVisionAPI, self).__init__(self.config.credentials, self.config) client_models = {k: v for k, v in models.__dict__.items() if isinstance(v, type)} - self.api_version = '1.0' + self.api_version = '2.0' self._serialize = Serializer(client_models) self._deserialize = Deserializer(client_models) @@ -625,18 +625,18 @@ def analyze_image_by_domain( analyze_image_by_domain.metadata = {'url': '/models/{model}/analyze'} def recognize_text( - self, url, detect_handwriting=False, custom_headers=None, raw=False, **operation_config): + self, url, mode, custom_headers=None, raw=False, **operation_config): """Recognize Text operation. When you use the Recognize Text interface, the response contains a field called 'Operation-Location'. The 'Operation-Location' field contains the URL that you must use for your - Get Handwritten Text Operation Result operation. + Get Recognize Text Operation Result operation. + :param mode: Type of text to recognize. Possible values include: + 'Handwritten', 'Printed' + :type mode: str or + ~azure.cognitiveservices.vision.computervision.models.TextRecognitionMode :param url: Publicly reachable URL of an image :type url: str - :param detect_handwriting: If 'true' is specified, handwriting - recognition is performed. If this parameter is set to 'false' or is - not specified, printed text recognition is performed. - :type detect_handwriting: bool :param dict custom_headers: headers that will be added to the request :param bool raw: returns the direct response alongside the deserialized response @@ -658,8 +658,7 @@ def recognize_text( # Construct parameters query_parameters = {} - if detect_handwriting is not None: - query_parameters['detectHandwriting'] = self._serialize.query("detect_handwriting", detect_handwriting, 'bool') + query_parameters['mode'] = self._serialize.query("mode", mode, 'TextRecognitionMode') # Construct headers header_parameters = {} @@ -693,7 +692,7 @@ def get_text_operation_result( returned from Recognize Text interface. :param operation_id: Id of the text operation returned in the response - of the 'Recognize Handwritten Text' + of the 'Recognize Text' :type operation_id: str :param dict custom_headers: headers that will be added to the request :param bool raw: returns the direct response alongside the @@ -1240,18 +1239,18 @@ def analyze_image_by_domain_in_stream( analyze_image_by_domain_in_stream.metadata = {'url': '/models/{model}/analyze'} def recognize_text_in_stream( - self, image, detect_handwriting=False, custom_headers=None, raw=False, callback=None, **operation_config): + self, image, mode, custom_headers=None, raw=False, callback=None, **operation_config): """Recognize Text operation. When you use the Recognize Text interface, the response contains a field called 'Operation-Location'. The 'Operation-Location' field contains the URL that you must use for your - Get Handwritten Text Operation Result operation. + Get Recognize Text Operation Result operation. :param image: An image stream. :type image: Generator - :param detect_handwriting: If 'true' is specified, handwriting - recognition is performed. If this parameter is set to 'false' or is - not specified, printed text recognition is performed. - :type detect_handwriting: bool + :param mode: Type of text to recognize. Possible values include: + 'Handwritten', 'Printed' + :type mode: str or + ~azure.cognitiveservices.vision.computervision.models.TextRecognitionMode :param dict custom_headers: headers that will be added to the request :param bool raw: returns the direct response alongside the deserialized response @@ -1276,8 +1275,7 @@ def recognize_text_in_stream( # Construct parameters query_parameters = {} - if detect_handwriting is not None: - query_parameters['detectHandwriting'] = self._serialize.query("detect_handwriting", detect_handwriting, 'bool') + query_parameters['mode'] = self._serialize.query("mode", mode, 'TextRecognitionMode') # Construct headers header_parameters = {} diff --git a/azure-cognitiveservices-vision-computervision/azure/cognitiveservices/vision/computervision/models/__init__.py b/azure-cognitiveservices-vision-computervision/azure/cognitiveservices/vision/computervision/models/__init__.py index f0dde33bd544..001e34518c57 100644 --- a/azure-cognitiveservices-vision-computervision/azure/cognitiveservices/vision/computervision/models/__init__.py +++ b/azure-cognitiveservices-vision-computervision/azure/cognitiveservices/vision/computervision/models/__init__.py @@ -16,6 +16,7 @@ from .text_operation_result_py3 import TextOperationResult from .face_rectangle_py3 import FaceRectangle from .celebrities_model_py3 import CelebritiesModel + from .landmarks_model_py3 import LandmarksModel from .category_detail_py3 import CategoryDetail from .category_py3 import Category from .adult_info_py3 import AdultInfo @@ -23,9 +24,9 @@ from .image_type_py3 import ImageType from .image_tag_py3 import ImageTag from .image_caption_py3 import ImageCaption - from .image_metadata_py3 import ImageMetadata from .image_description_details_py3 import ImageDescriptionDetails from .face_description_py3 import FaceDescription + from .image_metadata_py3 import ImageMetadata from .image_analysis_py3 import ImageAnalysis from .ocr_word_py3 import OcrWord from .ocr_line_py3 import OcrLine @@ -35,7 +36,6 @@ from .list_models_result_py3 import ListModelsResult from .domain_model_results_py3 import DomainModelResults from .celebrity_results_py3 import CelebrityResults - from .landmark_results_landmarks_item_py3 import LandmarkResultsLandmarksItem from .landmark_results_py3 import LandmarkResults from .image_description_py3 import ImageDescription from .tag_result_py3 import TagResult @@ -48,6 +48,7 @@ from .text_operation_result import TextOperationResult from .face_rectangle import FaceRectangle from .celebrities_model import CelebritiesModel + from .landmarks_model import LandmarksModel from .category_detail import CategoryDetail from .category import Category from .adult_info import AdultInfo @@ -55,9 +56,9 @@ from .image_type import ImageType from .image_tag import ImageTag from .image_caption import ImageCaption - from .image_metadata import ImageMetadata from .image_description_details import ImageDescriptionDetails from .face_description import FaceDescription + from .image_metadata import ImageMetadata from .image_analysis import ImageAnalysis from .ocr_word import OcrWord from .ocr_line import OcrLine @@ -67,7 +68,6 @@ from .list_models_result import ListModelsResult from .domain_model_results import DomainModelResults from .celebrity_results import CelebrityResults - from .landmark_results_landmarks_item import LandmarkResultsLandmarksItem from .landmark_results import LandmarkResults from .image_description import ImageDescription from .tag_result import TagResult @@ -79,6 +79,7 @@ ComputerVisionErrorCodes, VisualFeatureTypes, OcrLanguages, + TextRecognitionMode, AzureRegions, Details, ) @@ -90,6 +91,7 @@ 'TextOperationResult', 'FaceRectangle', 'CelebritiesModel', + 'LandmarksModel', 'CategoryDetail', 'Category', 'AdultInfo', @@ -97,9 +99,9 @@ 'ImageType', 'ImageTag', 'ImageCaption', - 'ImageMetadata', 'ImageDescriptionDetails', 'FaceDescription', + 'ImageMetadata', 'ImageAnalysis', 'OcrWord', 'OcrLine', @@ -109,7 +111,6 @@ 'ListModelsResult', 'DomainModelResults', 'CelebrityResults', - 'LandmarkResultsLandmarksItem', 'LandmarkResults', 'ImageDescription', 'TagResult', @@ -120,6 +121,7 @@ 'ComputerVisionErrorCodes', 'VisualFeatureTypes', 'OcrLanguages', + 'TextRecognitionMode', 'AzureRegions', 'Details', ] diff --git a/azure-cognitiveservices-vision-computervision/azure/cognitiveservices/vision/computervision/models/category_detail.py b/azure-cognitiveservices-vision-computervision/azure/cognitiveservices/vision/computervision/models/category_detail.py index eefbe5b691dc..92c3e355d3e6 100644 --- a/azure-cognitiveservices-vision-computervision/azure/cognitiveservices/vision/computervision/models/category_detail.py +++ b/azure-cognitiveservices-vision-computervision/azure/cognitiveservices/vision/computervision/models/category_detail.py @@ -18,12 +18,17 @@ class CategoryDetail(Model): :param celebrities: An array of celebrities if any identified. :type celebrities: list[~azure.cognitiveservices.vision.computervision.models.CelebritiesModel] + :param landmarks: An array of landmarks if any identified. + :type landmarks: + list[~azure.cognitiveservices.vision.computervision.models.LandmarksModel] """ _attribute_map = { 'celebrities': {'key': 'celebrities', 'type': '[CelebritiesModel]'}, + 'landmarks': {'key': 'landmarks', 'type': '[LandmarksModel]'}, } def __init__(self, **kwargs): super(CategoryDetail, self).__init__(**kwargs) self.celebrities = kwargs.get('celebrities', None) + self.landmarks = kwargs.get('landmarks', None) diff --git a/azure-cognitiveservices-vision-computervision/azure/cognitiveservices/vision/computervision/models/category_detail_py3.py b/azure-cognitiveservices-vision-computervision/azure/cognitiveservices/vision/computervision/models/category_detail_py3.py index b155197dab02..3e4350314499 100644 --- a/azure-cognitiveservices-vision-computervision/azure/cognitiveservices/vision/computervision/models/category_detail_py3.py +++ b/azure-cognitiveservices-vision-computervision/azure/cognitiveservices/vision/computervision/models/category_detail_py3.py @@ -18,12 +18,17 @@ class CategoryDetail(Model): :param celebrities: An array of celebrities if any identified. :type celebrities: list[~azure.cognitiveservices.vision.computervision.models.CelebritiesModel] + :param landmarks: An array of landmarks if any identified. + :type landmarks: + list[~azure.cognitiveservices.vision.computervision.models.LandmarksModel] """ _attribute_map = { 'celebrities': {'key': 'celebrities', 'type': '[CelebritiesModel]'}, + 'landmarks': {'key': 'landmarks', 'type': '[LandmarksModel]'}, } - def __init__(self, *, celebrities=None, **kwargs) -> None: + def __init__(self, *, celebrities=None, landmarks=None, **kwargs) -> None: super(CategoryDetail, self).__init__(**kwargs) self.celebrities = celebrities + self.landmarks = landmarks diff --git a/azure-cognitiveservices-vision-computervision/azure/cognitiveservices/vision/computervision/models/computer_vision_api_enums.py b/azure-cognitiveservices-vision-computervision/azure/cognitiveservices/vision/computervision/models/computer_vision_api_enums.py index 76ff22284087..66fb4dc38436 100644 --- a/azure-cognitiveservices-vision-computervision/azure/cognitiveservices/vision/computervision/models/computer_vision_api_enums.py +++ b/azure-cognitiveservices-vision-computervision/azure/cognitiveservices/vision/computervision/models/computer_vision_api_enums.py @@ -85,6 +85,12 @@ class OcrLanguages(str, Enum): sk = "sk" +class TextRecognitionMode(str, Enum): + + handwritten = "Handwritten" + printed = "Printed" + + class AzureRegions(str, Enum): westus = "westus" diff --git a/azure-cognitiveservices-vision-computervision/azure/cognitiveservices/vision/computervision/models/image_description.py b/azure-cognitiveservices-vision-computervision/azure/cognitiveservices/vision/computervision/models/image_description.py index 6a84422b6b65..18fbae6eb962 100644 --- a/azure-cognitiveservices-vision-computervision/azure/cognitiveservices/vision/computervision/models/image_description.py +++ b/azure-cognitiveservices-vision-computervision/azure/cognitiveservices/vision/computervision/models/image_description.py @@ -31,8 +31,8 @@ class ImageDescription(Model): _attribute_map = { 'tags': {'key': 'description.tags', 'type': '[str]'}, 'captions': {'key': 'description.captions', 'type': '[ImageCaption]'}, - 'request_id': {'key': 'description.requestId', 'type': 'str'}, - 'metadata': {'key': 'description.metadata', 'type': 'ImageMetadata'}, + 'request_id': {'key': 'requestId', 'type': 'str'}, + 'metadata': {'key': 'metadata', 'type': 'ImageMetadata'}, } def __init__(self, **kwargs): diff --git a/azure-cognitiveservices-vision-computervision/azure/cognitiveservices/vision/computervision/models/image_description_details.py b/azure-cognitiveservices-vision-computervision/azure/cognitiveservices/vision/computervision/models/image_description_details.py index 1e6afbb99ed7..f6b7524d1543 100644 --- a/azure-cognitiveservices-vision-computervision/azure/cognitiveservices/vision/computervision/models/image_description_details.py +++ b/azure-cognitiveservices-vision-computervision/azure/cognitiveservices/vision/computervision/models/image_description_details.py @@ -21,23 +21,14 @@ class ImageDescriptionDetails(Model): :param captions: A list of captions, sorted by confidence level. :type captions: list[~azure.cognitiveservices.vision.computervision.models.ImageCaption] - :param request_id: Id of the REST API request. - :type request_id: str - :param metadata: - :type metadata: - ~azure.cognitiveservices.vision.computervision.models.ImageMetadata """ _attribute_map = { 'tags': {'key': 'tags', 'type': '[str]'}, 'captions': {'key': 'captions', 'type': '[ImageCaption]'}, - 'request_id': {'key': 'requestId', 'type': 'str'}, - 'metadata': {'key': 'metadata', 'type': 'ImageMetadata'}, } def __init__(self, **kwargs): super(ImageDescriptionDetails, self).__init__(**kwargs) self.tags = kwargs.get('tags', None) self.captions = kwargs.get('captions', None) - self.request_id = kwargs.get('request_id', None) - self.metadata = kwargs.get('metadata', None) diff --git a/azure-cognitiveservices-vision-computervision/azure/cognitiveservices/vision/computervision/models/image_description_details_py3.py b/azure-cognitiveservices-vision-computervision/azure/cognitiveservices/vision/computervision/models/image_description_details_py3.py index 702d4ac029de..dcb990fc688e 100644 --- a/azure-cognitiveservices-vision-computervision/azure/cognitiveservices/vision/computervision/models/image_description_details_py3.py +++ b/azure-cognitiveservices-vision-computervision/azure/cognitiveservices/vision/computervision/models/image_description_details_py3.py @@ -21,23 +21,14 @@ class ImageDescriptionDetails(Model): :param captions: A list of captions, sorted by confidence level. :type captions: list[~azure.cognitiveservices.vision.computervision.models.ImageCaption] - :param request_id: Id of the REST API request. - :type request_id: str - :param metadata: - :type metadata: - ~azure.cognitiveservices.vision.computervision.models.ImageMetadata """ _attribute_map = { 'tags': {'key': 'tags', 'type': '[str]'}, 'captions': {'key': 'captions', 'type': '[ImageCaption]'}, - 'request_id': {'key': 'requestId', 'type': 'str'}, - 'metadata': {'key': 'metadata', 'type': 'ImageMetadata'}, } - def __init__(self, *, tags=None, captions=None, request_id: str=None, metadata=None, **kwargs) -> None: + def __init__(self, *, tags=None, captions=None, **kwargs) -> None: super(ImageDescriptionDetails, self).__init__(**kwargs) self.tags = tags self.captions = captions - self.request_id = request_id - self.metadata = metadata diff --git a/azure-cognitiveservices-vision-computervision/azure/cognitiveservices/vision/computervision/models/image_description_py3.py b/azure-cognitiveservices-vision-computervision/azure/cognitiveservices/vision/computervision/models/image_description_py3.py index 3ec3fa9c951a..0874f3c62e01 100644 --- a/azure-cognitiveservices-vision-computervision/azure/cognitiveservices/vision/computervision/models/image_description_py3.py +++ b/azure-cognitiveservices-vision-computervision/azure/cognitiveservices/vision/computervision/models/image_description_py3.py @@ -31,8 +31,8 @@ class ImageDescription(Model): _attribute_map = { 'tags': {'key': 'description.tags', 'type': '[str]'}, 'captions': {'key': 'description.captions', 'type': '[ImageCaption]'}, - 'request_id': {'key': 'description.requestId', 'type': 'str'}, - 'metadata': {'key': 'description.metadata', 'type': 'ImageMetadata'}, + 'request_id': {'key': 'requestId', 'type': 'str'}, + 'metadata': {'key': 'metadata', 'type': 'ImageMetadata'}, } def __init__(self, *, tags=None, captions=None, request_id: str=None, metadata=None, **kwargs) -> None: diff --git a/azure-cognitiveservices-vision-computervision/azure/cognitiveservices/vision/computervision/models/landmark_results.py b/azure-cognitiveservices-vision-computervision/azure/cognitiveservices/vision/computervision/models/landmark_results.py index 14bd0e21d982..7b0c45eb32c6 100644 --- a/azure-cognitiveservices-vision-computervision/azure/cognitiveservices/vision/computervision/models/landmark_results.py +++ b/azure-cognitiveservices-vision-computervision/azure/cognitiveservices/vision/computervision/models/landmark_results.py @@ -17,7 +17,7 @@ class LandmarkResults(Model): :param landmarks: :type landmarks: - list[~azure.cognitiveservices.vision.computervision.models.LandmarkResultsLandmarksItem] + list[~azure.cognitiveservices.vision.computervision.models.LandmarksModel] :param request_id: Id of the REST API request. :type request_id: str :param metadata: @@ -26,7 +26,7 @@ class LandmarkResults(Model): """ _attribute_map = { - 'landmarks': {'key': 'landmarks', 'type': '[LandmarkResultsLandmarksItem]'}, + 'landmarks': {'key': 'landmarks', 'type': '[LandmarksModel]'}, 'request_id': {'key': 'requestId', 'type': 'str'}, 'metadata': {'key': 'metadata', 'type': 'ImageMetadata'}, } diff --git a/azure-cognitiveservices-vision-computervision/azure/cognitiveservices/vision/computervision/models/landmark_results_py3.py b/azure-cognitiveservices-vision-computervision/azure/cognitiveservices/vision/computervision/models/landmark_results_py3.py index 1df503e9eccf..93228b4cf03f 100644 --- a/azure-cognitiveservices-vision-computervision/azure/cognitiveservices/vision/computervision/models/landmark_results_py3.py +++ b/azure-cognitiveservices-vision-computervision/azure/cognitiveservices/vision/computervision/models/landmark_results_py3.py @@ -17,7 +17,7 @@ class LandmarkResults(Model): :param landmarks: :type landmarks: - list[~azure.cognitiveservices.vision.computervision.models.LandmarkResultsLandmarksItem] + list[~azure.cognitiveservices.vision.computervision.models.LandmarksModel] :param request_id: Id of the REST API request. :type request_id: str :param metadata: @@ -26,7 +26,7 @@ class LandmarkResults(Model): """ _attribute_map = { - 'landmarks': {'key': 'landmarks', 'type': '[LandmarkResultsLandmarksItem]'}, + 'landmarks': {'key': 'landmarks', 'type': '[LandmarksModel]'}, 'request_id': {'key': 'requestId', 'type': 'str'}, 'metadata': {'key': 'metadata', 'type': 'ImageMetadata'}, } diff --git a/azure-cognitiveservices-vision-computervision/azure/cognitiveservices/vision/computervision/models/landmark_results_landmarks_item.py b/azure-cognitiveservices-vision-computervision/azure/cognitiveservices/vision/computervision/models/landmarks_model.py similarity index 89% rename from azure-cognitiveservices-vision-computervision/azure/cognitiveservices/vision/computervision/models/landmark_results_landmarks_item.py rename to azure-cognitiveservices-vision-computervision/azure/cognitiveservices/vision/computervision/models/landmarks_model.py index fd4de1c7afbc..74ff012df89a 100644 --- a/azure-cognitiveservices-vision-computervision/azure/cognitiveservices/vision/computervision/models/landmark_results_landmarks_item.py +++ b/azure-cognitiveservices-vision-computervision/azure/cognitiveservices/vision/computervision/models/landmarks_model.py @@ -12,7 +12,7 @@ from msrest.serialization import Model -class LandmarkResultsLandmarksItem(Model): +class LandmarksModel(Model): """A landmark recognized in the image. :param name: Name of the landmark. @@ -27,6 +27,6 @@ class LandmarkResultsLandmarksItem(Model): } def __init__(self, **kwargs): - super(LandmarkResultsLandmarksItem, self).__init__(**kwargs) + super(LandmarksModel, self).__init__(**kwargs) self.name = kwargs.get('name', None) self.confidence = kwargs.get('confidence', None) diff --git a/azure-cognitiveservices-vision-computervision/azure/cognitiveservices/vision/computervision/models/landmark_results_landmarks_item_py3.py b/azure-cognitiveservices-vision-computervision/azure/cognitiveservices/vision/computervision/models/landmarks_model_py3.py similarity index 90% rename from azure-cognitiveservices-vision-computervision/azure/cognitiveservices/vision/computervision/models/landmark_results_landmarks_item_py3.py rename to azure-cognitiveservices-vision-computervision/azure/cognitiveservices/vision/computervision/models/landmarks_model_py3.py index 69efe9302103..9b7ed4502168 100644 --- a/azure-cognitiveservices-vision-computervision/azure/cognitiveservices/vision/computervision/models/landmark_results_landmarks_item_py3.py +++ b/azure-cognitiveservices-vision-computervision/azure/cognitiveservices/vision/computervision/models/landmarks_model_py3.py @@ -12,7 +12,7 @@ from msrest.serialization import Model -class LandmarkResultsLandmarksItem(Model): +class LandmarksModel(Model): """A landmark recognized in the image. :param name: Name of the landmark. @@ -27,6 +27,6 @@ class LandmarkResultsLandmarksItem(Model): } def __init__(self, *, name: str=None, confidence: float=None, **kwargs) -> None: - super(LandmarkResultsLandmarksItem, self).__init__(**kwargs) + super(LandmarksModel, self).__init__(**kwargs) self.name = name self.confidence = confidence diff --git a/azure-cognitiveservices-vision-computervision/azure/cognitiveservices/vision/computervision/version.py b/azure-cognitiveservices-vision-computervision/azure/cognitiveservices/vision/computervision/version.py index 63d89bfb54fa..63f83465c874 100644 --- a/azure-cognitiveservices-vision-computervision/azure/cognitiveservices/vision/computervision/version.py +++ b/azure-cognitiveservices-vision-computervision/azure/cognitiveservices/vision/computervision/version.py @@ -9,5 +9,5 @@ # regenerated. # -------------------------------------------------------------------------- -VERSION = "1.0" +VERSION = "2.0"