Value

not co-occur with SPACE, LEADER_SPACE, or LINE_BREAK. Line break that ends a paragraph.

Likelihood

static

number

A bucketized representation of likelihood, which is intended to give clients highly stable results across model upgrades.

Value

UNKNOWN

Unknown likelihood.

VERY_UNLIKELY

It is very unlikely that the image belongs to the specified vertical.

UNLIKELY

It is unlikely that the image belongs to the specified vertical.

POSSIBLE

It is possible that the image belongs to the specified vertical.

LIKELY

It is likely that the image belongs to the specified vertical.

VERY_LIKELY

It is very likely that the image belongs to the specified vertical.

Type

static

number

Face landmark (feature) type. Left and right are defined from the vantage of the viewer of the image without considering mirror projections typical of photos. So, LEFT_EYE, typically, is the person's right eye.

Value

UNKNOWN_LANDMARK

Unknown face landmark detected. Should not be filled.

LEFT_EYE

Left eye.

RIGHT_EYE

Right eye.

LEFT_OF_LEFT_EYEBROW

Left of left eyebrow.

RIGHT_OF_LEFT_EYEBROW

Right of left eyebrow.

LEFT_OF_RIGHT_EYEBROW

Left of right eyebrow.

RIGHT_OF_RIGHT_EYEBROW

Right of right eyebrow.

MIDPOINT_BETWEEN_EYES

Midpoint between eyes.

NOSE_TIP

Nose tip.

UPPER_LIP

Upper lip.

LOWER_LIP

Lower lip.

MOUTH_LEFT

Mouth left.

MOUTH_RIGHT

Mouth right.

MOUTH_CENTER

Mouth center.

NOSE_BOTTOM_RIGHT

Nose, bottom right.

NOSE_BOTTOM_LEFT

Nose, bottom left.

NOSE_BOTTOM_CENTER

Nose, bottom center.

LEFT_EYE_TOP_BOUNDARY

Left eye, top boundary.

LEFT_EYE_RIGHT_CORNER

Left eye, right corner.

LEFT_EYE_BOTTOM_BOUNDARY

Left eye, bottom boundary.

LEFT_EYE_LEFT_CORNER

Left eye, left corner.

RIGHT_EYE_TOP_BOUNDARY

Right eye, top boundary.

RIGHT_EYE_RIGHT_CORNER

Right eye, right corner.

RIGHT_EYE_BOTTOM_BOUNDARY

Right eye, bottom boundary.

RIGHT_EYE_LEFT_CORNER

Right eye, left corner.

LEFT_EYEBROW_UPPER_MIDPOINT

Left eyebrow, upper midpoint.

RIGHT_EYEBROW_UPPER_MIDPOINT

Right eyebrow, upper midpoint.

LEFT_EAR_TRAGION

Left ear tragion.

RIGHT_EAR_TRAGION

Right ear tragion.

LEFT_EYE_PUPIL

Left eye pupil.

RIGHT_EYE_PUPIL

Right eye pupil.

FOREHEAD_GLABELLA

Forehead glabella.

CHIN_GNATHION

Chin gnathion.

CHIN_LEFT_GONION

Chin left gonion.

CHIN_RIGHT_GONION

Chin right gonion.

Type

static

number

Type of image feature.

Value

TYPE_UNSPECIFIED

Unspecified feature type.

FACE_DETECTION

Run face detection.

LANDMARK_DETECTION

Run landmark detection.

LOGO_DETECTION

Run logo detection.

LABEL_DETECTION

Run label detection.

TEXT_DETECTION

Run OCR.

DOCUMENT_TEXT_DETECTION

Run dense text document OCR. Takes precedence when both DOCUMENT_TEXT_DETECTION and TEXT_DETECTION are present.

SAFE_SEARCH_DETECTION

Run computer vision models to compute image safe-search properties.

IMAGE_PROPERTIES

Compute a set of image properties, such as the image's dominant colors.

CROP_HINTS

Run crop hints.

WEB_DETECTION

Run web detection.

Properties

BlockType

static

number

Type of a block (text, image etc) as identified by OCR.

Value

UNKNOWN

Unknown block type.

TEXT

Regular text block.

TABLE

Table block.

PICTURE

Image block.

RULER

Horizontal/vertical line box.

BARCODE

Barcode block.

BreakType

static

number

Enum to denote the type of break found. New line, space etc.

Value

UNKNOWN

Unknown break label type.

SPACE

Regular space.

SURE_SPACE

Sure space (very wide).

EOL_SURE_SPACE

Line-wrapping break.

HYPHEN

End-line hyphen that is not present in text; does

LINE_BREAK

not co-occur with SPACE, LEADER_SPACE, or LINE_BREAK. Line break that ends a paragraph.

Likelihood

static

number

A bucketized representation of likelihood, which is intended to give clients highly stable results across model upgrades.

Value

UNKNOWN

Unknown likelihood.

VERY_UNLIKELY

It is very unlikely that the image belongs to the specified vertical.

UNLIKELY

It is unlikely that the image belongs to the specified vertical.

POSSIBLE

It is possible that the image belongs to the specified vertical.

LIKELY

It is likely that the image belongs to the specified vertical.

VERY_LIKELY

It is very likely that the image belongs to the specified vertical.

Type

static

number

Face landmark (feature) type. Left and right are defined from the vantage of the viewer of the image without considering mirror projections typical of photos. So, LEFT_EYE, typically, is the person's right eye.

Value

UNKNOWN_LANDMARK

Unknown face landmark detected. Should not be filled.

LEFT_EYE

Left eye.

RIGHT_EYE

Right eye.

LEFT_OF_LEFT_EYEBROW

Left of left eyebrow.

RIGHT_OF_LEFT_EYEBROW

Right of left eyebrow.

LEFT_OF_RIGHT_EYEBROW

Left of right eyebrow.

RIGHT_OF_RIGHT_EYEBROW

Right of right eyebrow.

MIDPOINT_BETWEEN_EYES

Midpoint between eyes.

NOSE_TIP

Nose tip.

UPPER_LIP

Upper lip.

LOWER_LIP

Lower lip.

MOUTH_LEFT

Mouth left.

MOUTH_RIGHT

Mouth right.

MOUTH_CENTER

Mouth center.

NOSE_BOTTOM_RIGHT

Nose, bottom right.

NOSE_BOTTOM_LEFT

Nose, bottom left.

NOSE_BOTTOM_CENTER

Nose, bottom center.

LEFT_EYE_TOP_BOUNDARY

Left eye, top boundary.

LEFT_EYE_RIGHT_CORNER

Left eye, right corner.

LEFT_EYE_BOTTOM_BOUNDARY

Left eye, bottom boundary.

LEFT_EYE_LEFT_CORNER

Left eye, left corner.

RIGHT_EYE_TOP_BOUNDARY

Right eye, top boundary.

RIGHT_EYE_RIGHT_CORNER

Right eye, right corner.

RIGHT_EYE_BOTTOM_BOUNDARY

Right eye, bottom boundary.

RIGHT_EYE_LEFT_CORNER

Right eye, left corner.

LEFT_EYEBROW_UPPER_MIDPOINT

Left eyebrow, upper midpoint.

RIGHT_EYEBROW_UPPER_MIDPOINT

Right eyebrow, upper midpoint.

LEFT_EAR_TRAGION

Left ear tragion.

RIGHT_EAR_TRAGION

Right ear tragion.

LEFT_EYE_PUPIL

Left eye pupil.

RIGHT_EYE_PUPIL

Right eye pupil.

FOREHEAD_GLABELLA

Forehead glabella.

CHIN_GNATHION

Chin gnathion.

CHIN_LEFT_GONION

Chin left gonion.

CHIN_RIGHT_GONION

Chin right gonion.

Type

static

number

Type of image feature.

Value

TYPE_UNSPECIFIED

Unspecified feature type.

FACE_DETECTION

Run face detection.

LANDMARK_DETECTION

Run landmark detection.

LOGO_DETECTION

Run logo detection.

LABEL_DETECTION

Run label detection.

TEXT_DETECTION

Run OCR.

DOCUMENT_TEXT_DETECTION

Run dense text document OCR. Takes precedence when both DOCUMENT_TEXT_DETECTION and TEXT_DETECTION are present.

SAFE_SEARCH_DETECTION

Run computer vision models to compute image safe-search properties.

IMAGE_PROPERTIES

Compute a set of image properties, such as the image's dominant colors.

Block

Properties

Parameter

The bounding box for the block. The vertices are in the order of top-left, top-right, bottom-right, bottom-left. When a rotation of the bounding box is detected the rotation is represented as around the top-left corner as defined
when the text is read in the 'natural' orientation. For example:

when the text is horizontal it might look like:
0----1
| |
3----2
when it's rotated 180 degrees around the top-left corner it becomes:
2----3
| |
1----0
and the vertice order will still be (0, 1, 2, 3).

Property

Parameter

CropHintsParams

static

Parameters for crop hints annotation request.

Property

Parameter

aspectRatios

Array of number

Aspect ratios in floats, representing the ratio of the width to the height of the image. For example, if the desired aspect ratio is 4/3, the corresponding float value should be 1.33333. If not specified, the best possible crop
is returned. The number of provided aspect ratios is limited to a maximum of 16; any aspect ratios provided after the 16th are ignored.

EntityAnnotation

Properties

Parameter

The language code for the locale in which the entity textual
description is expressed.

description

string

Entity textual description, expressed in its locale language.

score

number

Overall score of the result. Range [0, 1].

confidence

number

The accuracy of the entity detection in an image. For example, for an image in which the "Eiffel Tower" entity is detected, this field represents the confidence that there is a tower in the query image. Range [0, 1].

topicality

number

The relevancy of the ICA (Image Content Annotation) label to the image. For example, the relevancy of "tower" is likely higher to an image containing the detected "Eiffel Tower" than to an image containing a
detected distant towering building, even though the confidence that there is a tower in each image may be the same. Range [0, 1].

boundingPoly

Object

Image region to which this entity belongs. Currently not produced for LABEL_DETECTION features. For TEXT_DETECTION (OCR), boundingPolys are produced for the entire text detected in an image
region, followed by
boundingPolys for each word within the detected text.

The location information for the detected entity. Multiple
LocationInfo elements can be present because one location may indicate the location of the scene in the image, and another location may indicate the location of the place where the image was taken. Location information
is usually present for landmarks.

FaceAnnotation

static

A face annotation object contains the results of face detection.

Properties

Parameter

boundingPoly

Object

The bounding polygon around the face. The coordinates of the bounding box are in the original image's scale, as returned in ImageParams. The bounding box is computed to "frame" the face in accordance with
human expectations. It is based on the landmarker results. Note that one or more x and/or y coordinates may not be generated in the
BoundingPoly (the polygon will be unbounded) if only a partial face appears in the image to be annotated.

The fd_bounding_poly bounding polygon is tighter than the
boundingPoly, and encloses only the skin part of the face. Typically, it is used to eliminate the face from any image analysis that detects the "amount of skin" visible in an image. It is not based on the
landmarker results, only on the initial face detection, hence the fd (face detection) prefix.

Feature

static

Users describe the type of Google Cloud Vision API tasks to perform over images by using Features. Each Feature indicates a type of image detection task to perform. Features encode the Cloud Vision API vertical to operate on and
the number of top-scoring results to return.

Properties

Parameter

List of languages to use for TEXT_DETECTION. In most cases, an empty value yields the best results since it enables automatic language detection. For languages based on the Latin alphabet, setting language_hints is
not needed. In rare cases, when the language of the text in the image is known, setting a hint will help get better results (although it will be a significant hindrance if the hint is wrong). Text detection returns an error if
one or more of the specified languages is not one of the
supported languages.

ImageSource

static

External image source (Google Cloud Storage image location).

Properties

Parameter

gcsImageUri

string

NOTE: For new code image_uri below is preferred. Google Cloud Storage image URI, which must be in the following form:
gs://bucket_name/object_name (for details, see
Google Cloud Storage Request
URIs). NOTE: Cloud Storage object versioning is not supported.

imageUri

string

Image URI which supports: 1) Google Cloud Storage image URI, which must be in the following form:
gs://bucket_name/object_name (for details, see
Google Cloud Storage Request
URIs). NOTE: Cloud Storage object versioning is not supported. 2) Publicly accessible image HTTP/HTTPS URL. This is preferred over the legacy gcs_image_uri above. When both
gcs_image_uri and image_uri are specified, image_uri takes precedence.

Landmark

static

A face-specific landmark (for example, a face feature). Landmark positions may fall outside the bounds of the image if the face is near one or more edges of the image. Therefore it is NOT guaranteed that 0 <= x < width or
0 <= y < height.

Properties

Parameter

The bounding box for the paragraph. The vertices are in the order of top-left, top-right, bottom-right, bottom-left. When a rotation of the bounding box is detected the rotation is represented as around the top-left corner as defined
when the text is read in the 'natural' orientation. For example:

when the text is horizontal it might look like:
0----1
| |
3----2
when it's rotated 180 degrees around the top-left corner it becomes:
2----3
| |
1----0
and the vertice order will still be (0, 1, 2, 3).

Symbol

Properties

Parameter

The bounding box for the symbol. The vertices are in the order of top-left, top-right, bottom-right, bottom-left. When a rotation of the bounding box is detected the rotation is represented as around the top-left corner as defined
when the text is read in the 'natural' orientation. For example:

when the text is horizontal it might look like:
0----1
| |
3----2
when it's rotated 180 degrees around the top-left corner it becomes:
2----3
| |
1----0
and the vertice order will still be (0, 1, 2, 3).

Word

Properties

Parameter

The bounding box for the word. The vertices are in the order of top-left, top-right, bottom-right, bottom-left. When a rotation of the bounding box is detected the rotation is represented as around the top-left corner as defined
when the text is read in the 'natural' orientation. For example:

when the text is horizontal it might look like:
0----1
| |
3----2
when it's rotated 180 degrees around the top-left corner it becomes:
2----3
| |
1----0
and the vertice order will still be (0, 1, 2, 3).