REST Resource: projects.locations.endpoints

Resource: Endpoint

This type has no fields.

Models are deployed into it, and afterwards Endpoint is called to obtain predictions and explanations.

Methods

computeTokens

Return a list of tokens based on the input text.

countTokens

Perform a token counting.

fetchPredictOperation

Fetch an asynchronous online prediction operation.

generateContent

Generate content with multimodal inputs.

getIamPolicy

Gets the access control policy for a resource.

predict

Request message for running inference on Google's generative AI models on Vertex AI.

predictLongRunning

rawPredict

Perform an online prediction with an arbitrary HTTP payload.

serverStreamingPredict

Perform a server-side streaming online prediction request for Vertex LLM streaming.

setIamPolicy

Sets the access control policy on the specified resource.

streamGenerateContent

Generate content with multimodal inputs with streaming support.

streamRawPredict

Perform a streaming online prediction with an arbitrary HTTP payload.

testIamPermissions

Returns permissions that a caller has on the specified resource.

Except as otherwise noted, the content of this page is licensed under the Creative Commons Attribution 4.0 License, and code samples are licensed under the Apache 2.0 License. For details, see the Google Developers Site Policies. Java is a registered trademark of Oracle and/or its affiliates.

Last updated 2025年10月20日 UTC.