Skip to content

/AWS1/CL_SGM=>CREATEINFERENCECOMPONENT()

About CreateInferenceComponent

Creates an inference component, which is a SageMaker hosting object that you can use to deploy a model to an endpoint. In the inference component settings, you specify the model, the endpoint, and how the model utilizes the resources that the endpoint hosts. You can optimize resource utilization by tailoring how the required CPU cores, accelerators, and memory are allocated. You can deploy multiple inference components to an endpoint, where each inference component contains one model and the resource utilization needs for that individual model. After you deploy an inference component, you can directly invoke the associated model when you use the InvokeEndpoint API action.

Method Signature

IMPORTING

Required arguments:

IV_INFERENCECOMPONENTNAME TYPE /AWS1/SGMINFERENCECOMPONENTN00 /AWS1/SGMINFERENCECOMPONENTN00

A unique name to assign to the inference component.

IV_ENDPOINTNAME TYPE /AWS1/SGMENDPOINTNAME /AWS1/SGMENDPOINTNAME

The name of an existing endpoint where you host the inference component.

IV_VARIANTNAME TYPE /AWS1/SGMVARIANTNAME /AWS1/SGMVARIANTNAME

The name of an existing production variant where you host the inference component.

IO_SPECIFICATION TYPE REF TO /AWS1/CL_SGMINFERENCECOMPONE00 /AWS1/CL_SGMINFERENCECOMPONE00

Details about the resources to deploy with this inference component, including the model, container, and compute resources.

IO_RUNTIMECONFIG TYPE REF TO /AWS1/CL_SGMINFERENCECOMPONE04 /AWS1/CL_SGMINFERENCECOMPONE04

Runtime settings for a model that is deployed with an inference component.

Optional arguments:

IT_TAGS TYPE /AWS1/CL_SGMTAG=>TT_TAGLIST TT_TAGLIST

A list of key-value pairs associated with the model. For more information, see Tagging Amazon Web Services resources in the Amazon Web Services General Reference.

RETURNING

OO_OUTPUT TYPE REF TO /AWS1/CL_SGMCREINFERENCECOMP01 /AWS1/CL_SGMCREINFERENCECOMP01