InferenceComponentDataCacheConfig
Settings that affect how the inference component caches data.
Contents
- EnableCaching
-
Sets whether the endpoint that hosts the inference component caches the model artifacts and container image.
With caching enabled, the endpoint caches this data in each instance that it provisions for the inference component. That way, the inference component deploys faster during the auto scaling process. If caching isn't enabled, the inference component takes longer to deploy because of the time it spends downloading the data.
Type: Boolean
Required: Yes
See Also
For more information about using this API in one of the language-specific Amazon SDKs, see the following: