Bedrock / Paginator / ListInferenceProfiles

ListInferenceProfiles#

class Bedrock.Paginator.ListInferenceProfiles#
paginator = client.get_paginator('list_inference_profiles')
paginate(**kwargs)#

Creates an iterator that will paginate through responses from Bedrock.Client.list_inference_profiles().

See also: AWS API Documentation

Request Syntax

response_iterator = paginator.paginate(
    typeEquals='SYSTEM_DEFINED'|'APPLICATION',
    PaginationConfig={
        'MaxItems': 123,
        'PageSize': 123,
        'StartingToken': 'string'
    }
)
Parameters:
  • typeEquals (string) –

    Filters for inference profiles that match the type you specify.

    • SYSTEM_DEFINED – The inference profile is defined by Amazon Bedrock. You can route inference requests across regions with these inference profiles.

    • APPLICATION – The inference profile was created by a user. This type of inference profile can track metrics and costs when invoking the model in it. The inference profile may route requests to one or multiple regions.

  • PaginationConfig (dict) –

    A dictionary that provides parameters to control pagination.

    • MaxItems (integer) –

      The total number of items to return. If the total number of items available is more than the value specified in max-items then a NextToken will be provided in the output that you can use to resume pagination.

    • PageSize (integer) –

      The size of each page.

    • StartingToken (string) –

      A token to specify where to start paginating. This is the NextToken from a previous response.

Return type:

dict

Returns:

Response Syntax

{
    'inferenceProfileSummaries': [
        {
            'inferenceProfileName': 'string',
            'description': 'string',
            'createdAt': datetime(2015, 1, 1),
            'updatedAt': datetime(2015, 1, 1),
            'inferenceProfileArn': 'string',
            'models': [
                {
                    'modelArn': 'string'
                },
            ],
            'inferenceProfileId': 'string',
            'status': 'ACTIVE',
            'type': 'SYSTEM_DEFINED'|'APPLICATION'
        },
    ],
    'NextToken': 'string'
}

Response Structure

  • (dict) –

    • inferenceProfileSummaries (list) –

      A list of information about each inference profile that you can use.

      • (dict) –

        Contains information about an inference profile.

        • inferenceProfileName (string) –

          The name of the inference profile.

        • description (string) –

          The description of the inference profile.

        • createdAt (datetime) –

          The time at which the inference profile was created.

        • updatedAt (datetime) –

          The time at which the inference profile was last updated.

        • inferenceProfileArn (string) –

          The Amazon Resource Name (ARN) of the inference profile.

        • models (list) –

          A list of information about each model in the inference profile.

          • (dict) –

            Contains information about a model.

            • modelArn (string) –

              The Amazon Resource Name (ARN) of the model.

        • inferenceProfileId (string) –

          The unique identifier of the inference profile.

        • status (string) –

          The status of the inference profile. ACTIVE means that the inference profile is ready to be used.

        • type (string) –

          The type of the inference profile. The following types are possible:

          • SYSTEM_DEFINED – The inference profile is defined by Amazon Bedrock. You can route inference requests across regions with these inference profiles.

          • APPLICATION – The inference profile was created by a user. This type of inference profile can track metrics and costs when invoking the model in it. The inference profile may route requests to one or multiple regions.

    • NextToken (string) –

      A token to resume pagination.