/* * Copyright 2018-2023 Amazon.com, Inc. or its affiliates. All Rights Reserved. * * Licensed under the Apache License, Version 2.0 (the "License"). You may not use this file except in compliance with * the License. A copy of the License is located at * * http://aws.amazon.com/apache2.0 * * or in the "license" file accompanying this file. This file is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR * CONDITIONS OF ANY KIND, either express or implied. See the License for the specific language governing permissions * and limitations under the License. */ package com.amazonaws.services.sagemaker.model; import java.io.Serializable; import javax.annotation.Generated; import com.amazonaws.protocol.StructuredPojo; import com.amazonaws.protocol.ProtocolMarshaller; /** *
* Describes the results of a transform job. *
* * @see AWS API * Documentation */ @Generated("com.amazonaws:aws-java-sdk-code-generator") public class TransformOutput implements Serializable, Cloneable, StructuredPojo { /** *
     * The Amazon S3 path where you want Amazon SageMaker to store the results of the transform job. For example,
     * s3://bucket-name/key-name-prefix.
     * 
     * For every S3 object used as input for the transform job, batch transform stores the transformed data with an .
     * out suffix in a corresponding subfolder in the location in the output prefix. For example, for the
     * input data stored at s3://bucket-name/input-name-prefix/dataset01/data.csv, batch transform stores
     * the transformed data at s3://bucket-name/output-name-prefix/input-name-prefix/data.csv.out. Batch
     * transform doesn't upload partially processed objects. For an input S3 object that contains multiple records, it
     * creates an .out file only if the transform job succeeds on the entire file. When the input contains
     * multiple S3 objects, the batch transform job processes the listed S3 objects and uploads only the output for
     * successfully processed objects. If any object fails in the transform job batch transform marks the job as failed
     * to prompt investigation.
     * 
* The MIME type used to specify the output data. Amazon SageMaker uses the MIME type with each http call to * transfer data from the transform job. *
*/ private String accept; /** *
     * Defines how to assemble the results of the transform job as a single S3 object. Choose a format that is most
     * convenient to you. To concatenate the results in binary format, specify None. To add a newline
     * character at the end of every transformed record, specify Line.
     * 
     * The Amazon Web Services Key Management Service (Amazon Web Services KMS) key that Amazon SageMaker uses to
     * encrypt the model artifacts at rest using Amazon S3 server-side encryption. The KmsKeyId can be any
     * of the following formats:
     * 
     * Key ID: 1234abcd-12ab-34cd-56ef-1234567890ab
     * 
     * Key ARN: arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab
     * 
     * Alias name: alias/ExampleAlias
     * 
     * Alias name ARN: arn:aws:kms:us-west-2:111122223333:alias/ExampleAlias
     * 
* If you don't provide a KMS key ID, Amazon SageMaker uses the default KMS key for Amazon S3 for your role's * account. For more information, see KMS-Managed Encryption Keys in * the Amazon Simple Storage Service Developer Guide. *
** The KMS key policy must grant permission to the IAM role that you specify in your CreateModel request. * For more information, see Using * Key Policies in Amazon Web Services KMS in the Amazon Web Services Key Management Service Developer * Guide. *
*/ private String kmsKeyId; /** *
     * The Amazon S3 path where you want Amazon SageMaker to store the results of the transform job. For example,
     * s3://bucket-name/key-name-prefix.
     * 
     * For every S3 object used as input for the transform job, batch transform stores the transformed data with an .
     * out suffix in a corresponding subfolder in the location in the output prefix. For example, for the
     * input data stored at s3://bucket-name/input-name-prefix/dataset01/data.csv, batch transform stores
     * the transformed data at s3://bucket-name/output-name-prefix/input-name-prefix/data.csv.out. Batch
     * transform doesn't upload partially processed objects. For an input S3 object that contains multiple records, it
     * creates an .out file only if the transform job succeeds on the entire file. When the input contains
     * multiple S3 objects, the batch transform job processes the listed S3 objects and uploads only the output for
     * successfully processed objects. If any object fails in the transform job batch transform marks the job as failed
     * to prompt investigation.
     * 
s3://bucket-name/key-name-prefix.
     *        
     *        For every S3 object used as input for the transform job, batch transform stores the transformed data with
     *        an .out suffix in a corresponding subfolder in the location in the output prefix. For
     *        example, for the input data stored at s3://bucket-name/input-name-prefix/dataset01/data.csv,
     *        batch transform stores the transformed data at
     *        s3://bucket-name/output-name-prefix/input-name-prefix/data.csv.out. Batch transform doesn't
     *        upload partially processed objects. For an input S3 object that contains multiple records, it creates an .
     *        out file only if the transform job succeeds on the entire file. When the input contains
     *        multiple S3 objects, the batch transform job processes the listed S3 objects and uploads only the output
     *        for successfully processed objects. If any object fails in the transform job batch transform marks the job
     *        as failed to prompt investigation.
     */
    public void setS3OutputPath(String s3OutputPath) {
        this.s3OutputPath = s3OutputPath;
    }
    /**
     * 
     * The Amazon S3 path where you want Amazon SageMaker to store the results of the transform job. For example,
     * s3://bucket-name/key-name-prefix.
     * 
     * For every S3 object used as input for the transform job, batch transform stores the transformed data with an .
     * out suffix in a corresponding subfolder in the location in the output prefix. For example, for the
     * input data stored at s3://bucket-name/input-name-prefix/dataset01/data.csv, batch transform stores
     * the transformed data at s3://bucket-name/output-name-prefix/input-name-prefix/data.csv.out. Batch
     * transform doesn't upload partially processed objects. For an input S3 object that contains multiple records, it
     * creates an .out file only if the transform job succeeds on the entire file. When the input contains
     * multiple S3 objects, the batch transform job processes the listed S3 objects and uploads only the output for
     * successfully processed objects. If any object fails in the transform job batch transform marks the job as failed
     * to prompt investigation.
     * 
s3://bucket-name/key-name-prefix.
     *         
     *         For every S3 object used as input for the transform job, batch transform stores the transformed data with
     *         an .out suffix in a corresponding subfolder in the location in the output prefix. For
     *         example, for the input data stored at s3://bucket-name/input-name-prefix/dataset01/data.csv,
     *         batch transform stores the transformed data at
     *         s3://bucket-name/output-name-prefix/input-name-prefix/data.csv.out. Batch transform doesn't
     *         upload partially processed objects. For an input S3 object that contains multiple records, it creates an
     *         .out file only if the transform job succeeds on the entire file. When the input contains
     *         multiple S3 objects, the batch transform job processes the listed S3 objects and uploads only the output
     *         for successfully processed objects. If any object fails in the transform job batch transform marks the
     *         job as failed to prompt investigation.
     */
    public String getS3OutputPath() {
        return this.s3OutputPath;
    }
    /**
     * 
     * The Amazon S3 path where you want Amazon SageMaker to store the results of the transform job. For example,
     * s3://bucket-name/key-name-prefix.
     * 
     * For every S3 object used as input for the transform job, batch transform stores the transformed data with an .
     * out suffix in a corresponding subfolder in the location in the output prefix. For example, for the
     * input data stored at s3://bucket-name/input-name-prefix/dataset01/data.csv, batch transform stores
     * the transformed data at s3://bucket-name/output-name-prefix/input-name-prefix/data.csv.out. Batch
     * transform doesn't upload partially processed objects. For an input S3 object that contains multiple records, it
     * creates an .out file only if the transform job succeeds on the entire file. When the input contains
     * multiple S3 objects, the batch transform job processes the listed S3 objects and uploads only the output for
     * successfully processed objects. If any object fails in the transform job batch transform marks the job as failed
     * to prompt investigation.
     * 
s3://bucket-name/key-name-prefix.
     *        
     *        For every S3 object used as input for the transform job, batch transform stores the transformed data with
     *        an .out suffix in a corresponding subfolder in the location in the output prefix. For
     *        example, for the input data stored at s3://bucket-name/input-name-prefix/dataset01/data.csv,
     *        batch transform stores the transformed data at
     *        s3://bucket-name/output-name-prefix/input-name-prefix/data.csv.out. Batch transform doesn't
     *        upload partially processed objects. For an input S3 object that contains multiple records, it creates an .
     *        out file only if the transform job succeeds on the entire file. When the input contains
     *        multiple S3 objects, the batch transform job processes the listed S3 objects and uploads only the output
     *        for successfully processed objects. If any object fails in the transform job batch transform marks the job
     *        as failed to prompt investigation.
     * @return Returns a reference to this object so that method calls can be chained together.
     */
    public TransformOutput withS3OutputPath(String s3OutputPath) {
        setS3OutputPath(s3OutputPath);
        return this;
    }
    /**
     * 
* The MIME type used to specify the output data. Amazon SageMaker uses the MIME type with each http call to * transfer data from the transform job. *
* * @param accept * The MIME type used to specify the output data. Amazon SageMaker uses the MIME type with each http call to * transfer data from the transform job. */ public void setAccept(String accept) { this.accept = accept; } /** ** The MIME type used to specify the output data. Amazon SageMaker uses the MIME type with each http call to * transfer data from the transform job. *
* * @return The MIME type used to specify the output data. Amazon SageMaker uses the MIME type with each http call to * transfer data from the transform job. */ public String getAccept() { return this.accept; } /** ** The MIME type used to specify the output data. Amazon SageMaker uses the MIME type with each http call to * transfer data from the transform job. *
* * @param accept * The MIME type used to specify the output data. Amazon SageMaker uses the MIME type with each http call to * transfer data from the transform job. * @return Returns a reference to this object so that method calls can be chained together. */ public TransformOutput withAccept(String accept) { setAccept(accept); return this; } /** *
     * Defines how to assemble the results of the transform job as a single S3 object. Choose a format that is most
     * convenient to you. To concatenate the results in binary format, specify None. To add a newline
     * character at the end of every transformed record, specify Line.
     * 
None. To add a
     *        newline character at the end of every transformed record, specify Line.
     * @see AssemblyType
     */
    public void setAssembleWith(String assembleWith) {
        this.assembleWith = assembleWith;
    }
    /**
     * 
     * Defines how to assemble the results of the transform job as a single S3 object. Choose a format that is most
     * convenient to you. To concatenate the results in binary format, specify None. To add a newline
     * character at the end of every transformed record, specify Line.
     * 
None. To add a
     *         newline character at the end of every transformed record, specify Line.
     * @see AssemblyType
     */
    public String getAssembleWith() {
        return this.assembleWith;
    }
    /**
     * 
     * Defines how to assemble the results of the transform job as a single S3 object. Choose a format that is most
     * convenient to you. To concatenate the results in binary format, specify None. To add a newline
     * character at the end of every transformed record, specify Line.
     * 
None. To add a
     *        newline character at the end of every transformed record, specify Line.
     * @return Returns a reference to this object so that method calls can be chained together.
     * @see AssemblyType
     */
    public TransformOutput withAssembleWith(String assembleWith) {
        setAssembleWith(assembleWith);
        return this;
    }
    /**
     * 
     * Defines how to assemble the results of the transform job as a single S3 object. Choose a format that is most
     * convenient to you. To concatenate the results in binary format, specify None. To add a newline
     * character at the end of every transformed record, specify Line.
     * 
None. To add a
     *        newline character at the end of every transformed record, specify Line.
     * @return Returns a reference to this object so that method calls can be chained together.
     * @see AssemblyType
     */
    public TransformOutput withAssembleWith(AssemblyType assembleWith) {
        this.assembleWith = assembleWith.toString();
        return this;
    }
    /**
     * 
     * The Amazon Web Services Key Management Service (Amazon Web Services KMS) key that Amazon SageMaker uses to
     * encrypt the model artifacts at rest using Amazon S3 server-side encryption. The KmsKeyId can be any
     * of the following formats:
     * 
     * Key ID: 1234abcd-12ab-34cd-56ef-1234567890ab
     * 
     * Key ARN: arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab
     * 
     * Alias name: alias/ExampleAlias
     * 
     * Alias name ARN: arn:aws:kms:us-west-2:111122223333:alias/ExampleAlias
     * 
* If you don't provide a KMS key ID, Amazon SageMaker uses the default KMS key for Amazon S3 for your role's * account. For more information, see KMS-Managed Encryption Keys in * the Amazon Simple Storage Service Developer Guide. *
** The KMS key policy must grant permission to the IAM role that you specify in your CreateModel request. * For more information, see Using * Key Policies in Amazon Web Services KMS in the Amazon Web Services Key Management Service Developer * Guide. *
* * @param kmsKeyId * The Amazon Web Services Key Management Service (Amazon Web Services KMS) key that Amazon SageMaker uses to * encrypt the model artifacts at rest using Amazon S3 server-side encryption. TheKmsKeyId can
     *        be any of the following formats: 
     *        
     *        Key ID: 1234abcd-12ab-34cd-56ef-1234567890ab
     *        
     *        Key ARN: arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab
     *        
     *        Alias name: alias/ExampleAlias
     *        
     *        Alias name ARN: arn:aws:kms:us-west-2:111122223333:alias/ExampleAlias
     *        
* If you don't provide a KMS key ID, Amazon SageMaker uses the default KMS key for Amazon S3 for your role's * account. For more information, see KMS-Managed Encryption * Keys in the Amazon Simple Storage Service Developer Guide. *
** The KMS key policy must grant permission to the IAM role that you specify in your CreateModel * request. For more information, see Using Key Policies in * Amazon Web Services KMS in the Amazon Web Services Key Management Service Developer Guide. */ public void setKmsKeyId(String kmsKeyId) { this.kmsKeyId = kmsKeyId; } /** *
     * The Amazon Web Services Key Management Service (Amazon Web Services KMS) key that Amazon SageMaker uses to
     * encrypt the model artifacts at rest using Amazon S3 server-side encryption. The KmsKeyId can be any
     * of the following formats:
     * 
     * Key ID: 1234abcd-12ab-34cd-56ef-1234567890ab
     * 
     * Key ARN: arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab
     * 
     * Alias name: alias/ExampleAlias
     * 
     * Alias name ARN: arn:aws:kms:us-west-2:111122223333:alias/ExampleAlias
     * 
* If you don't provide a KMS key ID, Amazon SageMaker uses the default KMS key for Amazon S3 for your role's * account. For more information, see KMS-Managed Encryption Keys in * the Amazon Simple Storage Service Developer Guide. *
** The KMS key policy must grant permission to the IAM role that you specify in your CreateModel request. * For more information, see Using * Key Policies in Amazon Web Services KMS in the Amazon Web Services Key Management Service Developer * Guide. *
* * @return The Amazon Web Services Key Management Service (Amazon Web Services KMS) key that Amazon SageMaker uses * to encrypt the model artifacts at rest using Amazon S3 server-side encryption. TheKmsKeyId
     *         can be any of the following formats: 
     *         
     *         Key ID: 1234abcd-12ab-34cd-56ef-1234567890ab
     *         
     *         Key ARN: arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab
     *         
     *         Alias name: alias/ExampleAlias
     *         
     *         Alias name ARN: arn:aws:kms:us-west-2:111122223333:alias/ExampleAlias
     *         
* If you don't provide a KMS key ID, Amazon SageMaker uses the default KMS key for Amazon S3 for your * role's account. For more information, see KMS-Managed Encryption * Keys in the Amazon Simple Storage Service Developer Guide. *
** The KMS key policy must grant permission to the IAM role that you specify in your CreateModel * request. For more information, see Using Key Policies in * Amazon Web Services KMS in the Amazon Web Services Key Management Service Developer Guide. */ public String getKmsKeyId() { return this.kmsKeyId; } /** *
     * The Amazon Web Services Key Management Service (Amazon Web Services KMS) key that Amazon SageMaker uses to
     * encrypt the model artifacts at rest using Amazon S3 server-side encryption. The KmsKeyId can be any
     * of the following formats:
     * 
     * Key ID: 1234abcd-12ab-34cd-56ef-1234567890ab
     * 
     * Key ARN: arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab
     * 
     * Alias name: alias/ExampleAlias
     * 
     * Alias name ARN: arn:aws:kms:us-west-2:111122223333:alias/ExampleAlias
     * 
* If you don't provide a KMS key ID, Amazon SageMaker uses the default KMS key for Amazon S3 for your role's * account. For more information, see KMS-Managed Encryption Keys in * the Amazon Simple Storage Service Developer Guide. *
** The KMS key policy must grant permission to the IAM role that you specify in your CreateModel request. * For more information, see Using * Key Policies in Amazon Web Services KMS in the Amazon Web Services Key Management Service Developer * Guide. *
* * @param kmsKeyId * The Amazon Web Services Key Management Service (Amazon Web Services KMS) key that Amazon SageMaker uses to * encrypt the model artifacts at rest using Amazon S3 server-side encryption. TheKmsKeyId can
     *        be any of the following formats: 
     *        
     *        Key ID: 1234abcd-12ab-34cd-56ef-1234567890ab
     *        
     *        Key ARN: arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab
     *        
     *        Alias name: alias/ExampleAlias
     *        
     *        Alias name ARN: arn:aws:kms:us-west-2:111122223333:alias/ExampleAlias
     *        
* If you don't provide a KMS key ID, Amazon SageMaker uses the default KMS key for Amazon S3 for your role's * account. For more information, see KMS-Managed Encryption * Keys in the Amazon Simple Storage Service Developer Guide. *
** The KMS key policy must grant permission to the IAM role that you specify in your CreateModel * request. For more information, see Using Key Policies in * Amazon Web Services KMS in the Amazon Web Services Key Management Service Developer Guide. * @return Returns a reference to this object so that method calls can be chained together. */ public TransformOutput withKmsKeyId(String kmsKeyId) { setKmsKeyId(kmsKeyId); return this; } /** * Returns a string representation of this object. This is useful for testing and debugging. Sensitive data will be * redacted from this string using a placeholder value. * * @return A string representation of this object. * * @see java.lang.Object#toString() */ @Override public String toString() { StringBuilder sb = new StringBuilder(); sb.append("{"); if (getS3OutputPath() != null) sb.append("S3OutputPath: ").append(getS3OutputPath()).append(","); if (getAccept() != null) sb.append("Accept: ").append(getAccept()).append(","); if (getAssembleWith() != null) sb.append("AssembleWith: ").append(getAssembleWith()).append(","); if (getKmsKeyId() != null) sb.append("KmsKeyId: ").append(getKmsKeyId()); sb.append("}"); return sb.toString(); } @Override public boolean equals(Object obj) { if (this == obj) return true; if (obj == null) return false; if (obj instanceof TransformOutput == false) return false; TransformOutput other = (TransformOutput) obj; if (other.getS3OutputPath() == null ^ this.getS3OutputPath() == null) return false; if (other.getS3OutputPath() != null && other.getS3OutputPath().equals(this.getS3OutputPath()) == false) return false; if (other.getAccept() == null ^ this.getAccept() == null) return false; if (other.getAccept() != null && other.getAccept().equals(this.getAccept()) == false) return false; if (other.getAssembleWith() == null ^ this.getAssembleWith() == null) return false; if (other.getAssembleWith() != null && other.getAssembleWith().equals(this.getAssembleWith()) == false) return false; if (other.getKmsKeyId() == null ^ this.getKmsKeyId() == null) return false; if (other.getKmsKeyId() != null && other.getKmsKeyId().equals(this.getKmsKeyId()) == false) return false; return true; } @Override public int hashCode() { final int prime = 31; int hashCode = 1; hashCode = prime * hashCode + ((getS3OutputPath() == null) ? 0 : getS3OutputPath().hashCode()); hashCode = prime * hashCode + ((getAccept() == null) ? 0 : getAccept().hashCode()); hashCode = prime * hashCode + ((getAssembleWith() == null) ? 0 : getAssembleWith().hashCode()); hashCode = prime * hashCode + ((getKmsKeyId() == null) ? 0 : getKmsKeyId().hashCode()); return hashCode; } @Override public TransformOutput clone() { try { return (TransformOutput) super.clone(); } catch (CloneNotSupportedException e) { throw new IllegalStateException("Got a CloneNotSupportedException from Object.clone() " + "even though we're Cloneable!", e); } } @com.amazonaws.annotation.SdkInternalApi @Override public void marshall(ProtocolMarshaller protocolMarshaller) { com.amazonaws.services.sagemaker.model.transform.TransformOutputMarshaller.getInstance().marshall(this, protocolMarshaller); } }