/** * Copyright Amazon.com, Inc. or its affiliates. All Rights Reserved. * SPDX-License-Identifier: Apache-2.0. */ #pragma once #include #include #include #include #include namespace Aws { template class AmazonWebServiceResult; namespace Utils { namespace Json { class JsonValue; } // namespace Json } // namespace Utils namespace LookoutEquipment { namespace Model { class ListInferenceExecutionsResult { public: AWS_LOOKOUTEQUIPMENT_API ListInferenceExecutionsResult(); AWS_LOOKOUTEQUIPMENT_API ListInferenceExecutionsResult(const Aws::AmazonWebServiceResult& result); AWS_LOOKOUTEQUIPMENT_API ListInferenceExecutionsResult& operator=(const Aws::AmazonWebServiceResult& result); /** *

An opaque pagination token indicating where to continue the listing of * inference executions.

*/ inline const Aws::String& GetNextToken() const{ return m_nextToken; } /** *

An opaque pagination token indicating where to continue the listing of * inference executions.

*/ inline void SetNextToken(const Aws::String& value) { m_nextToken = value; } /** *

An opaque pagination token indicating where to continue the listing of * inference executions.

*/ inline void SetNextToken(Aws::String&& value) { m_nextToken = std::move(value); } /** *

An opaque pagination token indicating where to continue the listing of * inference executions.

*/ inline void SetNextToken(const char* value) { m_nextToken.assign(value); } /** *

An opaque pagination token indicating where to continue the listing of * inference executions.

*/ inline ListInferenceExecutionsResult& WithNextToken(const Aws::String& value) { SetNextToken(value); return *this;} /** *

An opaque pagination token indicating where to continue the listing of * inference executions.

*/ inline ListInferenceExecutionsResult& WithNextToken(Aws::String&& value) { SetNextToken(std::move(value)); return *this;} /** *

An opaque pagination token indicating where to continue the listing of * inference executions.

*/ inline ListInferenceExecutionsResult& WithNextToken(const char* value) { SetNextToken(value); return *this;} /** *

Provides an array of information about the individual inference executions * returned from the ListInferenceExecutions operation, including * model used, inference scheduler, data configuration, and so on.

*/ inline const Aws::Vector& GetInferenceExecutionSummaries() const{ return m_inferenceExecutionSummaries; } /** *

Provides an array of information about the individual inference executions * returned from the ListInferenceExecutions operation, including * model used, inference scheduler, data configuration, and so on.

*/ inline void SetInferenceExecutionSummaries(const Aws::Vector& value) { m_inferenceExecutionSummaries = value; } /** *

Provides an array of information about the individual inference executions * returned from the ListInferenceExecutions operation, including * model used, inference scheduler, data configuration, and so on.

*/ inline void SetInferenceExecutionSummaries(Aws::Vector&& value) { m_inferenceExecutionSummaries = std::move(value); } /** *

Provides an array of information about the individual inference executions * returned from the ListInferenceExecutions operation, including * model used, inference scheduler, data configuration, and so on.

*/ inline ListInferenceExecutionsResult& WithInferenceExecutionSummaries(const Aws::Vector& value) { SetInferenceExecutionSummaries(value); return *this;} /** *

Provides an array of information about the individual inference executions * returned from the ListInferenceExecutions operation, including * model used, inference scheduler, data configuration, and so on.

*/ inline ListInferenceExecutionsResult& WithInferenceExecutionSummaries(Aws::Vector&& value) { SetInferenceExecutionSummaries(std::move(value)); return *this;} /** *

Provides an array of information about the individual inference executions * returned from the ListInferenceExecutions operation, including * model used, inference scheduler, data configuration, and so on.

*/ inline ListInferenceExecutionsResult& AddInferenceExecutionSummaries(const InferenceExecutionSummary& value) { m_inferenceExecutionSummaries.push_back(value); return *this; } /** *

Provides an array of information about the individual inference executions * returned from the ListInferenceExecutions operation, including * model used, inference scheduler, data configuration, and so on.

*/ inline ListInferenceExecutionsResult& AddInferenceExecutionSummaries(InferenceExecutionSummary&& value) { m_inferenceExecutionSummaries.push_back(std::move(value)); return *this; } inline const Aws::String& GetRequestId() const{ return m_requestId; } inline void SetRequestId(const Aws::String& value) { m_requestId = value; } inline void SetRequestId(Aws::String&& value) { m_requestId = std::move(value); } inline void SetRequestId(const char* value) { m_requestId.assign(value); } inline ListInferenceExecutionsResult& WithRequestId(const Aws::String& value) { SetRequestId(value); return *this;} inline ListInferenceExecutionsResult& WithRequestId(Aws::String&& value) { SetRequestId(std::move(value)); return *this;} inline ListInferenceExecutionsResult& WithRequestId(const char* value) { SetRequestId(value); return *this;} private: Aws::String m_nextToken; Aws::Vector m_inferenceExecutionSummaries; Aws::String m_requestId; }; } // namespace Model } // namespace LookoutEquipment } // namespace Aws