[ aws . iotsitewise ]

create-bulk-import-job

Description

Note

This API operation is in preview release for IoT SiteWise and is subject to change. We recommend that you use this operation only with test data, and not in production environments.

Defines a job to ingest data to IoT SiteWise from Amazon S3. For more information, see Create a bulk import job (CLI) in the Amazon Simple Storage Service User Guide .

Warning

You must enable IoT SiteWise to export data to Amazon S3 before you create a bulk import job. For more information about how to configure storage settings, see PutStorageConfiguration .

See also: AWS API Documentation

See ‘aws help’ for descriptions of global parameters.

Synopsis

  create-bulk-import-job
--job-name <value>
--job-role-arn <value>
--files <value>
--error-report-location <value>
--job-configuration <value>
[--cli-input-json | --cli-input-yaml]
[--generate-cli-skeleton <value>]

Options

--job-name (string)

The unique name that helps identify the job request.

--job-role-arn (string)

The ARN of the IAM role that allows IoT SiteWise to read Amazon S3 data.

--files (list)

The files in the specified Amazon S3 bucket that contain your data.

(structure)

The file in Amazon S3 where your data is saved.

bucket -> (string)

The name of the Amazon S3 bucket from which data is imported.

key -> (string)

The key of the Amazon S3 object that contains your data. Each object has a key that is a unique identifier. Each object has exactly one key.

versionId -> (string)

The version ID to identify a specific version of the Amazon S3 object that contains your data.

Shorthand Syntax:

bucket=string,key=string,versionId=string ...

JSON Syntax:

[
  {
    "bucket": "string",
    "key": "string",
    "versionId": "string"
  }
  ...
]

--error-report-location (structure)

The Amazon S3 destination where errors associated with the job creation request are saved.

bucket -> (string)

The name of the Amazon S3 bucket to which errors associated with the bulk import job are sent.

prefix -> (string)

Amazon S3 uses the prefix as a folder name to organize data in the bucket. Each Amazon S3 object has a key that is its unique identifier in the bucket. Each object in a bucket has exactly one key. The prefix must end with a forward slash (/). For more information, see Organizing objects using prefixes in the Amazon Simple Storage Service User Guide .

Shorthand Syntax:

bucket=string,prefix=string

JSON Syntax:

{
  "bucket": "string",
  "prefix": "string"
}

--job-configuration (structure)

Contains the configuration information of a job, such as the file format used to save data in Amazon S3.

fileFormat -> (structure)

The file format of the data in Amazon S3.

csv -> (structure)

The .csv file format.

columnNames -> (list)

The column names specified in the .csv file.

(string)

JSON Syntax:

{
  "fileFormat": {
    "csv": {
      "columnNames": ["ALIAS"|"ASSET_ID"|"PROPERTY_ID"|"DATA_TYPE"|"TIMESTAMP_SECONDS"|"TIMESTAMP_NANO_OFFSET"|"QUALITY"|"VALUE", ...]
    }
  }
}

--cli-input-json | --cli-input-yaml (string) Reads arguments from the JSON string provided. The JSON string follows the format provided by --generate-cli-skeleton. If other arguments are provided on the command line, those values will override the JSON-provided values. It is not possible to pass arbitrary binary values using a JSON-provided value as the string will be taken literally. This may not be specified along with --cli-input-yaml.

--generate-cli-skeleton (string) Prints a JSON skeleton to standard output without sending an API request. If provided with no value or the value input, prints a sample input JSON that can be used as an argument for --cli-input-json. Similarly, if provided yaml-input it will print a sample input YAML that can be used with --cli-input-yaml. If provided with the value output, it validates the command inputs and returns a sample output JSON for that command. The generated JSON skeleton is not stable between versions of the AWS CLI and there are no backwards compatibility guarantees in the JSON skeleton generated.

See ‘aws help’ for descriptions of global parameters.

Output

jobId -> (string)

The ID of the job.

jobName -> (string)

The unique name that helps identify the job request.

jobStatus -> (string)

The status of the bulk import job can be one of following values.

  • PENDING – IoT SiteWise is waiting for the current bulk import job to finish.

  • CANCELLED – The bulk import job has been canceled.

  • RUNNING – IoT SiteWise is processing your request to import your data from Amazon S3.

  • COMPLETED – IoT SiteWise successfully completed your request to import data from Amazon S3.

  • FAILED – IoT SiteWise couldn’t process your request to import data from Amazon S3. You can use logs saved in the specified error report location in Amazon S3 to troubleshoot issues.

  • COMPLETED_WITH_FAILURES – IoT SiteWise completed your request to import data from Amazon S3 with errors. You can use logs saved in the specified error report location in Amazon S3 to troubleshoot issues.