ingest command submits a new data ingestion request to upload data from an S3 bucket.
Options
Options
--delimiterstringoptional: Character used to separate fields in the input file. Default:,.--domainstringoptional: DataHub logical category used to organize the ingestion job.--file-idstringoptional: Unique ID of the uploaded file.--file-pathstringoptional: S3 URL to the file you want to ingest.--formatstringoptional: Format of the file. Values:csv,orc,xml,parquet, orxls.--headerbooleanoptional: Indicates whether the file contains a header row.--merge-keysstringoptional: Comma-separated columns used to match and merge data when themodeismerge.--modestringoptional: How to write data to the table. Values:append,overwrite, ormerge.--namestringrequired: Name of the Airflow ingestion job.--schemastringrequired: Target schema name for the data.--schedulestringoptional: Cron expression that specifies when or how often the ingestion runs.--tablestringrequired: Target table name for the data.--tagsstringoptional: Comma-separated DataHub tags.--typestringoptional: Type of data ingestion. Values:file,jdbc,lakehouse.
Global options
Global options
--api-keystringoptional: API key or theNX1_API_KEYvariable.--config-filestringoptional: YAML config file.--help-hoptional: Shows help messages and exit.--hoststringoptional: API host.--no-verify-ssloptional: Disables SSL verification.--output-ooptional: Output format. Allowed values:json,yaml, ortable.--profile-pstringoptional: Profile name from~/.nx1/profiles.--timeoutintegeroptional: Requests when to timeout. Default is30.--verbose-voptional: Verbose output.