Skip to main content
Version: Latest-4.0

loads

loads provides the results of load jobs. This view is supported from StarRocks v3.1 onwards.

The following fields are provided in loads:

FieldDescription
IDGlobally unique identifier.
LABELLabel of the load job.
PROFILE_IDThe ID of the Profile, which can be analyzed via ANALYZE PROFILE.
DB_NAMEThe database to which the target table belongs.
TABLE_NAMEThe target table.
USERThe user who initiates the load job.
WAREHOUSEThe warehouse to which the load job belongs.
STATEThe state of the load job. Valid values:
  • PENDING/BEGIN: The load job is created.
  • QUEUEING/BEFORE_LOAD: The load job is in the queue waiting to be scheduled.
  • LOADING: The load job is running.
  • PREPARING: The transaction is being pre-committed.
  • PREPARED: The transaction has been pre-committed.
  • COMMITED: The transaction has been committed.
  • FINISHED: The load job succeeded.
  • CANCELLED: The load job failed.
PROGRESSThe progress of the ETL stage and LOADING stage of the load job.
TYPEThe type of the load job. For Broker Load, the return value is BROKER. For INSERT, the return value is INSERT. For Stream Load, the return value is STREAM. For Routine Load Load, the return value is ROUTINE.
PRIORITYThe priority of the load job. Valid values: HIGHEST, HIGH, NORMAL, LOW, and LOWEST.
SCAN_ROWSThe number of data rows that are scanned.
SCAN_BYTESThe number of bytes that are scanned.
FILTERED_ROWSThe number of data rows that are filtered out due to inadequate data quality.
UNSELECTED_ROWSThe number of data rows that are filtered out due to the conditions specified in the WHERE clause.
SINK_ROWSThe number of data rows that are loaded.
RUNTIME_DETAILSLoad runtime metadata. For details, see RUNTIME_DETAILS.
CREATE_TIMEThe time at which the load job was created. Format: yyyy-MM-dd HH:mm:ss. Example: 2023-07-24 14:58:58.
LOAD_START_TIMEThe start time of the LOADING stage of the load job. Format: yyyy-MM-dd HH:mm:ss. Example: 2023-07-24 14:58:58.
LOAD_COMMIT_TIMEThe time at which the loading transaction was committed. Format: yyyy-MM-dd HH:mm:ss. Example: 2023-07-24 14:58:58.
LOAD_FINISH_TIMEThe end time of the LOADING stage of the load job. Format: yyyy-MM-dd HH:mm:ss. Example: 2023-07-24 14:58:58.
PROPERTIESThe static properties of the load job. For details, see PROPERTIES.
ERROR_MSGThe error message of the load job. If the load job did not encounter any error, NULL is returned.
TRACKING_SQLThe SQL statement that can be used to query the tracking log of the load job. A SQL statement is returned only when the load job involves unqualified data rows. If the load job does not involve any unqualified data rows, NULL is returned.
REJECTED_RECORD_PATHThe path from which you can access all the unqualified data rows that are filtered out in the load job. The number of unqualified data rows logged is determined by the log_rejected_record_num parameter configured in the load job. You can use the wget command to access the path. If the load job does not involve any unqualified data rows, NULL is returned.

RUNTIME_DETAILS​

  • Universal metrics:
MetricDescription
load_idGlobally unique ID of the load execution plan.
txn_idLoad transaction ID.
  • Specific metrics for Broker Load, INSERT INTO, and Spark Load:
MetricDescription
etl_infoETL Details. This field is only valid for Spark Load jobs. For other types of load jobs, the value will be empty.
etl_start_timeThe start time of the ETL stage of the load job. Format: yyyy-MM-dd HH:mm:ss. Example: 2023-07-24 14:58:58.
etl_start_timeThe end time of the ETL stage of the load job. Format: yyyy-MM-dd HH:mm:ss. Example: 2023-07-24 14:58:58.
unfinished_backendsList of BEs with unfinished executions.
backendsList of BEs participating in execution.
file_numNumber of files read.
file_sizeTotal size of files read.
task_numNumber of subtasks.
  • Specific metrics for Routine Load:
MetricDescription
schedule_intervalThe interval for Routine Load to be scheduled.
wait_slot_timeTime elapsed while the Routine Load task waits for execution slots.
check_offset_timeTime consumed when checking offset information during Routine Load task scheduling.
consume_timeTime consumed by the Routine Load task to read upstream data.
plan_timeTime for generating the execution plan.
commit_publish_timeTime consumed to execute the COMMIT RPC.
  • Specific metrics for Stream Load:
MetricDescription
timeoutTimeout for load tasks.
begin_txn_msTime consumed to begin the transaction.
plan_time_msTime for generating the execution plan.
receive_data_time_msTime for receiving data.
commit_publish_time_msTime consumed to execute the COMMIT RPC.
client_ipClient IP address.

PROPERTIES​

  • Specific properties for Broker Load, INSERT INTO, and Spark Load:
PropertyDescription
timeoutTimeout for load tasks.
max_filter_ratioMaximum ratio of data rows that are filtered out due to inadequate data quality.
  • Specific properties for Routine Load:
PropertyDescription
job_nameRoutine Load job name.
task_numNumber of subtasks actually executed in parallel.
timeoutTimeout for load tasks.