Loading options
Data loading is the process of cleansing and transforming raw data from various data sources based on your business requirements and loading the resulting data into StarRocks to facilitate analysis.
StarRocks provides a variety of options for data loading:
- Loading methods: Insert, Stream Load, Broker Load, Pipe, Routine Load, and Spark Load
- Ecosystem tools: StarRocks Connector for Apache Kafka® (Kafka connector for short), StarRocks Connector for Apache Spark™ (Spark connector for short), StarRocks Connector for Apache Flink® (Flink connector for short), and other tools such as SMT, DataX, CloudCanal, and Kettle Connector
- API: Stream Load transaction interface
These options each have its own advantages and support its own set of data source systems to pull from.
This topic provides an overview of these options, along with comparisons between them to help you determine the loading option of your choice based on your data source, business scenario, data volume, data file format, and loading frequency.
Introduction to loading options
This section mainly describes the characteristics and business scenarios of the loading options available in StarRocks.

In the following sections, "batch" or "batch loading" refers to the loading of a large amount of data from a specified source all at a time into StarRocks, whereas "stream" or "streaming" refers to the continuous loading of data in real time.