Install Chango Components
There are several components supported by Chango.
See What is Chango? for more details.
NOTE: If Chango initialization failed or you want to reset Chango, then you need to move to
http://[admin-host]:8123/cp-reset.html
to reset Chango.
Initialize Chango
If you have not initialized Chango, you will get the following picture to initialze Chango.
There are mandatory Chango Components like MySQL
, Object Storage
, Chango Authorizer
, and
Chango REST Catalog
which must be installed when Chango is initialized.
The other optional compoents can be installed after Chango initialization.
Configure Hosts and SSH Private Key
All the Chango Component Nodes
need to be registered, and SSH private key on Chango Admin Node
need to be added to access Chango Component Nodes
from
the host of Chango Admin Node
.
NOTE: Take a note that you need to register all the nodes of
Chango Component Nodes
exceptChango Admin Node
.
Get SSH private key on Chango Admin Node
with the following command, and paste it to the text box SSH Private Key
.
Configure LVM
Raw disks attached to the Chango Component Nodes
in comma separated list need to be added to mount as logical volume.
If the disks attached are same on all the nodes, use this.
If the disks attached are different on every node, use this.
Configure MySQL
MySQL
is used by open source components like Apache Superset
and Apache Ozone
in Chango.
Select the host on which MySQL
will be installed.
Configure Object Storage
Select the options for object storage. Apache Ozone
is default object storage provided by Chango,
which will be used in disconnected environment in most cases.
In public, you may select the external object storage like AWS S3, MinIO, OCI Object Storage.
If Apache Ozone is used as object storage, enter the values this below.
If external object storage is used, enter values of S3 credentials.
Configure Chango Authorizer
Chango Authorizer
is used to authenticate and authorize all the data access to Chango.
Configure Chango REST Catalog
Chango REST Catalog
is used as data catalog in Chango.
For now, the components you configured are mandatory. After configuring mandatory components, you can skip configuration.
You can install other optional components later after finishing Chango initialization.
Install Configured Components
Install all the configured components.
When the installation is finished, press the button of Installation Finished
.
Then, you will move to the main page.
Show Log
You can see current log produced by installed components.
Click the host name of components in Status
to show log.
Chango Authorizer
Scale Server
Chango Authorizer can be scaled out or unscaled.
First, select hosts for scaling out or unscaling Authorizer servers, and then, press the button of Scale Out Server
to scale out Authorizer servers
or press the button of Unscale Server
to unscale Authorizer servers.
Chango REST Catalog
Scale Server
Chango REST Catalog can be scaled out or unscaled.
First, select hosts for scaling out or unscaling REST Catalog servers, and then, press the button of Scale Out Server
to scale out REST Catalog servers
or press the button of Unscale Server
to unscale REST Catalog servers.
Apache Kafka
Apache Kafka
is used as event streaming platform in Chango.
Multiple Kafka clusters are supported by Chango, that is, you can install kafka clusters as many as you want.
Install Kafka
If you want to install Apache Kafka
, press Go to Install
button.
NOTE: If you have not installed any kafka cluster, then, enter
default
for the cluster name.
After installing kafka, you will see kafka page like this.
Because Chango supports multiple kafka clusters, you can install another kafka cluster.
NOTE: Because you have already installed default kafka cluster, you can enter anything for the cluster name.
After installing another kafka cluster, new created kafka cluster will be shown in tab list.
Scale Broker
Kafka Broker can be scaled out or unscaled.
First, select hosts for scaling out or unscaling kafka brokers, and then, press the button of Scale Out Broker
to scale out brokers
or press the button of Unscale Broker
to unscale brokers.
Configure Kafka
You can update kafka configurations like heap memory and server.properties
.
First, select kafka cluster which you want to configure.
After modifying configuration, press Update
to update the selected kafka cluster.
Apache Spark
Apache Spark
is used as computing engine to run batch and streaming jobs in Chango.
Install Spark
Select hosts to install master, worker and history server of Spark.
After installing Spark, you will see the spark page like this.
Scale Worker
You can scale out and unscale spark workers.
UI
There are URL links to get Spark Master UI and Spark History Server UI in UI
of Spark Page.
Spark Master UI looks as below.
Spark History Server UI looks like this.
Trino
Trino
is used as query engine to run interactive and long running ETL query in Chango.
Chango provides multiple trino clusters, so, you can install trino clusers as many as you want.
Install Trino
Enter default
for the cluser name if default trino cluster is not installed.
After installing default trino cluster, trino page looks like this.
If you want to install another trino cluster, enter any name for cluster name.
After installing another trino cluster, new created trino cluster will be shown in the cluster tab list.
Scale Worker
You can scale out and unscale trino workers.
Configure Trino
You can update trino memory configurations and catalogs.
To update memory properties in trino, select trino cluster.
Update catalogs in trino.
You can also add catalogs.
UI
You can get Trino UI clicking link in UI
of the selected trino cluster.
Chango Trino Gateway
Chango Trino Gateway
is used to route trino queries to the backend trino clusters in Chango.
In addition, Chango Trino Gateway
also provides the following functions.
- authenticate trino users and authorize the queries run by trino users.
- activate and deactivate the backend trino clusters.
Install Chango Trino Gateway
Select hosts for Chango Trino Gateway
servers, host for NGINX proxy, and host for Redis cache.
It looks as below after installing it.
Scale Server
Chango Trino Gateway can be scaled out or unscaled.
First, select hosts for scaling out or unscaling Trino Gateway servers, and then, press the button of Scale Out Server
to scale out Trino Gateway servers
or press the button of Unscale Server
to unscale Trino Gateway servers.
Apache Superset
Apache Superset
is used as BI tool in Chango.
Install Superset
Select host for superset server.
UI
You can get Superset UI clicking link in UI
of superset page.
Azkaban
Azkaban
is used as workflow to integrate all the batch jobs like spark ETL jobs and trino ETL jobs in Chango.
Install Azkaban
Select host for web and hosts for executors.
UI
You can get Azkaban UI clicking link in UI
of azkaban page.
Azkaban CLI
Azkaban CLI
is CLI to create and update azkaban project on Azkaban.
Install Azkaban CLI
Select hosts for azkaban CLI.
Chango Data API
Chango provides data ingestion especially for streaming events.
Chango Data API
is used to collect streaming events and produce them to kafka.
Install Chango Data API
Select hosts for Chango Data API servers and host for NGINX proxy.
Scale Server
You can scale out and unscale Chango Data API
servers.
Chango Streaming
Chango Streaming
is a spark streaming job and used to consume streaming events from kafka and save them to Iceberg table in Chango.
Install Chango Streaming
Enter spark configurations for Chango Streaming
job.
After installation, you will see the driver host of Chango Streaming
spark job.
In spark master UI, Chango Streaming
job will be found.
Install Chango Streaming Tx
This is transactional spark streaming jobs. Enter spark configurations for Chango Streaming Tx
job.
You need to enter iceberg
schema
andtable
to which streaming messages will be saved.
After installation, you will see chango streaming tx applications in running status in spark ui.
Chango Spark Thrift Server
Chango Spark Thrift Server
is a spark streaming job to run spark sql queries. Clients can connect to spark thrift server,
for example through JDBC/Thrift.
Install Chango Spark Thrift Server
Enter spark configurations for Chango Spark Thrift Server
job.
After installation, you will see spark thrift server in running status in spark ui.
Chango Spark SQL Runner
Chango Spark SQL Runner
is a spark streaming job which exposes REST API to run spark sql queries requested by clients through REST.
Install Chango Spark SQL Runner
Enter spark configurations for Chango Spark SQL Runner
job.
After installation, you will see spark sql runner in running status in spark ui.
Chango Query Exec
Chango Query Exec
is a REST application to execute trino ETL queries to transform data in Chango.
It may be used as alternative to dbt to transform data in Chango.
Install Chango Query Exec
Select hosts for Chango Query Exec servers and host for NGINX proxy.
Scale Server
You can scale out and unscale Chango Query Exec
servers.