Skip to content

Install Chango Components

There are several components supported by Chango.

See What is Chango? for more details.

NOTE: If Chango initialization failed or you want to reset Chango, then you need to move to http://[admin-host]:8123/cp-reset.html to reset Chango.

Initialize Chango

If you have not initialized Chango, you will get the following picture to initialze Chango.

There are mandatory Chango Components like MySQL, Object Storage, Chango Authorizer, and Chango REST Catalog which must be installed when Chango is initialized. The other optional compoents can be installed after Chango initialization.

Configure Hosts and SSH Private Key

All the Chango Component Nodes need to be registered, and SSH private key on Chango Admin Node need to be added to access Chango Component Nodes from the host of Chango Admin Node.

NOTE: Take a note that you need to register all the nodes of Chango Component Nodes except Chango Admin Node.

Get SSH private key on Chango Admin Node with the following command, and paste it to the text box SSH Private Key.

cat ~/.ssh/id_rsa

Configure LVM

Raw disks attached to the Chango Component Nodes in comma separated list need to be added to mount as logical volume.

If the disks attached are same on all the nodes, use this.

If the disks attached are different on every node, use this.

Configure MySQL

MySQL is used by open source components like Apache Superset and Apache Ozone in Chango. Select the host on which MySQL will be installed.

Configure Object Storage

Select the options for object storage. Apache Ozone is default object storage provided by Chango, which will be used in disconnected environment in most cases. In public, you may select the external object storage like AWS S3, MinIO, OCI Object Storage.

If Apache Ozone is used as object storage, enter the values this below.

If external object storage is used, enter values of S3 credentials.

Configure Chango Authorizer

Chango Authorizer is used to authenticate and authorize all the data access to Chango.

Configure Chango REST Catalog

Chango REST Catalog is used as data catalog in Chango.

For now, the components you configured are mandatory. After configuring mandatory components, you can skip configuration.

You can install other optional components later after finishing Chango initialization.

Install Configured Components

Install all the configured components.

When the installation is finished, press the button of Installation Finished.

Then, you will move to the main page.

Show Log

You can see current log produced by installed components.

Click the host name of components in Status to show log.

Chango Authorizer

Scale Server

Chango Authorizer can be scaled out or unscaled.

First, select hosts for scaling out or unscaling Authorizer servers, and then, press the button of Scale Out Server to scale out Authorizer servers or press the button of Unscale Server to unscale Authorizer servers.

Chango REST Catalog

Scale Server

Chango REST Catalog can be scaled out or unscaled.

First, select hosts for scaling out or unscaling REST Catalog servers, and then, press the button of Scale Out Server to scale out REST Catalog servers or press the button of Unscale Server to unscale REST Catalog servers.

Apache Kafka

Apache Kafka is used as event streaming platform in Chango. Multiple Kafka clusters are supported by Chango, that is, you can install kafka clusters as many as you want.

Install Kafka

If you want to install Apache Kafka, press Go to Install button.

NOTE: If you have not installed any kafka cluster, then, enter default for the cluster name.

After installing kafka, you will see kafka page like this.

Because Chango supports multiple kafka clusters, you can install another kafka cluster.

NOTE: Because you have already installed default kafka cluster, you can enter anything for the cluster name.

After installing another kafka cluster, new created kafka cluster will be shown in tab list.

Scale Broker

Kafka Broker can be scaled out or unscaled.

First, select hosts for scaling out or unscaling kafka brokers, and then, press the button of Scale Out Broker to scale out brokers or press the button of Unscale Broker to unscale brokers.

Configure Kafka

You can update kafka configurations like heap memory and server.properties.

First, select kafka cluster which you want to configure.

After modifying configuration, press Update to update the selected kafka cluster.

Apache Spark

Apache Spark is used as computing engine to run batch and streaming jobs in Chango.

Install Spark

Select hosts to install master, worker and history server of Spark.

After installing Spark, you will see the spark page like this.

Scale Worker

You can scale out and unscale spark workers.

UI

There are URL links to get Spark Master UI and Spark History Server UI in UI of Spark Page.

Spark Master UI looks as below.

Spark History Server UI looks like this.

Trino

Trino is used as query engine to run interactive and long running ETL query in Chango. Chango provides multiple trino clusters, so, you can install trino clusers as many as you want.

Install Trino

Enter default for the cluser name if default trino cluster is not installed.

After installing default trino cluster, trino page looks like this.

If you want to install another trino cluster, enter any name for cluster name.

After installing another trino cluster, new created trino cluster will be shown in the cluster tab list.

Scale Worker

You can scale out and unscale trino workers.

Configure Trino

You can update trino memory configurations and catalogs.

To update memory properties in trino, select trino cluster.

Update catalogs in trino.

You can also add catalogs.

UI

You can get Trino UI clicking link in UI of the selected trino cluster.

Chango Trino Gateway

Chango Trino Gateway is used to route trino queries to the backend trino clusters in Chango.

In addition, Chango Trino Gateway also provides the following functions. - authenticate trino users and authorize the queries run by trino users. - activate and deactivate the backend trino clusters.

Install Chango Trino Gateway

Select hosts for Chango Trino Gateway servers, host for NGINX proxy, and host for Redis cache.

It looks as below after installing it.

Scale Server

Chango Trino Gateway can be scaled out or unscaled.

First, select hosts for scaling out or unscaling Trino Gateway servers, and then, press the button of Scale Out Server to scale out Trino Gateway servers or press the button of Unscale Server to unscale Trino Gateway servers.

Apache Superset

Apache Superset is used as BI tool in Chango.

Install Superset

Select host for superset server.

UI

You can get Superset UI clicking link in UI of superset page.

Azkaban

Azkaban is used as workflow to integrate all the batch jobs like spark ETL jobs and trino ETL jobs in Chango.

Install Azkaban

Select host for web and hosts for executors.

UI

You can get Azkaban UI clicking link in UI of azkaban page.

Azkaban CLI

Azkaban CLI is CLI to create and update azkaban project on Azkaban.

Install Azkaban CLI

Select hosts for azkaban CLI.

Chango Data API

Chango provides data ingestion especially for streaming events. Chango Data API is used to collect streaming events and produce them to kafka.

Install Chango Data API

Select hosts for Chango Data API servers and host for NGINX proxy.

Scale Server

You can scale out and unscale Chango Data API servers.

Chango Streaming

Chango Streaming is a spark streaming job and used to consume streaming events from kafka and save them to Iceberg table in Chango.

Install Chango Streaming

Enter spark configurations for Chango Streaming job.

After installation, you will see the driver host of Chango Streaming spark job.

In spark master UI, Chango Streaming job will be found.

Install Chango Streaming Tx

This is transactional spark streaming jobs. Enter spark configurations for Chango Streaming Tx job.

You need to enter iceberg schema and table to which streaming messages will be saved.

After installation, you will see chango streaming tx applications in running status in spark ui.

Chango Spark Thrift Server

Chango Spark Thrift Server is a spark streaming job to run spark sql queries. Clients can connect to spark thrift server, for example through JDBC/Thrift.

Install Chango Spark Thrift Server

Enter spark configurations for Chango Spark Thrift Server job.

After installation, you will see spark thrift server in running status in spark ui.

Chango Spark SQL Runner

Chango Spark SQL Runner is a spark streaming job which exposes REST API to run spark sql queries requested by clients through REST.

Install Chango Spark SQL Runner

Enter spark configurations for Chango Spark SQL Runner job.

After installation, you will see spark sql runner in running status in spark ui.

Chango Query Exec

Chango Query Exec is a REST application to execute trino ETL queries to transform data in Chango. It may be used as alternative to dbt to transform data in Chango.

Install Chango Query Exec

Select hosts for Chango Query Exec servers and host for NGINX proxy.

Scale Server

You can scale out and unscale Chango Query Exec servers.