Code Monkey home page Code Monkey logo

kylin-gcp's Introduction

Kylin on GCP

This project demonstrates the use of Apache Kylin on GCP backed by Dataproc.

Prerequisites

To get started you will need a GCP project, to create one see here. This demo uses Terraform along with GCP provider tool which can be installed on your machine directly or via docker accessed via docker. The demo can also be run from cloud shell.

Demo

Create Hadoop/Kylin cluster

The Terraform components assume that a GCP project exists from which to deploy the Kylin cluster. It also assumes the existence of GCP bucket in order to store terraform state, also known as backends in Terraform. If you do not wish to use the included GCS backend configuration simply comment-out the following in the gcp.tf file (note: backend configuration cannot be populated by TF-variables as of v0.11.11):

# comment this out to store Terraform state locally
# otherwise, enter the name of the ops-bucket in this configuration
terraform {
  backend "gcs" {
    bucket  = "<OPS_BUCKET_NAME>"
    prefix  = "terraform/state/kylin"
  }
}

The kylin.tf file contains the deployment configuration for the Kylin cluster. To deploy it, first provide deployment info via environment variables (note: you may elect to use a different method of providing TF-variables, see here):

# ID of existing project to deploy too
export GOOGLE_PROJECT="<YOUR_PROJECT>"
# name of resource bucket (created by deployment)
export TF_VAR_resource_bucket="${GOOGLE_PROJECT}"
# master node for shell access to dataproc cluster
export MASTER='kylin-m-2'

Now that the deployments have the parameters, verify the deployment plan using:

terraform init # only required once
terraform plan

and deploy using:

terraform apply
# use "-auto-approve" to skip prompt

If successful, the deployment will deploy a resource bucket and upload the necessary init-scripts, then create the Kylin cluster.

Once the cluster is up and running, start Kylin on one of the nodes; note a master node is being used in this example, however any node can be seleted:

gcloud compute ssh ${MASTER} --command="source /etc/profile && kylin.sh start"

This will start the Kylin service. With Kylin installed and running yo can now tunnel to the master node to bring up the kylin UI:

gcloud compute ssh ${MASTER} -- -L 7070:${MASTER}:7070
# then open browser to http://localhost:7070/kylin
# default creds ADMIN/KYLIN

Writting data to cluster

Kylin gets its data from structured sources such as Hive and other JDBC/ODBC compliant sources. Hive in turn relies on the Hadoop filesystem HDFS to store the unerlying data. In standard (on-prem) Hadoop deployments, HDFS would be deployed accross the disks attached to the workers of the Hadoop cluster. In GCP Dataproc however, the HDFS filesystem can (and is by default) backed by the GCS object storage service using the Cloud Storage Connector. This provides the decoupling of storage and compute resources, providing the cost-benefits of both cheaper/scalable storage in comparison to persistant disks, and allows cluster nodes (or the at least the workers) to be ephemiral because they are only used for computation and not for persistent storage.

Building on the former, Hive can be used by creating table/schema definitions and then pointing them to "directories" in HDFS (GCS). Once a table is created, data can be written to GCS from any source using the standard GCS APIs without the need to use any hive/hadoop interfaces. To demonstrate this, create a table with a schema pointing to a path in GCS:

# example creation of hive table
hive -e "CREATE EXTERNAL TABLE test
(
unique_key STRING,
complaint_type STRING,
complaint_description STRING,
owning_department STRING,
source STRING,
status STRING,
status_change_date TIMESTAMP,
created_date TIMESTAMP,
last_update_date TIMESTAMP,
close_date TIMESTAMP,
incident_address STRING,
street_number STRING,
street_name STRING,
city STRING,
incident_zip INTEGER,
county STRING,
state_plane_x_coordinate STRING,
state_plane_y_coordinate FLOAT,
latitude FLOAT,
longitude FLOAT,
location STRING,
council_district_code INTEGER,
map_page STRING,
map_tile STRING
)
row format delimited
fields terminated by ','
LOCATION 'gs://${BUCKET}/data/311/csv/' ;"

the above can be run on one of the dataproc nodes (gcloud compute ssh ${MASTER}) or using gcloud dataproc jobs submit .... Note that we are using CSV format for this demo, however, there are many many options for serialization with HDFS/Hive. for more info see SerDe.

To populate the table location path with data, run the following dataflow job to which will extract public 311-service-request data from the city of Austin, TX which is stored in BigQuery:

cd dataflow
./gradlew ronLocal # run Apache Beam job locally
# OR
./gradlew runFlow # submit job to run on GCP Dataflow

Note that this Beam job is simply extracting BigQuery data via query, converting it to csv form, and writting it to GCS. Writting data to hive can also be achieved in Beam using the HCatalog IO, however, is not recommeded if the primary query engine is Kylin and not Hive.

As data is added to the path, new queries will pick up this new data:

hive -e "select count(unique_key) from test;"

With tables created and populated, return to the Kylin web interface; if the session has terminated restart it with the tunnel:

gcloud compute ssh ${MASTER} -- -L 7070:${MASTER}:7070
# open browser to http://localhost:7070/kylin
# default creds ADMIN/KYLIN

From here, the table can be loaded in Kylin:

  • Model tab -> Data Sources sub-tab
  • Load Table (blue button)
  • enter table name "test"
  • Under Tables the hive database DEFAULT should be displayed with the test just loaded

Once tables are loaded you can continue to create models/cubes within the kylin interface, for more see the Kylin Docs.

Cleanup

Deleting the Kylin cluster and associated resources:

terraform destroy #(yes at prompt)

OR delete the project entirely to ensure no other resources are incurring costs.

Future Options:

  • Secure cluster and access
  • configure Kylin init-action for HA-mode (with load-balancing)
  • Autoscaling setup
  • BigTable (HBase) cube storage substitution
  • persistant disk HDFS substitution
  • Spark substitution
  • Streaming cube (Kafka) sample
  • JDBC source sample
  • hadoop resource optimization (disk, cpu, preemptible workers, etc)
  • cube creation sample (real-world data)
  • hive-metastore Cloud-SQL substition

kylin-gcp's People

Contributors

rsimm-muvaki avatar

Stargazers

 avatar  avatar  avatar  avatar  avatar

Watchers

 avatar  avatar

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    ๐Ÿ–– Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. ๐Ÿ“Š๐Ÿ“ˆ๐ŸŽ‰

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google โค๏ธ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.