Bootstrapping GeoMesa HBase on AWS S3¶
GeoMesa can be run on top of HBase using S3 as the underlying storage engine. This mode of running GeoMesa is cost-effective as one sizes the database cluster for the compute and memory requirements, not the storage requirements. The following guide describes how to bootstrap GeoMesa in this manner. This guide assumes you have an Amazon Web Services account already provisioned as well as an IAM key pair.
An EMR cluster can be bootstrapped either via the AWS Web Console (recommended for new users) or from another EC2 instance via the AWS CLI.
Creating an EMR Cluster (Web Console)¶
To begin, sign into the AWS Web Console. Ensure that you have created a keypair before beginning. Select “EMR” from the list of services and then select “Create Cluster” to begin. Once you have entered the wizard switch to the “Advanced View” and select HBase, Spark, and Hadoop for your software packages. Deselect all others. After selecting HBase you will see an “HBase storage settings” configuration area where you can enter a bucket to use as the HBase Root directory. You’ll want to ensure this bucket is in the same region as your HBase cluster for performance and cost reasons. On the next pages you can select and customize your hardware and give your cluster a good name.
After your cluster has bootstrapped you can view the hardware associated with your cluster. Find the public IP of the MASTER server and connect to it via SSH:
$ ssh -i /path/to/key ec2-user@<your-master-ip>
Creating an EMR Cluster (AWS CLI)¶
This section is meant for advanced users. If you have bootstrapped a cluster via the Web Console you can skip this section and continue on to Installing GeoMesa. The instructions below were executed on an AWS EC2 machine running Amazon Linux. To set up the AWS command line tools, follow the instructions found in the AWS online documentation.
First, you will need to configure an S3 bucket for use by HBase. Make sure to replace <bucket-name>
with your bucket
name. You can also use a different root directory for HBase if you desire. If you’re using the AWS CLI you can create a
bucket and the root “directory” this:
$ aws s3 mb s3://<bucket-name>
$ aws s3api put-object --bucket <bucket-name> --key hbase-root/
You should now be able to list the contents of your bucket:
$ aws s3 ls s3://<bucket-name>/
PRE hbase-root/
Next, create a local json file named geomesa-hbase-on-s3.json
with the following content. Make sure to replace
<bucket-name>/hbase-root
with a unique root directory for HBase that you configured in the previous step.
[
{
"Classification": "hbase-site",
"Properties": {
"hbase.rootdir": "s3://<bucket-name>/hbase-root"
}
},
{
"Classification": "hbase",
"Properties": {
"hbase.emr.storageMode": "s3"
}
}
]
Then, use the following command to bootstrap an EMR cluster with HBase. You will need to change __KEY_NAME__
to
the IAM key pair you intend to use for this cluster and __SUBNET_ID__
to the id of the subnet if that key is
associated with a specific subnet. You can also edit the instance types to a size appropriate for your use case.
Specify the appropriate path to the json file you created in the last step.
You may desire to run aws configure
before running this command. If you don’t you’ll need to specify a region
something like --region us-west-2
. Also, you’ll need to ensure that your EC2 instance has the IAM Role to perform
the elasticmapreduce:RunJobFlow
action. The config below will create a single master and 3 worker nodes. You may
wish to increase or decrease the number of worker nodes or change the instance types to suit your query needs.
Note
In the code below, $VERSION
= 2.0.2
$ export CID=$(
aws emr create-cluster \
--name "GeoMesa HBase on S3" \
--release-label emr-5.5.0 \
--output text \
--use-default-roles \
--ec2-attributes KeyName=__KEY_NAME__,SubnetId=__SUBNET_ID__ \
--applications Name=Hadoop Name=Zookeeper Name=Spark Name=HBase \
--instance-groups \
Name=Master,InstanceCount=1,InstanceGroupType=MASTER,InstanceType=m4.2xlarge \
Name=Workers,InstanceCount=3,InstanceGroupType=CORE,InstanceType=m4.xlarge \
--configurations file:///path/to/geomesa-hbase-on-s3.json \
)
After executing that command, you can monitor the state of the EMR bootstrap process by going to the Management Console. Or by running the following command:
watch 'aws emr describe-cluster --cluster-id $CID | grep MasterPublic | cut -d "\"" -f 4'
Once the cluster is provisioned you can run the following code to retrieve its hostname.
export MASTER=$(aws emr describe-cluster --cluster-id $CID | grep MasterPublic | cut -d "\"" -f 4)
Optionally you can find the hostname for the master node on the AWS management console. Find the name (as specified in
the aws emr
command) of the cluster and click through to its details page. Under the Hardware section, you can
find the master node and its IP address. Copy the IP address and then run the
following command.
export MASTER=<ip_address>
To configure GeoMesa, remote into the master node of your new AWS EMR cluster using the following command:
$ ssh -i /path/to/key ec2-user@$MASTER
Installing GeoMesa¶
Now that you have SSH’d into your master server you can test out your HBase and Hadoop installations by running these commands:
hbase version
hadoop version
If everything looks good, download the GeoMesa HBase distribution, replacing ${VERSION}
with the appropriate GeoMesa
Version (e.g. 1.3.4) or setting the VERSION
environment variable.
$ wget "https://repo.locationtech.org/content/repositories/geomesa-releases/org/locationtech/geomesa/geomesa-hbase-dist_2.11/${VERSION}/geomesa-hbase-dist_2.11-${VERSION}-bin.tar.gz" -o /tmp/geomesa-hbase-dist_2.11-${VERSION}-bin.tar.gz
$ cd /opt
$ sudo tar zxvf /tmp/geomesa-hbase-dist_2.11-${VERSION}-bin.tar.gz
Then, bootstrap GeoMesa on HBase on S3 by executing the provided script. This script sets up the needed environment variables, copies hadoop jars into GeoMesa’s lib directory, copies the GeoMesa distributed runtime into S3 where HBase can utilize it, sets up the GeoMesa coprocessor registration among other administrative tasks.
$ sudo /opt/geomesa-hbase_2.11-${VERSION}/bin/bootstrap-geomesa-hbase-aws.sh
Now, log out and back in and your environment will be set up appropriately.
Ingest Public GDELT data¶
GeoMesa ships with predefined data models for many open spatio-temporal data sets such as GDELT. To ingest the most recent 7 days of GDELT from Amazon’s public S3 bucket, one can copy the files locally to the cluster or use a distributed ingest:
Local ingest:
mkdir gdelt
cd gdelt
seq 7 -1 1 | xargs -n 1 -I{} sh -c "date -d'{} days ago' +%Y%m%d" | xargs -n 1 -I{} aws s3 cp s3://gdelt-open-data/events/{}.export.csv .
# you'll need to ensure the hbase-site.xml is provided on the classpath...by default it is picked up by the tools from standard locations
geomesa-hbase ingest -c geomesa.gdelt -C gdelt -f gdelt -s gdelt \*.csv
Distributed ingest:
# we need to package up the hbase-site.xml for use in the distributed classpath
# zip and jar files found in GEOMESA_EXTRA_CLASSPATHS are picked up for the distributed classpath
cd /etc/hadoop/conf
zip /tmp/hbase-site.zip hbase-site.xml
export GEOMESA_EXTRA_CLASSPATHS=/tmp/hbase-site.zip
# now lets ingest
files=$(for x in `seq 7 -1 1 | xargs -n 1 -I{} sh -c "date -d'{} days ago' +%Y%m%d"`; do echo "s3a://gdelt-open-data/events/$x.export.csv"; done)
geomesa-hbase ingest -c geomesa.gdelt -C gdelt -f gdelt -s gdelt $files
You can then query the data using the GeoMesa command line export tool.
geomesa-hbase export -c geomesa.gdelt -f gdelt -m 50
Setup GeoMesa and SparkSQL¶
To start executing SQL queries using Spark over your GeoMesa on HBase on S3 cluster, set up the following variable, replacing VERSION
with the appropriate version of GeoMesa.
$ JARS=file:///opt/geomesa/dist/spark/geomesa-hbase-spark-runtime_2.11-${VERSION}.jar,file:///usr/lib/hbase/conf/hbase-site.xml
Then, start up the Spark shell
$ spark-shell --jars $JARS
Within the Spark shell, you can connect to GDELT and issues some queries.
scala> val df = spark.read.format("geomesa").option("bigtable.table.name", "geomesa.gdelt").option("geomesa.feature", "gdelt").load()
scala> df.createOrReplaceTempView("gdelt")
scala> spark.sql("SELECT globalEventId,geom,dtg FROM gdelt LIMIT 5").show()