#!/bin/sh sudo mkdir /apache && sudo chown hadoop.hadoop /apache sudo curl https://s3.amazonaws.com/apache-atlas-setup-on-emr/apache-atlas-1.0.0-bin.tar.gz -o /tmp/apache-atlas-1.0.0-bin.tar.gz && sudo tar xvpfz /tmp/apache-atlas-1.0.0-bin.tar.gz -C /apache sudo curl https://s3.amazonaws.com/apache-atlas-setup-on-emr/kafka_2.11-1.1.0.tgz -o /tmp/kafka_2.11-1.1.0.tgz && sudo tar xvpfz /tmp/kafka_2.11-1.1.0.tgz -C /apache # To avoid yum install fails: At least 47MB more space needed on the / filesystem, sudo sh -c "echo 'diskspacecheck=0' >> /etc/yum.conf " sudo yum install -y https://s3.amazonaws.com/apache-atlas-setup-on-emr/jdk-8u171-linux-x64.rpm # Create symlinks sudo ln -s /apache/kafka_2.11-1.1.0 /apache/kafka sudo ln -s /apache/apache-atlas-1.0.0 /apache/atlas # Change default port for zookeeper and kafka sudo sed -i 's/clientPort=2181/clientPort=3000/' /apache/kafka/config/zookeeper.properties sudo sed -i 's/zookeeper.connect=localhost:2181/zookeeper.connect=localhost:3000/' /apache/kafka/config/server.properties #set these variables in hadoop user's bash profile sudo cat << EOL >> /home/hadoop/.bash_profile export JAVA_HOME=/usr/java/jdk1.8.0_171-amd64/ export MANAGE_LOCAL_HBASE=false export MANAGE_LOCAL_SOLR=true export HIVE_HOME=/usr/lib/hive export HIVE_CONF_DIR=/usr/lib/hive/conf EOL export JAVA_HOME=/usr/java/jdk1.8.0_171-amd64/ sudo /apache/kafka/bin/zookeeper-server-start.sh -daemon /apache/kafka/config/zookeeper.properties sudo /apache/kafka/bin/kafka-server-start.sh -daemon /apache/kafka/config/server.properties # Create a symlink in native hive's conf directory sudo ln -s /apache/atlas/conf/atlas-application.properties /usr/lib/hive/conf/atlas-application.properties # add hive hook in /etc/hive/conf/hive-site.xml # Ensure that following is present: # # hive.exec.post.hooks # org.apache.atlas.hive.hook.HiveHook # sudo cp /etc/hive/conf/hive-site.xml /etc/hive/conf/hive-site.xml.orig sudo sed -i "s## \n hive.exec.post.hooks\n org.apache.atlas.hive.hook.HiveHook\n \n\n#" /etc/hive/conf/hive-site.xml || mv /etc/hive/conf/hive-site.xml.orig /etc/hive/conf/hive-site.xml # Create symlinks to the jar files under atlas folder cd /usr/lib/hive/lib sudo ln -s /apache/atlas/hook/hive/atlas-plugin-classloader-1.0.0.jar sudo ln -s /apache/atlas/hook/hive/hive-bridge-shim-1.0.0.jar for i in /apache/atlas/hook/hive/atlas-hive-plugin-impl/*; do sudo ln -s $i; done # Restart hive server export HIVE_HOME=/usr/lib/hive export HIVE_CONF_DIR=/usr/lib/hive/conf sudo systemctl stop hive-server2 sudo systemctl start hive-server2 # To run Apache Atlas with local Apache HBase & Apache Solr instances that are started/stopped along with Atlas start/stop, run following commands: sudo sed -i 's?#export JAVA_HOME=?export JAVA_HOME=/usr/java/jdk1.8.0_171-amd64?' /apache/atlas/conf/atlas-env.sh sudo sed -i 's/export MANAGE_LOCAL_HBASE=true/export MANAGE_LOCAL_HBASE=false/' /apache/atlas/conf/atlas-env.sh sudo /apache/atlas/bin/atlas_start.py && sudo /apache/atlas/bin/atlas_stop.py && sudo /apache/atlas/bin/atlas_start.py