forked from D-Net/dnet-hadoop
addded command line params to allow hive actions to run
This commit is contained in:
parent
0353f93d54
commit
6fc9ef53f6
|
@ -9,6 +9,8 @@ fi
|
||||||
CONTEXT_API=$1
|
CONTEXT_API=$1
|
||||||
TARGET_DB=$2
|
TARGET_DB=$2
|
||||||
|
|
||||||
|
export HIVE_OPTS="-hiveconf mapred.job.queue.name=analytics -hiveconf hive.spark.client.connect.timeout=120000ms -hiveconf hive.spark.client.server.connect.timeout=300000ms -hiveconf spark.executor.memory=4831838208 -hiveconf spark.yarn.executor.memoryOverhead=450"
|
||||||
|
|
||||||
TMP=/tmp/stats-update-`tr -dc A-Za-z0-9 </dev/urandom | head -c 6`
|
TMP=/tmp/stats-update-`tr -dc A-Za-z0-9 </dev/urandom | head -c 6`
|
||||||
|
|
||||||
echo "Downloading context ids"
|
echo "Downloading context ids"
|
||||||
|
@ -29,13 +31,16 @@ hdfs dfs -copyFromLocal categories.csv ${TMP}
|
||||||
hdfs dfs -copyFromLocal concepts.csv ${TMP}
|
hdfs dfs -copyFromLocal concepts.csv ${TMP}
|
||||||
hdfs dfs -chmod -R 777 ${TMP}
|
hdfs dfs -chmod -R 777 ${TMP}
|
||||||
|
|
||||||
|
export HADOOP_USER="antonis.lempesis"
|
||||||
|
export HADOOP_USER_NAME="antonis.lempesis"
|
||||||
|
|
||||||
echo "Creating and populating impala tables"
|
echo "Creating and populating impala tables"
|
||||||
hive -e "create table ${TARGET_DB}.context (id string, name string) row format delimited fields terminated by ','"
|
hive $HIVE_OPTS -e "create table ${TARGET_DB}.context (id string, name string) row format delimited fields terminated by ','"
|
||||||
hive -e "create table ${TARGET_DB}.category (context string, id string, name string) row format delimited fields terminated by ','"
|
hive $HIVE_OPTS -e "create table ${TARGET_DB}.category (context string, id string, name string) row format delimited fields terminated by ','"
|
||||||
hive -e "create table ${TARGET_DB}.concept (category string, id string, name string) row format delimited fields terminated by ','"
|
hive $HIVE_OPTS -e "create table ${TARGET_DB}.concept (category string, id string, name string) row format delimited fields terminated by ','"
|
||||||
hive -e "load data inpath '${TMP}/contexts.csv' into table ${TARGET_DB}.context"
|
hive $HIVE_OPTS -e "load data inpath '${TMP}/contexts.csv' into table ${TARGET_DB}.context"
|
||||||
hive -e "load data inpath '${TMP}/categories.csv' into table ${TARGET_DB}.category"
|
hive $HIVE_OPTS -e "load data inpath '${TMP}/categories.csv' into table ${TARGET_DB}.category"
|
||||||
hive -e "load data inpath '${TMP}/concepts.csv' into table ${TARGET_DB}.concept"
|
hive $HIVE_OPTS -e "load data inpath '${TMP}/concepts.csv' into table ${TARGET_DB}.concept"
|
||||||
|
|
||||||
echo "Cleaning up"
|
echo "Cleaning up"
|
||||||
rm concepts.csv
|
rm concepts.csv
|
||||||
|
|
|
@ -8,7 +8,8 @@ fi
|
||||||
|
|
||||||
export TARGET=$1
|
export TARGET=$1
|
||||||
export SCRIPT_PATH=$2
|
export SCRIPT_PATH=$2
|
||||||
export HIVE_OPTS="-hiveconf mapred.job.queue.name=analytics -hiveconf hive.spark.client.connect.timeout=120000ms -hiveconf hive.spark.client.server.connect.timeout=300000ms"
|
export HIVE_OPTS="-hiveconf mapred.job.queue.name=analytics -hiveconf hive.spark.client.connect.timeout=120000ms -hiveconf hive.spark.client.server.connect.timeout=300000ms -hiveconf spark.executor.memory=4831838208 -hiveconf spark.yarn.executor.memoryOverhead=450"
|
||||||
|
export HADOOP_USER="antonis.lempesis"
|
||||||
|
|
||||||
echo "Getting file from " $SCRIPT_PATH
|
echo "Getting file from " $SCRIPT_PATH
|
||||||
hdfs dfs -copyToLocal $SCRIPT_PATH
|
hdfs dfs -copyToLocal $SCRIPT_PATH
|
||||||
|
|
|
@ -74,7 +74,7 @@
|
||||||
</configuration>
|
</configuration>
|
||||||
</global>
|
</global>
|
||||||
|
|
||||||
<start to="Step16-createIndicatorsTables"/>
|
<start to="Contexts"/>
|
||||||
|
|
||||||
<kill name="Kill">
|
<kill name="Kill">
|
||||||
<message>Action failed, error message[${wf:errorMessage(wf:lastErrorNode())}]</message>
|
<message>Action failed, error message[${wf:errorMessage(wf:lastErrorNode())}]</message>
|
||||||
|
@ -280,7 +280,7 @@
|
||||||
<argument>${wf:appPath()}/scripts/step16-createIndicatorsTables.sql</argument>
|
<argument>${wf:appPath()}/scripts/step16-createIndicatorsTables.sql</argument>
|
||||||
<file>indicators.sh</file>
|
<file>indicators.sh</file>
|
||||||
</shell>
|
</shell>
|
||||||
<ok to="End"/>
|
<ok to="Step16_1-definitions"/>
|
||||||
<error to="Kill"/>
|
<error to="Kill"/>
|
||||||
</action>
|
</action>
|
||||||
|
|
||||||
|
|
Loading…
Reference in New Issue