sparksql和hive集成

########################################
alter database hive character set latin1;
ALTER TABLE hive.* DEFAULT CHARACTER SET latin1;
########################################


1.安装hive
CREATE USER 'hive'@'%' IDENTIFIED BY '123456';
GRANT all privileges ON hive.* TO 'hive'@'%';
flush privileges;


2.将配置好的hive-site.xml放入$SPARK-HOME/conf目录下


3.启动spark-shell时指定mysql连接驱动位置
bin/spark-shell \
 --master spark://node1.itcast.cn:7077 \
 --executor-memory 1g \
 --total-executor-cores 2 \
 --driver-class-path /usr/local/apache-hive-0.13.1-bin/lib/mysql-connector-java-5.1.35-bin.jar 


4.使用sqlContext.sql调用HQL
sqlContext.sql("select * from spark.person limit 2")


或使用org.apache.spark.sql.hive.HiveContext
import org.apache.spark.sql.hive.HiveContext
val hiveContext = new HiveContext(sc)
hiveContext.sql("select * from spark.person")




bin/spark-sql \
 --master spark://node1.itcast.cn:7077 \
 --executor-memory 1g \
 --total-executor-cores 2 \
 --driver-class-path /usr/local/apache-hive-0.13.1-bin/lib/mysql-connector-java-5.1.35-bin.jar 




1.安装hive


2.将配置好的hive-site.xml放入$SPARK-HOME/conf目录下


3.启动spark-shell时指定mysql连接驱动位置
bin/spark-shell --master spark://spark1.itcast.cn:7077 --executor-memory 1g --total-executor-cores 2 --driver-class-path /usr/local/spark-1.3.1-bin-hadoop2.4/lib/mysql-connector-java-5.1.35-bin.jar


4.使用sqlContext.sql调用HQL
sqlContext.sql("select * from spark.person limit 2")


或使用org.apache.spark.sql.hive.HiveContext
import org.apache.spark.sql.hive.HiveContext
val hiveContext = new HiveContext(sc)
hiveContext.sql("select * from spark.person")

你可能感兴趣的:(spark)