#sqoop
解压软件
$ sudo tar -zxvf sqoop-1.4.6.bin__hadoop-0.23.tar.gz -C /usr/local/
$ pwd
/usr/local/sqoop-1.4.6.bin__hadoop-0.23
$ cd conf/
$ sudo cp sqoop-env-template.sh sqoop-env.sh
$ sudo vi sqoop-env.sh
export HADOOP_COMMON_HOME=/usr/local/hadoop-2.7.4/
export HADOOP_MAPRED_HOME=/usr/local/hadoop-2.7.4/
#sqoop 导入Oracle数据库到HDFS中
$ sqoop import --connect jdbc:oracle:thin:@192.168.240.137:1521:orcl --username scott --password tiger --table emp --columns 'empno,ename,job,sal,deptno' -m 1 --target-dir '/sqoop/emp'
#sqoop 导入Oracle到hive中
$ sqoop import --hive-import --connect jdbc:oracle:thin:@192.168.240.137:1521:orcl --username scott --password tiger --table emp -m 1 --columns 'empno,ename,job,sal,deptno'
#导入Oracle,并指明表名
$ sqoop import --hive-import --connect jdbc:oracle:thin:@192.168.240.137:1521:orcl --username scott --password tiger --table emp -m 1 --columns 'empno,ename,job,sal,deptno' --hive-table emp1
#导入Oracle,指定whare条件
$ sqoop import --hive-import --connect jdbc:oracle:thin:@192.168.240.137:1521:orcl --username scott --password tiger --table emp -m 1 --columns 'empno,ename,job,sal,deptno' --hive-table emp2 where 'deptno=10'
#导入Oracle,使用select
$ sqoop import --hive-import --connect jdbc:oracle:thin:@192.168.240.137:1521:orcl --username scott --password tiger -m 1 --query 'select * from emp where sal < 1000 and $CONDITIONS' --target-dir '/sqoop/emp3' --hive-table emp3
$ sqoop import --hive-import --connect jdbc:oracle:thin:@192.168.240.137:1521:orcl --username scott --password tiger -m 1 --table myemp --export-dir '/sqoop/emp3'
免责声明:本站发布的内容(图片、视频和文字)以原创、转载和分享为主,文章观点不代表本网站立场,如果涉及侵权请联系站长邮箱:is@yisu.com进行举报,并提供相关证据,一经查实,将立刻删除涉嫌侵权内容。