Skip to main content

sqoop

----------for database connect--------
mysql -u root -pcloudera
----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------
>>sqoop import --connect jdbc://localhost/retail_db --username root --password cloudera --table customers --target-dir /data/sqoop/test01
>>sqoop import --connect jdbc:mysql://localhost/retail_db --username root --password cloudera --table customers --target-dir /data/sqoop/test02 -m 10
>>sqoop import --connect jdbc:mysql://localhost/retail_db --username root --password cloudera --table customers --target-dir /data/sqoop/test03 --split-by customer_id --fields-terminated-by '|' -m 10
>>sqoop list-databases --connect jdbc:mysql://localhost/retail_db --username root --password cloudera
>>sqoop list-tables --connect jdbc:mysql://localhost/retail_db --username root --password cloudera
----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------
sqoop import --connect jdbc:mysql://localhost/retail_db --username root --password cloudera --table customers  --split-by customer_id --fields-terminated-by '|' -m 10 --hive-import --create-hive-table --hive-table veer.customers2

---------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------
My job,export,dynamic partition...



create table customer5 (customer_id int , customer_fname varchar , customer_lname varchar,
       customer_email varchar,customer_password varchar, customer_street varchar, customer_city varchar ,
       customer_state varchar,customer_zipcode varchar);


hive>> create table customer_inc (customer_id int , customer_fname string , customer_lname string,
       customer_email string,customer_password string, customer_street string, customer_city string ,
       customer_state string ,customer_zipcode string);

sqoop import --connect jdbc:mysql://localhost/retail_db --username root --password cloudera --table customers --split-by customer_id --fields-terminated-by ',' -m 10 --hcatalog-database veer --hcatalog-table customer4;


>>>sqoop job --create import --connect jdbc:mysql://localhost/retail_db --username root --password cloudera --table customers --hive-import --hive-table customer4 --hive-partition-key customer_zipcode;
>>>
sqoop import --connect jdbc:mysql://localhost/retail_db \
--username root --password cloudera \
--hive-import \
--query "select customer_id ,customer_fname,customer_lname,customer_email,customer_password,customer_street,customer_city,customer_state from retail_db.customers where customer_zipcode='00725' and \$CONDITIONS" \
--hive-table customer4 \
--hive-partition-key customer_zipcode \
--hive-partition-value 'PR00725' \
--target-dir /user/cloudera/customer_temp \
--split-by customer_id

sqoop import --connect jdbc:mysql://localhost/retail_db \
--username root --password cloudera \
--hive-import \
--query "select * from retail_db.customers where \$CONDITIONS" \
--hive-table customer5 \
--hive-partition-key customer_zipcode \
--target-dir /user/cloudera/customer_temp \
--split-by customer_id


--table customers --where "customer_zipcode='PR00725'"




_____________________________________________________________________________________________________________________________________________________________________
create table ct1 select * from customers where 1=2;
sqoop export --connect jdbc:mysql://localhost/retail_db \
--username root --password cloudera \
--table temp \
--export-dir /user/hive/warehouse/temp
______________________________________________________________________________________________
sqoop job -create inc_upt import --connect jdbc:mysql://localhost/retail_db \
--username root --password cloudera \
--table customers \
--incremental append \
--check-column customer_id  last-value 12435 \
--target-dir /data/sqoop/test02


sqoop job –create jobname –<space>import –connect jdbc:mysql://localhost/DBname –username username –password password –table tablename –incremental append –check-column colname last-value 101

Comments

Popular posts from this blog

How to learn Hadoop

First of all, I want to tell you. here all content you will get  practical only. for the theoretical part, you can follow edureka videos and Durgasoft videos. once you complete some videos related to map-reduce and Hadoop ecosystem..after that you can follow me... you need to go step by step:- 1. What is big data 2. What is the need of Hadoop 3. How Hadoop works. 4. Tools in Hadoop ecosystem. If you are a beginner then first you need some configuration in your system which is: 1: RAM:-             Minimum RAM size should be 8gb otherwise you will get frustrated because of a slow system.             if you go with 16gb RAM then it will be great. My opinion is to go with 16gb RAM. 2: Download VMWARE. https://my.vmware.com/en/web/vmware/info/slug/desktop_end_user_computing/vmware_workstation_pro/14_0 3. Download Cloudera -> https://www.cloudera.com/downloads/cdh/5-14-0.html . inst...

Big Data in Business

How is big data impacting business and people? Have you ever searched for or bought a product on Amazon? Did you notice that Amazon started making recommendations related to the product you searched for Recommendation engines are a common application of big data. Companies like Amazon, Netflix and Spotify use algorithms based on big data to make specific recommendations based on customer preferences and historical behavior. Personal assistants like Siri on Apple devices use big data to devise answers to the infinite number of questions end users may ask. Google now makes recommendations based on the big data on a user's device. Now that we have an idea of how consumers are using big data, let's take a look at how big data is impacting business. In 2011, McKinsey & Company said that big data was going to become the key basis of competition supporting new waves of productivity growth and innovation. In 2013, UPS announced that it was ...