Data Ingestion: Files : Stage the files & use Hadoop/Hive CLI Database: Sqoop, no CDC for smaller tables but only for larger (10M+) , use -m option for large db dump, NiFi is another option Streaming: NiFi , Flume , Streamsets. NiFi is popular. File Ingestion: CSV into TEXTFILE : Overwrite: Move the file to HDFS, create an external TEXTFILE table on top of the HDFS location. You can also create the table and use "LOAD DATA INPATH LOCAL localpath OVERWRITE INTO tablename". This approach will be handy for internal tables where location is not specified and if you don't know the HDFS warehouse location where table is created. You can use LOAD DATA command for loading data from local as well as hdfs file. Append: You can still use "LOAD DATA INPATH ....INSERT INTO tablename" . create a temporary table using overwrite approach and then insert into original table from temporary table. Same approach will work for parti