Tag - Sqoop

Data Absorption (Ingestion)

Read more...

Data Engineering, Streaming & Cloud Solutions

The data tsunami The term Big Data is a buzzword now-a-days which has no specific formats and defies traditional storage systems. The data can be structured, semi-structured or unstructured and generated faster than ever. For example, information post over social media, video upload in YouTube etc. Machine generated data like data recorded from sensor and logs are also becoming part of big data having more magnitude then human generated data. Though Apache Hadoop is a popular framework for distributed storage and...

Read more...

Steering number of mapper (MapReduce) in sqoop for parallelism of data ingestion into Hadoop Distributed File System (HDFS)

To import data from most the data source like RDBMS, sqoop internally use mapper. Before delegating the responsibility to the mapper, sqoop performs few initial operations in a sequence once we execute the command on a terminal in any node in the Hadoop cluster. Ideally, in production environment, sqoop installed in the separate node and updated .bashrc file to append sqoop's binary and configuration which helps to execute sqoop command from anywhere in the multi-node cluster. Most of the...

Read more...

Transfer structured data from Oracle to Hadoop storage system

Using Apache's sqoop, we can transfer structured data from Relational Database Management System to Hadoop distributed file system (HDFS). Because of distributed storage mechanism in Hadoop Distributed File System (HDFS), we can store any format of data in huge volume in terms of capacity. In RDBMS, data persists in the row and column format (Known as Structured Data). In order to process the huge volume of enterprise data, we can leverage HDFS as a basic data lake. In this...

Read more...

Data Ingestion phase for migrating enterprise data into Hadoop Data Lake

The Big Data solutions helps to achieve valuable information to iron out the accurate strategic business decision. Exponential growth of digitalization, social media, telecommunication etc. are fueling enormous data generation everywhere. Prior to process of huge volume of data, we should have efficient data storage mechanism in a distributed manner to hold any form of data starting from structured to unstructured. Hadoop distributed file systems (HDFS) can be leveraged efficiently as data lake by installing on multi node cluster....

Read more...
Technical Capabilities

Irisidea’s Competencies

The technical leadership team is the core strength of Irisidea Technologies. Irisidea's business and technical experts not only work closely with clients to provide cost effective, robust solutions to their business needs, but also dedicate a part of their time regularly in research work and keeping themselves on the edge of technology. Each individual of this team is not only an expert in his domain, but he also strives to expand his knowledge horizon by learning new technologies that...

Read more...