J
Jaisrinadh Kalluri
Jaisrinadh Kalluri
About
Detail
Data Engineer
SUMMARY • Experienced Data Engineer with 6+ Years of professional experience in IT which includes comprehensive experience in working with Hadoop Ecosystem Components, Hive, HBase, Spark, Cloud Engineering Platforms (AWS, Azure), Python, Scala, Unix, DBT, Rest API's data and ETL data processing. • Experience with working different big data distributions like Cloudera, Hortonworks and MapR. • Implemented various ETL transformations with MapReduce and Pig. • Experienced with the Implementation of optimized data pipelines in Hive and Spark for various data transformations. • Strong expertise developing most efficient Spark applications for performing various kinds types of data transformations such as data cleansing, de-normalization, various types of joins, and data aggregation. • Hands-on experience with data mining processes, including implementing complex business logic, optimizing queries using Hive-QL, and controlling data distribution with partitioning and bucketing techniques to improve efficiency. • Experience working with Huge Complex Data Sets, as well as identifying patterns and insights in structured and unstructured data. • Experience working with Hive data, including extending the Hive library with custom UDF's to query data in non-standard formats. • Experience in migrating the data from HDFS to Relational Database Systems using Sqoop and vice-versa. • Extensive knowledge of data import and export utilizing stream processing platforms such as Flume and Kafka. • Experience with physical and logical data modelling, dimensional modelling utilizing Star and Snowflake schemas, data marts, OLAP, FACT, and Dimensions tables. • Excellent knowledge of NoSQL databases, as well as hands-on experience building applications for No SQL databases such as HBase, Cassandra, and MongoDB. • Expert in working with various optimized data formats such as Orc, Parquet, Avro, and Sequence files. • Implemented optimized data pipelines in Hive and Spark for various data transformations. • Strong experience in writing applications using python using different libraries like Pandas, NumPy, SciPy, Matplotlib etc. • Expert at working with AWS services such as S3, RDS, Redshift, Elastic Cache, and Dynamo DB. • Experience in configuring workflows for managing and scheduling Hadoop jobs using Direct Acyclic Graph (DAG) of actions with control flows in Apache Airflow and the Oozie workflow engine. • Experience in database design using PL/SQL to write Stored Procedures, Functions, Triggers and strong experience in writing complex queries for Oracle. • Strong experience in working with UNIX/LINUX environments, and expert at writing automation scripts in Python, Bash, and Shell. • Good knowledge working on components such as Azure Databricks, Azure Data Factory, Blob Storage and Azure Data Lake. • Excellent Knowledge on DOCKER components like Docker Engine, Docker Hub, Machine, Compose and Docker Registry. • Exposed application metrics and logs using tools like Kibana and Grafana. • Experience in working on different BI tools like Tableau, QlikView and Power BI.