Poojith Reddy

Poojith Reddy

About

Detail

Senior Data Engineer/Snowflake Engineer at CVS Health
Plano, Texas, United States

Contact Poojith regarding: 

work
Full-time jobs

Timeline


work
Job

Résumé


Jobs verified_user 0% verified
  • CVS Health
    Senior Data Engineer/Snowflake Engineer
    CVS Health
    Feb 2022 - Current (3 years 5 months)
    I have a strong background in developing data pipelines and processing large datasets in the big data ecosystem. I have also developed custom multi-threaded Java and Sqoop jobs for ingesting data from FTP servers and data warehouses as well. Additionally, I have worked extensively with the Scala and Spark to perform data cleansing, aggregation, and preparation for machine learning and reporting teams. I have experience troubleshooting Spark applications to improve their error tolerance and have integrated Jenkins with various tools for CI/CD automation. Furthermore, I have worked with Docker and Kubernetes for containerization and orchestration, and have hands-on experience with AWS and Azure cloud services. I have developed REST APIs using
  • O
    Cloud Data Engineer
    Old National Bank
    Dec 2020 - Feb 2022 (1 year 3 months)
    I have been actively involved in working with various components of the Azure cloud platform, including HDInsight, Databricks, Data Lake, Blob storage, Data Factory, Synapse, SQL Database, and SQL Data Warehouse as well. I have used Azure Data Factory, SQL API, and MongoDB API to integrate data from MongoDB, MS SQL, and cloud services. Leveraging Linked Services, Datasets, and Pipelines, I extracted, transformed, and loaded data from multiple sources such as Azure SQL, Blob storage, and Azure SQL Data Warehouse, creating ADF pipelines. Utilizing Databricks and Spark, I performed information purging and applied changes, often relying on Databricks notebooks for interactive analysis. Implementing versatile microservices to handle concurrency
  • B
    Big Data Engineer/Hadoop Developer
    Burlington Stores Inc
    Mar 2020 - Dec 2020 (10 months)
    I have configured Spark streaming with Kafka, storing data in HDFS and HBase, and working on data ingestion pipelines using Talend and bash scripting. I have used Spark-Streaming APIs to perform transformations and actions on Kafka data, persisting it into Cassandra. In addition, I have designed and developed data integration programs using Hadoop and HBase, worked with Flume and NiFi for loading log files, and utilized various Spark transformations for data cleansing. I am also proficient in Python, Django, and have experience with Spark Streaming APIs for real-time data processing. Furthermore, I have expertise in working with Hive, Impala, Kudu, and optimizing Hive queries using Spark. I have written Templates forAWS infrastructure as a
  • F
    Data Engineer
    Foot Locker
    Mar 2017 - Dec 2019 (2 years 10 months)
    I have extensive experience in installing, configuring, and maintaining Apache Hadoop clusters, including tools like Hive, Pig, Zookeeper, and Sqoop. I have implemented advanced features in Hive such as partitioning, dynamic partitions, and buckets. I have also set up and configured Sqoop for seamless data import and export between Hive and relational databases. I have successfully managed large Hadoop environments, focusing on cluster setup, performance tuning and monitoring. I have optimized Hadoop cluster components to achieve high performance by monitoring and analyzing MapReduce job executions. I have also developed data pipelines using Flume, Sqoop, Pig, and Java MapReduce. Using Python and SAS, I have extracted, transformed, and load
  • Consumers Energy
    Data Engineer & Reporting Analyst
    Consumers Energy
    Jan 2015 - Mar 2017 (2 years 3 months)
    I have performed data transformations using Pig and created Sqoop jobs to import data from SQL, Oracle, and Teradata into HDFS. I also have experience in creating Hive tables to push data to MongoDB and writing complex aggregate queries in MongoDB for generating reports. Additionally, I developed scripts using Oozie to run scheduled batch cycles and present data for reports. I developed a big data ingestion framework that processed multi TB data, including data quality checks and transformations, and stored it in efficient formats like Parquet, loaded into Amazon S3 using Spark Scala API and Spark. I have implemented automation, traceability, and transparency using Python, Java, Hadoop streaming, Apache Spark, Spark SQL, Scala, Hive, and Pi
Education verified_user 0% verified
  • V
    Bachelor of Technology - BTech, Computer Science (Data Analytics)
    Vel Tech Rangarajan Dr Sagunthala R D Institute of Science Technology Chennai