B

BHANU PRAKASH

About

Detail

Dallas, Texas, United States

Contact BHANU regarding: 
Flexible work
Starting at USD45/hour
id_card
Internships
Starting at USD3K/month
connect_without_contact
Finding mentors
groups
Networking

Timeline


work
Job
school
Education
folder
Project

Résumé


Jobs verified_user 0% verified
  • Visa
    Senior Data Engineer
    Visa
    Sep 2024 - Current (1 year 8 months)
    As a Senior Data Engineer at Visa, I led the migration of over 20TB of financial and transactional data from Teradata and Redshift to Snowflake, utilizing Talend and AWS Glue to create scalable ingestion pipelines. I designed Snowflake multi-cluster virtual warehouses, implementing workload segregation and scaling strategies that improved query performance by 40%. My responsibilities included building robust ETL/ELT pipelines with Talend Big Data to process structured and unstructured data from S3, landing it into Redshift staging zones. I created and maintained Snowflake schemas (Star and Snowflake) to support compliance, claims, and reporting systems, ensuring audit readiness with GDPR-compliant masking. Additionally, I implemented Snowfl
  • Onix
    Data Engineer II
    Onix
    May 2023 - Aug 2024 (1 year 4 months)
    As a Data Engineer II at Onix, I designed and developed ETL pipelines using PySpark on AWS EMR, integrating seamlessly with legacy Hadoop/Hive systems for efficient historical batch processing. I successfully migrated data from SQL Server to Snowflake utilizing Python, SnowSQL, and AWS DMS, ensuring meticulous schema mapping, deduplication, and validation. My role involved engineering real-time ingestion workflows with Snowpipe and COPY commands, automating file arrival triggers from S3 into Snowflake landing zones. I refactored legacy SQL Server ETL logic into modular Snowflake SQL scripts and dbt models, significantly enhancing reusability, performance, and compatibility with source control. In addition, I implemented robust data profili
  • TCS
    Data Engineer I
    TCS
    Jan 2020 - Dec 2021 (2 years)
    As a Data Engineer I at TCS, I developed complex ETL workflows using Informatica PowerCenter and Informatica Cloud to process large datasets efficiently. I built and maintained Azure Data Factory (ADF) pipelines for ingesting SAP and Oracle data into Azure Data Lake Gen2, while utilizing Apache Kafka and Hadoop HDFS for real-time ingestion and batch storage of enterprise data. My role involved writing SQL and PL/SQL scripts to cleanse, transform, and validate incoming data before loading it into Redshift. I developed reusable mapping templates in Informatica for faster pipeline deployment and QA validation, and participated in the on-prem to cloud migration of ETL workflows to AWS and Azure platforms. Additionally, I worked on data integrat
  • A
    Data Engineer I
    Alsoft
    Jun 2018 - Mar 2021 (2 years 10 months)
    As a Data Engineer I at Alsoft, I built cloud-based data pipelines using AWS S3, Glue, Lambda, Redshift, and Kinesis for scalable ingestion, developing real-time ingestion solutions with Kinesis Data Streams and Lambda to process millions of events daily. I created ETL workflows in AWS Glue using Python and Spark for cleansing and structuring raw data from diverse sources, while also designing optimized Redshift schemas and distribution/sort keys to enhance analytical performance for reporting teams. My responsibilities included deploying processing jobs on EC2 with auto-scaling groups to reduce infrastructure costs during non-peak hours and building CloudFormation templates for infrastructure-as-code deployment across environments. I worke
Education verified_user 0% verified
  • Southern Arkansas University
    Masters in Computer and Information Science
    Southern Arkansas University
    Jan 2022 - Dec 2022 (1 year)
    Relevant courses: Cloud Computing, Machine Learning, Advanced Data Structures & Algorithms
  • S
    Snowflake Certified - SnowPro Core Certification
    Jan 2024 - Jun 2025 (1 year 6 months)
  • M
    Microsoft Certified – Azure Data Engineer Associate
    Jan 2020 - Dec 2022 (3 years)
Projects (professional or personal) verified_user 0% verified
  • G
    Global Commerce Backend | Spring Boot | gRPC | PostgreSQL | Redis | AWS EKS
    Jan 2024 - Jun 2024 (6 months)
    As a Global Commerce Backend Developer, I built a distributed commerce backend utilizing Saga/Outbox patterns for transaction consistency and Resilience4j for fault tolerance, achieving P99 latency of less than 250ms at over 5,000 requests per second. My role involved extensive API integration and database design, leveraging technologies such as Spring Boot, PostgreSQL, and Redis, while also employing AWS EKS for scalable deployments. Although the primary focus was on backend development, I collaborated closely with data engineering teams to design services that ingested data into data lakes and warehouses, effectively supporting ETL workflows and downstream data marts for reporting and analytics. This integration ensured that backend APIs
  • E
    Enterprise DevSecOps Toolchain | GitHub Actions | Argo CD | Terraform
    Aug 2023 - Dec 2023 (5 months)
    As an Enterprise DevSecOps Toolchain specialist, I designed a CI/CD automation framework that included SBOM generation, container image signing (cosign), and integrated security scans (SAST/DAST/SCA). I also exposed DORA metrics (lead time, deployment frequency, MTTR) in Grafana dashboards to enhance leadership visibility. Although the role title focused on DevSecOps, my work directly supported data engineering platforms and analytics systems. I collaborated with data engineering teams to design and deploy data lakes, data warehouses, and data marts on cloud platforms, ensuring reliable CI/CD for Spark, Hadoop, Redshift, and BigQuery workloads. I worked closely with engineers developing SQL- and Spark-based pipelines, integrating API-drive
  • S
    Streaming Analytics Pipeline | Kafka → Spark/Flink → Snowflake/BigQuery
    Jan 2023 - Jun 2023 (6 months)
    As a Streaming Analytics Pipeline Engineer, I built a robust real-time data ingestion pipeline utilizing Change Data Capture (CDC) with Debezium, stateful aggregations, and schema evolution handling. This initiative enabled the delivery of sub-minute analytics dashboards through Looker and GraphQL, while optimizing compute costs with auto-suspend policies. My work was part of a broader enterprise data engineering ecosystem, where I integrated streaming workflows with Hadoop-based data lakes for historical storage and batch processing, as well as with Redshift for downstream analytics and reporting alongside Snowflake and BigQuery. Additionally, I operated in a consulting data engineering context, collaborating directly with client stakehold
  • A
    AI Conversational Agent Platform | Python | Node.js | React | Kubernetes
    Sep 2019 - Jan 2020 (5 months)
    As an AI Conversational Agent Platform developer, I spearheaded the creation of a multi-channel conversational AI system that seamlessly integrated large language models (LLMs) with guardrails for chat, email, and voice automation, while also managing significant banking data engineering and governance responsibilities. I successfully scaled the system on Kubernetes, implementing auto-scaling and circuit breakers, which resulted in a 28% reduction in average handling time and improved first-contact resolution rates. My role involved designing and maintaining robust data pipelines in Databricks, where I enforced data governance standards, including data quality checks, lineage, and access controls, to ensure compliance with internal risk and