Staff Engineer - Data at Safe Security | Torre

Staff Engineer - Data

You'll architect the future of autonomous cyber risk management at scale.
Emma highlights
This highlight was written by Emma’s AI. Ask Emma to edit it.
Full-time

Legal agreement: Employment

Provide your expected compensation while applying
location_on
Bengaluru, Karnataka, India
skeleton-gauges
You have opted out of job matches in .
To undo this, go to the 'Skills and Interests' section of your preferences.
Review preferences
Posted 5 months ago

Requirements and responsibilities


About SAFE SecurityAt SAFE Security, our mission is bold and ambitious: We Will Build CyberAGI — a super-specialized system of intelligence that autonomously predicts, detects, and remediates threats. This isn’t just a vision—it’s the future we’re building every day, with the best minds in AI, cybersecurity, and risk. At SAFE, we empower individuals and teams with the freedom and responsibility to align their goals, ensuring we all move towards this goal together.We operate with radical transparency, autonomy, and accountability—there’s no room for brilliant jerks. We embrace a culture-first approach, offering an unlimited vacation policy, a high-trust work environment, and a commitment to continuous learning. For us, Culture is Our Strategy—check out our Culture Memo to dive deeper into what makes SAFE unique.We’re looking for a Staff Data Engineer who thrives on solving complex data challenges at scale. You’ll be the technical force multiplier, leading the design of data platforms, pipelines, and lakehouse architectures that fuel AI-driven cyber risk quantification globally. If you’ve been waiting for a role where you can set data strategy, lead bold ideas, and shape large-scale data ecosystems—this is it.What You’ll DoBe the Data Tech Leader: Mentor engineers, champion data engineering best practices, and raise the bar for technical excellence across the org.Architect at Scale: Design and lead petabyte-scale data ingestion, processing, and analytics platforms using Snowflake, Apache Spark, Iceberg, Parquet, and AWS-native services.Own the Data Flow: Build streaming and batch pipelines handling billions of events daily, orchestrated through Apache Airflow for reliability and fault tolerance.Set the Standards: Define frameworks for data modeling, schema evolution, partitioning strategies, and data quality/observability for analytics and AI workloads.Code Like a Pro: Stay hands-on, writing high-performance data processing jobs in Python, SQL, and Scala, and conducting deep-dive reviews when it matters most.Master the Lakehouse: Architect data lakes and warehouse solutions that balance cost, performance, and scalability, leveraging AWS S3 and Snowflake.Solve Complex Problems: Elegantly and efficiently debug and optimize long-running jobs, data skew, and high-volume ETL bottlenecks.Collaborate and influence: Work with the Product, AI/ML, and Platform teams to ensure that data solutions directly power real-time cyber risk analytics.Innovate Constantly: Evaluate and introduce emerging data technologies (e.g., Flink, Druid, Rockset) to keep SAFE at the forefront of data engineering innovation.What We’re Looking For8+ years of experience in data engineering, with a proven track record of designing and scaling distributed data systems.Deep expertise in big data processing frameworks (Apache Spark, Flink) and workflow orchestration (Airflow).Strong hands-on experience with data warehousing (Snowflake) and data lakehouse architectures (Iceberg, Parquet).Proficiency in Python, SQL, Scala, Go/Nodejs with an ability to optimize large-scale ETL/ELT workloads.Expertise in real-time data ingestion pipelines using Kafka or Kinesis, handling billions of events daily.Experience operating in cloud-native environments (AWS) and leveraging services like S3, Lambda, ECS, Glue, and Athena.Strong understanding of data modeling, schema design, indexing, and query optimization for analytical workloads.Proven leadership in mentoring engineers, driving architectural decisions, and aligning data initiatives with product goals.Experience in streaming architectures, CDC pipelines, and data observability frameworks.Ability to navigate ambiguous problems, high-scale challenges, and lead teams toward innovative solutions.Proficient in deploying containerized applications (Docker, Kubernetes, ECS).Familiarity with using AI Coding assistants like Cursor, Claude Code, or GitHub CopilotPreferred QualificationsExposure to CI/CD pipelines, automated testing, and infrastructure-as-code for data workflows.Familiarity with real-time analytics engines (Druid, Pinot, Rockset) or machine learning data pipelines.Contributions to open-source data projects or thought leadership in the data engineering community.Prior experience in cybersecurity, risk quantification, or other high-scale SaaS domainClosingIf you’re passionate about cyber risk, thrive in a fast-paced environment, and want to be part of a team that’s redefining security—we want to hear from you! 🚀AI in HiringWe may use artificial intelligence (AI) tools to support parts of the hiring process, such as reviewing applications, analyzing resumes, or assessing responses. These tools assist our recruitment team but do not replace human judgment. Final hiring decisions are ultimately made by humans. If you would like more information about how your data is processed, please contact us.
Optionally, you can add more information later (benefits, pre-screening questions, etc.)
check_circle

Payment confirmed

A member of the Torre team will contact you shortly

In the meantime, continue adding information to your job opening.