Data Engineer - Development (AWS Data Architect & Pipeline Operations Engineer)
Location
Bengaluru, Karnataka, IN / Remote (IN)
Job Type
Full-Time
Experience Level
Mid-Level
Salary Range
₹1.8M - ₹3M INR
Job Description
We are looking for a skilled Data Engineer to design, build, and optimize scalable data pipelines and cloud-based data architectures. You will play a key role in enabling data-driven decision-making by ensuring reliable, high-quality data flows across systems. 🎯 Key Responsibilities 🔹 Data Pipeline Development Design and build scalable ETL/ELT pipelines Develop batch and real-time data pipelines using tools like Spark, Kafka, or similar Ensure seamless data ingestion from APIs, databases, logs, and streaming sources 🔹 Data Architecture & Warehousing Design and maintain data models (star schema, data vault, dimensional models) Work with cloud data warehouses like Redshift, BigQuery, or Snowflake Build scalable data lake and warehouse solutions 🔹 AWS & Cloud Engineering Develop and manage data solutions using AWS services (S3, Glue, Lambda, EMR, etc.) Optimize infrastructure for performance, scalability, and cost efficiency Implement cloud-native architectures and best practices 🔹 Data Quality & Governance Implement data validation, monitoring, and quality checks Ensure data lineage, schema evolution, and governance standards Maintain data integrity across systems 🔹 Pipeline Operations & DevOps Deploy pipelines using Docker/Kubernetes or managed services Build CI/CD pipelines for data workflows Monitor systems using tools like Prometheus, Grafana, or ELK stack 🔹 Collaboration Work closely with Data Scientists, Analysts, and Product teams Translate business requirements into scalable data solutions Document pipeline architecture and workflows 🛠️ Required Skills Strong programming skills in Python / Java / Scala Expertise in SQL & relational databases (PostgreSQL, MySQL) Hands-on experience with AWS or other cloud platforms Experience with data pipeline tools (Airflow, Prefect, Dagster) Knowledge of big data technologies (Spark, Kafka) Experience with data warehousing solutions ⭐ Good to Have Experience with dbt, Fivetran, or Matillion Knowledge of real-time streaming systems Exposure to ML/AI pipelines or data platforms Understanding of data observability & monitoring tools
About cyble
Cyble is the world’s first intelligence-driven, AI-native security platform that brings together detection, protection, and remediation across the entire digital attack surface. The company was born in Melbourne (Australia) and has since gone global, serving customers in more than 35 countries, with its headquarters now in Cupertino, California. Backed by Y Combinator and other leading venture capital firms, Cyble is on a mission to make the world a safer place by rethinking how cybersecurity works — through an AI-native platform built for predictive and autonomous defense. By combining deep intelligence, automation, and real-time response, Cyble helps organizations stay ahead of threats and build stronger, more resilient digital environments.
Connections
Sai Charan
Senior Developer
Kalpana Sharma
Team Lead
Rahul Patel
Full Stack Developer
Priya Singh
Frontend Developer
Connect with professionals in your network
Skill Match Analysis
??% skills matched (?? of skills)
💡 This is keyword matching for reference only. Your actual match score uses AI semantic analysis.
Login to see your score