Overview
The CoinDCX Journey: Building the Future of FinanceAt CoinDCX, our mission is clear - to make crypto and blockchain accessible to every Indian and enable them to participate in the future of finance.
As India’s first crypto unicorn valued at $2.45B, we are reshaping the financial ecosystem by building safe, transparent, and scalable products that power adoption at scale.
We believe that change starts together. It begins with bold ideas, relentless execution and people who want to build what’s next.
If you’re driven by purpose and thrive in environments where your work defines the next chapter of an industry, you’ll feel right at home here.
About The Role
- This role combines core data engineering excellence with optional exposure to financial correctness & reconciliation pipelines.
- You will design, build, and optimize data systems that handle high-volume, high-variety datasets , power analytics, reporting, financial systems, and operational products.
- The work will be across ingestion, transformation, data modeling, data marts, and workflow orchestration, ensuring reliability, scalability, and data correctness across the entire platform.
- Design and build scalable, high-performance data platforms leveraging Databricks, Lakehouse architectures, Spark ETL, and Kafka-based streaming pipelines.
- Build & Optimize Large-Scale Data Pipelines
- Own end-to-end ETL/ELT pipelines using Databricks (PySpark, Spark SQL).
- Build streaming and batch workflows processing billions of events per day.
- Implement efficient ingestion frameworks using Kafka/MSK, Auto Loader, CDC.
- Optimize jobs for performance, cost efficiency, cluster utilisation, and scalability.
- Ensure pipelines meet defined SLAs for latency, throughput, and freshness.
- Own Lakehouse Design & Data Modeling
- Design bronze → silver → gold layers using Delta Lake best practices.
- Build curated data marts for analytics, BI, finance, risk, and product teams.
- Implement dimensional models, fact/event stores, and conformed dimensions.
- Drive schema governance, catalog organization, versioning, and lineage.
- Build High-Quality Data Integrations
- Integrate with internal microservices, third-party APIs, streaming sources, and databases.
- Develop connectors/pipelines for structured, semi-structured, and unstructured data.
- Implement robust idempotency, incremental ingestion, and change data capture.
- Build scalable workflows for multi-system integration (custody, payments, trading, CRM, etc.)
- Workflow Orchestration & Reliability
- Build highly reliable workflows using Airflow/Databricks Workflows/Lambda/Step Functions.
- Implement retry logic, DLQs, backfill strategies, and pipeline auto-recovery.
- Maintain >99.5% pipeline uptime for critical systems.
- Data Quality, Observability & Governance
- Implement DQ checks: schema, completeness, freshness, referential integrity.
- Build monitoring and alerting for ingestion lag, pipeline failures, and cost anomalies.
- Own documentation, runbooks, dashboards (Datadog/CloudWatch/Lakehouse monitoring).
- Ensure compliance with internal governance (S3 structure, catalog rules, PII handling).
- Cross-Functional Collaboration
- Work closely with Analytics, Product, Finance, Ops, Risk, and Engineering teams.
- Translate requirements into scalable data pipelines and optimized datasets.
- Participate in design reviews, code reviews, architecture discussions.
- Mentor SDE-1/SDE-2 engineers and uplift engineering excellence.
Have strong hands-on experience in:
Data Engineering Core
- Databricks (PySpark, Spark SQL, Delta Lake) – mandatory
- Kafka/MSK / Streaming pipelines
- PySpark, Python — expert level
- Spark SQL optimisation & query tuning
- ETL engineering for large-scale workflows
- Designing data marts, fact tables, dimensions, business layers
- Strong experience in data warehouse concepts
- AWS (S3, Glue, Lambda, EMR/Databricks, DMS, IAM, CloudWatch)
- Cluster management, autoscaling, job tuning
- CI/CD for data pipelines
- REST APIs, event ingestion, CDC, RDBMS ingestion (Postgres/MySQL)
- Handling schema evolution & incremental loads
- 4-6 key pipelines redesigned or optimized for performance/cost
- High-reliability ingestion for multiple systems in production
- Data marts operational with measurable improvement in query performance
- Zero severe data-quality issues attributed to owned systems
- Documented and automated workflows replacing manual processes
Here’s what your journey with us looks like:
- Application Review – We assess for skills, alignment, and intent
- Recruiter Connect – A short conversation to understand you better
- Functional Round(s) – Deep dive into your approach, craft, and problem-solving
- Assignment / Simulation Round – A take-home task or live problem-solving exercise to understand how you think and execute in real scenarios
- Culture & Values Discussion – A conversation to understand our ways of working and how you thrive best
- Founder Conversation (Optional) – For certain roles and senior levels, you may meet our founders to explore strategic alignment and long-term fit
We believe the best ideas emerge when people build together. Collaboration, speed and trust come alive when teams share the same space.
With this belief, we operate as a work-from-office organisation. This role is based out of our Bengaluru office, where energy, alignment and innovation move in real time.
Perks That Empower You
We believe great people deserve great experiences.
- Design Your Own Benefits: Flexible perks to match your lifestyle
- Unlimited Wellness Leaves: Rest and recharge as you need
- Mental Wellness Support: Access to therapy and wellness resources
- Learning Sessions: Bi-weekly learning and growth opportunities
If you’re looking for a role that gives you direct access to high-stakes decisions, deep impact and a chance to build the future of finance, this is it.
Join CoinDCX and help us make crypto accessible to every Indian, together.