Free cookie consent management tool by TermsFeed Senior Big Data Engineer (Python | AWS) | Antal Tech Jobs
Back to Jobs
1 Week ago

Senior Big Data Engineer (Python | AWS)

decor
Bangalore, Karnataka, India
Information Technology
Full-Time
BEO Software Private Limited

Overview

Responsibilities

Role Summary

Own the data platform that powers clinician discovery, credential unification, and care-quality analytics. Design resilient, low-latency ingestion and transformation at scale (batch + streaming) with GDPR-by-design. Your work underpins search, matching, and ML features in our telemedicine platform across Germany.

Key Responsibilities

  • Design and operate AWS-native data lakehouse: Amazon S3 + Lake Formation (governance),

Glue/Athena (ELT), and optional Amazon Redshift for warehousing.

  • Build high-throughput ingestion and CDC pipelines from partner APIs, files, and databases

using EventBridge, SQS/SNS, Kinesis/MSK, AWS DMS, and Lambda/ECS Fargate.

  • Implement idempotent upserts, deduplication, and delta detection; define source-of-truth

governance and survivorship rules across authorities/insurers/partners.

  • Model healthcare provider data (DDD) and normalize structured/semi-structured payloads

(JSON/CSV/XML, FHIR/HL7 if present) into curated zones.

  • Engineer vector-aware datasets for clinician/patient matching; operate pgvector on Amazon.

Aurora PostgreSQL or use OpenSearch k-NN for hybrid search.

  • Establish data quality (freshness, accuracy, coverage, cost-per-item) with automated checks.

(e.g., Great Expectations/Deequ) and publish KPIs/dashboards.

  • Harden security & privacy: IAM least-privilege, KMS encryption, Secrets Manager, VPC

endpoints, audit logs, pseudonymised telemetry; enforce GDPR and right-to-erasure.

  • Observability-first pipelines using OpenTelemetry (ADOT), CloudWatch, X-Ray; DLQ

handling, replay tooling, resiliency/chaos tests; SLOs and runbooks.

  • Performance tuning for Aurora PostgreSQL (incl. indexing, partitioning, vacuum/analyze)

and cost-aware Spark (EMR/Glue) jobs.

  • CI/CD for data (Terraform/CDK, GitHub Actions/CodeBuild/CodePipeline); test automation

(pytest/DBT) and blue/green or canary for critical jobs.

Desired Candidate Profile

  • 6+ years in data engineering at scale; proven delivery in production systems (regulated

domains a plus).

  • Expertise in Python and SQL; hands-on with Spark (EMR/Glue) and stream processing

(Kinesis/MSK/Flink/Spark Streaming).

  • Deep AWS experience across S3, Glue, Athena, Redshift or Aurora PostgreSQL, Lake

Formation, DMS, Lambda/ECS, Step Functions, EventBridge, SQS/SNS.

  • PostgreSQL mastery incl. query planning, indexing, and performance tuning; familiarity with

pgvector or OpenSearch vector search.

  • Strong grasp of idempotency, deduplication, CDC, schema evolution, SCDs, and contract

testing for data products.

  • Observability (OpenTelemetry), CI/CD, and IaC (Terraform/CDK) best practices; strong

incident response and on-call hygiene.

  • Security-by-design mindset: data minimization, encryption, secrets, PII-safe logging; working knowledge of GDPR and auditability.
  • Effective communicator across Product, Platform, Data Science, and Compliance; pragmatic,metrics-driven delivery.

Nice to Have

  • Experience with FHIR/HL7, German TI/ePrescription/ePA integrations.
  • DBT for transformations; OpenMetadata/Amundsen for catalog/lineage.
  • Go for high-throughput services; experience with Bedrock or SageMaker for embedding

generation.

How We Work & Benefits

  • API-first, clean architecture, and pairing culture; mandatory code reviews.
  • Remote-friendly with defined core hours; mission-led, patient-safety-first.
  • Ownership mindset: you build it, you run it (with sensible SLOs and error budgets).

Compliance & Notes

  • All PHI/PII processed within EU regions (e.g., eu-central-1); strict key management via AWS

KMS and Secrets Manager.

  • Right-to-erasure and lawful-basis handling embedded in data lifecycle (tombstones, purge

workflows, and immutable audit trails).

Back

Share job
Similar Jobs
View All
22 Hours ago
Data Engineer
Fintech
  • 3 - 5 Yrs
  • Mumbai
Data Engineer Mumbai | Full-Time  Experience: 3–6 Years Budget: Up to ₹27 LPA Industry: General Insurance (Digital-First Organization) We’re rebuilding insurance from the ground up digital-first, transparent, fast, and fair. No legacy te...
decor
1 Day ago
QA Manager
Fintech
  • 10 - 18 Yrs
  • Pune
Job Description We are seeking an experienced and dynamic QA Manager to lead our quality assurance team in delivering high-quality software products for our organization. The ideal candidate will have a strong background in manual and automation tes...
decor
1 Day ago
Database Administrator (DBA)
Information Technology
  • Bangalore, Karnataka, India
This role is for one of our clients Company Name: cloudtechner Seniority level: Mid-Senior level Min Experience: 5 years Location: Gurgaon, NCR JobType: full-time We are looking for an experienced and detail-oriented Database Administrator (DBA) to ...
decor
1 Day ago
Salesforce Data Engineer
Information Technology
  • Bangalore, Karnataka, India
DescriptionRole Summary :We are seeking a highly skilled Salesforce Data Engineer with deep expertise in the Salesforce platform and a strong focus on building and operating Salesforce Data Cloud (D360) solutions. The ideal candidate will design, int...
decor
1 Day ago
Business Analyst I
Information Technology
  • Bangalore, Karnataka, India
Through our dedicated associates, Conduent delivers mission-critical services and solutions on behalf of Fortune 100 companies and over 500 governments - creating exceptional outcomes for our clients and the millions of people who count on them. You ...
decor
1 Day ago
Associate Software Engineer - Test Automation (Infra)
Information Technology
  • Bangalore, Karnataka, India
Veeva Systems is a mission-driven organization and pioneer in industry cloud, helping life sciences companies bring therapies to patients faster. As one of the fastest-growing SaaS companies in history, we surpassed $2B in revenue in our last fiscal ...
decor
1 Day ago
Interesting Job Opportunity: Data Analyst - SQL/Python
Information Technology
  • Bangalore, Karnataka, India
DescriptionWe are seeking a skilled Data Analyst with strong expertise in Python, SQL, and Excel, coupled with a solid foundation in statistics and a good understanding of retail demand processes.The ideal candidate will be responsible for transformi...
decor
1 Day ago
EY - GDS Consulting - AI and DATA - GCP Data Engineer - Senior
Information Technology
  • Bangalore, Karnataka, India
At EY, you’ll have the chance to build a career as unique as you are, with the global scale, support, inclusive culture and technology to become the best version of you. And we’re counting on your unique voice and perspective to help EY become even b...
decor

Talk to us

Feel free to call, email, or hit us up on our social media accounts.
Social media