Free cookie consent management tool by TermsFeed Senior Data Engineer | Antal Tech Jobs
Back to Jobs
3 Days ago

Senior Data Engineer

decor
Information Technology
Full-Time
Wadhwani Entrepreneurship

Overview

The Role Context

We are seeking a highly motivated and detail-oriented individual to join our team as a Data Engineer with experience in the designing, constructing, and maintaining the architecture and infrastructure necessary for data generation, storage and processing.

Key Responsibilities

  • Data Architecture Design: Design, develop, and maintain scalable data pipelines and infrastructure for ingesting, processing, storing, and analyzing large volumes of data efficiently. This involves understanding business requirements and translating them into technical solutions.
  • Data Integration: Integrate data from various sources such as databases, APIs, streaming platforms, and third-party systems. Should ensure the data is collected reliably and efficiently, maintaining data quality and integrity throughout the process as per the Ministries/government data standards.
  • Data Modeling: Design and implement data models to organize and structure data for efficient storage and retrieval. They use techniques such as dimensional modeling, normalization, and denormalization depending on the specific requirements of the project.
  • Data Pipeline Development/ ETL (Extract, Transform, Load): Develop data pipeline/ETL processes to extract data from source systems, transform it into the desired format, and load it into the target data systems. This involves writing scripts or using ETL tools or building data pipelines to automate the process and ensure data accuracy and consistency.
  • Data Quality and Governance: Implement data quality checks and data governance policies to ensure data accuracy, consistency, and compliance with regulations. Should be able to design and track data lineage, data stewardship, metadata management, building business glossary etc.
  • Data lakes or Warehousing: Design and maintain data lakes and data warehouse to store and manage structured data from relational databases, semi-structured data like JSON or XML, and unstructured data such as text documents, images, and videos at any scale. Should be able to integrate with big data processing frameworks such as Apache Hadoop, Apache Spark, and Apache Flink, as well as with machine learning and data visualization tools.
  • Data Security: Implement security practices, technologies, and policies designed to protect data from unauthorized access, alteration, or destruction throughout its lifecycle. It should include data access, encryption, data masking and anonymization, data loss prevention, compliance, and regulatory requirements such as DPDP, GDPR, etc.
  • Database Management: Administer and optimize databases, both relational and NoSQL, to manage large volumes of data effectively.
  • Data Migration: Plan and execute data migration projects to transfer data between systems while ensuring data consistency and minimal downtime.
  • Performance Optimization: Optimize data pipelines and queries for performance and scalability. Identify and resolve bottlenecks, tune database configurations, and implement caching and indexing strategies to improve data processing speed and efficiency.
  • Collaboration: Collaborate with data scientists, analysts, and other stakeholders to understand their data requirements and provide them with access to the necessary data resources. They also work closely with IT operations teams to deploy and maintain data infrastructure in production environments.
  • Documentation and Reporting: Document their work including data models, data pipelines/ETL processes, and system configurations. Create documentation and provide training to other team members to ensure the sustainability and maintainability of data systems.
  • Continuous Learning: Stay updated with the latest technologies and trends in data engineering and related fields. Should participate in training programs, attend conferences, and engage with the data engineering community to enhance their skills and knowledge.

Desired Skills/ Competencies

  • Education: A Bachelor's or Master's degree in Computer Science, Software Engineering, Data Science, or equivalent with at least 5 to 10 years of experience.
  • Database Management: Strong expertise in working with databases, such as SQL databases (e.g., MySQL, PostgreSQL) and NoSQL databases (e.g., MongoDB, Cassandra).
  • Big Data Technologies: Familiarity with big data technologies, such as Apache Hadoop, Spark, and related ecosystem components, for processing and analyzing large-scale datasets.
  • ETL Tools: Experience with ETL tools (e.g., Apache NiFi, Talend, Apache Airflow, Talend Open Studio, Pentaho, Infosphere) for designing and orchestrating data workflows.
  • Data Modeling and Warehousing: Knowledge of data modeling techniques and experience with data warehousing solutions (e.g., Amazon Redshift, Google BigQuery, Snowflake).
  • Data Governance and Security: Understanding of data governance principles and best practices for ensuring data quality and security.
  • Cloud Computing: Experience with cloud platforms (e.g., AWS, Azure, Google Cloud) and their data services for scalable and cost-effective data storage and processing.
  • Streaming Data Processing: Familiarity with real-time data processing frameworks (e.g., Apache Kafka, Apache Flink) for handling streaming data.
  • Familiar with Python Programming and Prompt Engineering

KPIs

  • Data Pipeline Efficiency: Measure the efficiency of data pipelines in terms of data processing time, throughput, and resource utilization. KPIs could include average time to process data, data ingestion rates, and pipeline latency.
  • Data Quality Metrics: Track data quality metrics such as completeness, accuracy, consistency, and timeliness of data. KPIs could include data error rates, missing values, data duplication rates, and data validation failures.
  • System Uptime and Availability: Monitor the uptime and availability of data infrastructure, including databases, data warehouses, and data processing systems. KPIs could include system uptime percentage, mean time between failures (MTBF), and mean time to repair (MTTR).
  • Data Storage Efficiency: Measure the efficiency of data storage systems in terms of storage utilization, data compression rates, and data retention policies. KPIs could include storage utilization rates, data compression ratios, and data storage costs per unit.
  • Data Security and Compliance:
Share job
Similar Jobs
View All
1 Day ago
.Net Developer
Information Technology
Position Title: .NET DeveloperCTC: 8 LPAExp: 2YrLocation: Mumbai ThaneWorking Mode: OnsitePrimary Skill: .NET development, C#, ASP.NET Core/.NET Framework, API development and integration, LINQ, SQL databases, GitQualification: Any GraduationKey Res...
decor
1 Day ago
Full Stack Developer in Chennai, Bangalore
Information Technology
Key Responsibilities Build and optimize APIs using Supabase (or Node.js where needed). Implement push notifications using Firebase Cloud Messaging for both web & mobile. Develop the call scheduling and status system (Pending, Connected, Cancelled...
decor
1 Day ago
Serosoft - Senior Software Engineer - Java Technologies
Information Technology
Job Category : Technical.Department : Engineering.Job Location : Indore.Experience : 6+ years.About The RoleWe are seeking a highly skilled Senior Software Engineer (Java) to join our team.In this role, you will design, develop, and deliver robu...
decor
1 Day ago
Java Fullstack Developer -Azure
Information Technology
This job is with Kyndryl, an inclusive employer and a member of myGwork – the largest global platform for the LGBTQ+ business community. Please do not contact the recruiter directly. Who We AreAt Kyndryl, we design, build, manage and modernize the ...
decor
1 Day ago
Senior .NET Developer
Information Technology
Position Title: Senior .NET DeveloperCTC: 15 LPAExp:4+ YrLocation: Mumbai ThaneWorking Mode: OnsitePrimary Skill: .NET development, C#, ASP.NET Core/.NET Framework, API development and integration, LINQ, SQL databases, Git, Restful APIsQualification...
decor
1 Day ago
Software Engineer, QA
Information Technology
Who We AreMotive empowers the people who run physical operations with tools to make their work safer, more productive, and more profitable. For the first time ever, safety, operations and finance teams can manage their drivers, vehicles, equipment, ...
decor
1 Day ago
Software Engineer
Information Technology
The purpose of this role is to develop required software features, achieving timely delivery in compliance with the performance and quality standards of the company.Job Description:Job Responsibilities: Works independently with a broader overview on...
decor
1 Day ago
Junior Web Developer in Mumbai
Information Technology
  • Bangalore, Karnataka, India
Key Responsibilities Develop and maintain responsive web applications using Angular, Laravel, and WordPress Implement modern styling with Tailwind CSS and Bootstrap Customize and extend WordPress themes and plugins Translate Figma designs into c...
decor

Talk to us

Feel free to call, email, or hit us up on our social media accounts.
Social media