Overview
Position: Cloud Data Engineer- GCP
Location: Nagpur/Pune/Chennai/Bangalore
Type of Employment: Full-time
InfoCepts is a global leader of end-to-end data and analytics solutions with nearly 20 years’ experience enabling customers to derive value from a variety of data-driven capabilities. Unique among its peers, InfoCepts operates with the scale of a global consulting firm, yet the expertise of a niche partner.
At InfoCepts, you’ll be challenged to think innovatively, while growing your own personal and professional skills designed with the future in mind. We have more than 1200 global professionals working on cutting edge technology solutions with a single mission – transforming our customers' journey with data-driven modernization.
- InfoCepts has been recognized as Gartner Peer Insights ™ Customers' Choice for two consecutive years in 2020 and 2021 which recognizes our best-in-class services that help our customers approach any data analytics problem.
- We are certified by Great Place to Work®, India in 2021 and recognizes our high-trust and high-performance work culture, and great employee experience in the industry.
- Our award-winning reusable solutions approach is well recognized in the D&A industry and lets our associates leverage our collective and proven consultative expertise, accelerate solution delivery through automation, enable faster time to value with reusable toolkits – all while delivering exceptional customer experience.
Our success has been unique, and we are looking for professionals who are enthusiastic and passionate about data & analytics, delivering differentiated experiences, and solving real world problems for our global customers. This position works with the competency and delivery team to execute on the assigned responsibilities.
Purpose of the Position:
As a Cloud Data Engineer, this role is ideal for candidates passionate about data and committed to helping organizations solve their AI & Analytics challenges. You will support clients in their data and analytics journey by designing scalable data pipelines, improving data consistency, and building infrastructure to support cloud-based analytics platforms on Google Cloud Platform (GCP).
Key Result Areas and Activities:
- Develop and maintain data pipelines using GCP services like Cloud Functions, Dataflow, and BigQuery
- Assist in designing and implementing ETL/ELT workflows for structured and semi-structured data
- Contribute to building and maintaining data models and schemas using best practices
- Support in the development and optimization of Spark jobs using PySpark or Scala
- Collaborate with senior engineers to perform data quality checks and resolve data inconsistencies
Essential Skills:
- Strong experience in Python and Spark
- Solid understanding of data engineering concepts and GCP architecture
- Hands-on experience with GCP services:
Cloud Storage (GCS)
Cloud Functions / Cloud Run
Pub/Sub, BigQuery, Dataflow, Cloud Composer
- Experience with BigQuery and other cloud DWHs (Snowflake, Redshift)
- Familiarity with CI/CD pipelines and Git for version control
- Experience designing and implementing ETL workflows with complex transformations (SCDs, deduplication, aggregation, etc.)
- Ability to maintain scalable data pipelines and build new API/data integrations for growing data volumes
Desirable Skills:
- Familiarity with streaming platforms like Kafka or Google Cloud Data Streams
- Working knowledge of Linux and ETL tools such as Informatica or Talend
Qualifications:
- Bachelor’s degree in computer science, Engineering, or related field (Master’s degree is a plus)
- 5 to 7 years of IT experience, including 3+ years of hands-on experience with Google Cloud Platform (GCP)
- Strong understanding of GCP services such as BigQuery, Cloud Storage, Dataflow, Cloud Functions, Cloud Composer, and Pub/Sub
India
4 to 7 years