Data Engineer
About Techversant:
Techversant is a global technology services and solutions company providing high-quality end-to-end IT Solutions and Services to its clients across the globe for more than 17 years. Techversant has been providing Technology services in the key areas of Digital Transformation, Understanding the Evolving Industrial Scenarios, Investment Priorities and Technology Disruption. We offers expertise and consultation in multiple technology domains like – ColdFusion, Python, Java, .NET, Node JS, Ruby on Rails, Golang, PHP, Artificial intelligence, Blockchain, IOS, Android, Flutter, React Native, Angular, React Vue, etc catering to various industry verticals globally.
What’s important to us:
We are looking for a skilled and experienced Data Engineer with over 4 years of professional experience in building, automating, and optimizing data pipelines and cloud-based architectures. The ideal candidate will have hands-on experience with cloud data services (AWS, Azure, or GCP) and CI/CD pipelines for deploying scalable, reliable, and secure data solutions.
The candidate will collaborate with cross-functional teams including data analysts, data scientists, and software engineers to design and maintain robust data infrastructure that supports analytics, AI/ML workflows, and enterprise reporting systems.
Key Responsibilities:
Design, build, and maintain end-to-end ETL/ELT pipelines using both on-premise and cloud-based technologies.
Architect and operate data storage and streaming solutions leveraging cloud-based services on AWS, Azure, or GCP.
Design and implement data ingestion and transformation workflows using Airflow, AWS Glue, or Azure Data Factory.
Develop and optimize data pipelines using Python and PySpark for large-scale distributed data processing.
Build data models — normalized, denormalized, and dimensional (Star/Snowflake) — for analytics and warehousing solutions.
Implement data quality, lineage, and governance using metadata management and monitoring tools.
Collaborate with cross-functional teams to deliver clean, reliable, and timely data for analytics and machine learning use cases.
Integrate CI/CD pipelines for data infrastructure deployment using GitHub Actions, Jenkins, or Azure DevOps.
Automate infrastructure provisioning using Infrastructure as Code (IaC) tools such as AWS CloudFormation or Terraform.
Monitor and optimize data processing performance for scalability, reliability, and cost-efficiency.
Enforce data security policies and ensure compliance with standards such as GDPR and HIPAA.
Must-Have Skills & Qualifications:
Education: Bachelor’s or Master’s degree in Computer Science, Information Technology, Data Engineering, or a related field.
Experience: Minimum 4 years of hands-on experience as a Data Engineer or in data-intensive environments.
SQL Expertise: Advanced proficiency in SQL for complex queries, joins, window functions, and performance tuning.
Analytical Databases: Experience working with Snowflake, Amazon Redshift, Google BigQuery, Azure Synapse, and PostgreSQL.
Query Optimization: Skilled in query optimization, indexing, and execution plan analysis for high-performance analytics workloads.
Programming: Proficient in Python and PySpark for data manipulation, automation, and pipeline orchestration.
Data Processing Frameworks: Strong understanding of Apache Spark (RDD, DataFrame, Spark SQL, optimization), Hive, Hadoop, and Flink for large-scale distributed data processing.
ETL/ELT Frameworks: Hands-on experience designing and maintaining pipelines using Airflow, AWS Glue, or Azure Data Factory.
Data Integration Patterns: Familiar with incremental loading, Slowly Changing Dimensions (SCD), Change Data Capture (CDC), and error handling in data pipelines.
Data Modeling: Expertise in data modeling, schema design, and building normalized, denormalized, and dimensional (Star/Snowflake) schemas.
Data Architecture: Strong understanding of Data Warehousing, Data Lakes, and Lakehouse architectures, including Delta Lake, ACID transactions, and partitioning strategies.
Cloud Platforms: Practical experience with major cloud ecosystems —
AWS: S3, Glue, Redshift, Athena, Lambda, Step Functions, EMR
Azure: Data Factory, Data Lake Gen2, Synapse, Databricks
Cloud Security: Experience managing IAM roles, access control, and encryption in cloud environments.
Pipeline Optimization: Skilled in optimizing data pipelines for performance, scalability, and cost-efficiency.
CI/CD and DevOps: Hands-on experience with CI/CD tools such as GitHub Actions, GitLab CI, or Azure DevOps.
Version Control: Proficient with Git and familiar with agile development practices.
Preferred Skills
Experience with containerization and orchestration.
Exposure to data cataloging and governance tools.
Experience with monitoring tools .
Familiarity with data APIs and microservices architecture.
Certification in cloud data engineering (e.g., AWS Certified Data Engineer, Azure Data Engineer Associate, or GCP Professional Data Engineer).
Experience supporting machine learning and analytics pipelines.
Soft Skills:
Strong analytical and problem-solving mindset.
Excellent communication and documentation skills.
Ability to work collaboratively in a cross-functional, fast-paced environment.
Strong attention to detail with a focus on data accuracy and reliability.
Eagerness to learn and adopt emerging data technologies.
What Company Offers:
Excellent career growth opportunities and exposure to multiple technologies.
Fixed weekday day schedule, meaning, you’ll have your weekends off!
Family Medical Insurance.
Unique leave benefits and encashment options based on performance.
Long term growth opportunities.
Fun family environment surrounded by experienced developers.
Various internal employee rewards programs based on Performance.
Opportunities for various other Bonus programs – for training hours taken, certifications, special value to business through idea and innovation.
Work life Balance – flexible work timings, early out Fridays, various social and cultural activities etc.
Company Sponsored International Tours.