
Virtusa
PySpark Data Engineer
- Permanent
- Dubai, United Arab Emirates
- Experience 5 - 10 yrs
Job expiry date: 08/01/2026
Job overview
Date posted
24/11/2025
Location
Dubai, United Arab Emirates
Salary
AED 15,000 - 20,000 per month
Compensation
Salary only
Experience
5 - 10 yrs
Seniority
Senior & Lead
Qualification
Bachelors degree
Expiration date
08/01/2026
Job description
The role focuses on designing and implementing large-scale end-to-end Data Management and Analytics solutions across complex organizations, with deep expertise in data architecture, data strategy, and the transformation of traditional data warehouses into modern Big Data platforms. The position requires hands-on experience in big-data processing frameworks such as Hadoop, Presto, Tez, Hive, and Spark, with strong proficiency in PySpark, Python, Linux, GIT, and Jenkins. The engineer must utilize DW dimensional modeling techniques including star and snowflake schemas, slowly changing dimensions, role-playing dimensions, dimensional hierarchies, and data classification. Responsibilities further include working with cloud-native principles, enhancing CI/CD environments, ensuring robust data quality, profiling, governance, security, metadata management, and archival practices. The role requires defining workload migration strategies, driving delivery in a matrixed environment, managing risks, ensuring data security, and handling simultaneous tasks under tight deadlines. A self-starter mindset with excellent problem solving, communication, influencing, and presentation skills is essential, along with the ability to work independently and produce strategic planning, estimation, and scheduling outputs.
Required skills
Key responsibilities
- Design and implement end-to-end Data Management and Analytics solutions for large and complex environments
- Transform traditional data warehousing systems into big data–based architectures
- Apply DW dimensional modeling techniques including star and snowflake schemas, slowly changing dimensions, and dimensional hierarchies
- Implement cloud-native designs and deployments for data workloads
- Enhance and manage Continuous Integration (CI) and Continuous Development (CD) environments
- Ensure data quality, profiling, governance, security, metadata management, and archival
- Define workload migration strategies using appropriate tools
- Drive delivery across matrixed teams and multiple internal IT partners
- Manage risks, ensure data security, and execute tasks under tight deadlines
- Plan, estimate, schedule, and strategize project tasks independently
Experience & skills
- Expertise in data architecture, data strategy, and data roadmaps for large and complex systems
- Experience implementing large-scale end-to-end Data Management and Analytics solutions
- Experience transforming traditional data warehousing into big data architectures
- Strong knowledge of DW dimensional modeling including star and snowflake schemas and SCDs
- Experience with cloud-native principles and deployments
- Extensive experience with CI/CD environments
- Expertise in data quality, profiling, governance, security, metadata management, and archival
- Hands-on experience with Hadoop, Presto, Tez, Hive, Spark, and PySpark
- Proficiency in Linux, Python, GIT, and Jenkins
- Proven ability to work independently in fast-paced environments handling multiple projects
- Strong problem-solving, communication, influencing, and presentation skills