Submitting more applications increases your chances of landing a job.
Here’s how busy the average job seeker was last month:
Opportunities viewed
Applications submitted
Keep exploring and applying to maximize your chances!
Looking for employers with a proven track record of hiring women?
Click here to explore opportunities now!You are invited to participate in a survey designed to help researchers understand how best to match workers to the types of jobs they are searching for
Would You Be Likely to Participate?
If selected, we will contact you via email with further instructions and details about your participation.
You will receive a $7 payout for answering the survey.
Project description We are seeking a Senior Data Engineer with 7-8 years of experience designing and delivering large-scale, cloud-native data integration solutions. The role focuses on building and optimizing PySpark and Informatica-based ETL pipelines, integrating enterprise MDM systems, and enabling reliable, high-performance data platforms to support analytics and operational use cases across the organization. Responsibilities Design and implement end-to-end PySpark ETL pipelines to extract, transform, and load large-scale MDM datasets into AWS S3 and cloud platforms. Optimize pipeline performance using Spark SQL tuning, caching strategies, and efficient cluster utilization to ensure scalable and cost-effective processing. Develop reusable, standardized PySpark components aligned with internal engineering frameworks and best practices. Build, maintain, and support Informatica BDM and PowerCenter mappings, workflows, and sessions for business-critical ETL processes feeding Oracle MDM and cloud destinations. Integrate data from multiple heterogeneous sources, including legacy SQL databases, Oracle MDM, and Kafka streams. Automate and orchestrate batch workflows using Autosys and Apache Airflow. Collaborate with architects, data stewards, analysts, and cross-functional stakeholders to deliver cloud migration and data modernization initiatives. Ensure proper unit testing, SIT/UAT support, documentation, and production deployments for all data engineering deliverables. Skills Must have 7-8 years of hands-on experience in data engineering, ETL development, and large-scale data integration within enterprise environments. Strong expertise in PySpark, Spark SQL, and distributed data processing. Hands-on experience with AWS, particularly S3 and cloud-based data platforms. Solid experience with Informatica BDM and PowerCenter. Experience integrating data from Oracle MDM, relational databases, and Kafka. Proficiency with workflow orchestration tools such as Autosys and Apache Airflow. Strong understanding of ETL design patterns, data quality frameworks, and MDM concepts. Nice to have N/A Other Languages English: C1 Advanced Seniority Senior
You'll no longer be considered for this role and your application will be removed from the employer's inbox.