Submitting more applications increases your chances of landing a job.

Here’s how busy the average job seeker was last month:

Opportunities viewed

Applications submitted

Keep exploring and applying to maximize your chances!

Looking for employers with a proven track record of hiring women?

Click here to explore opportunities now!
We Value Your Feedback

You are invited to participate in a survey designed to help researchers understand how best to match workers to the types of jobs they are searching for

Would You Be Likely to Participate?

If selected, we will contact you via email with further instructions and details about your participation.

You will receive a $7 payout for answering the survey.


User unblocked successfully
https://bayt.page.link/GXe7K8wcT6VTtCYv6
Back to the job results

Lead Software Engineer - Pyspark and Scala

Today 2026/09/12
Other Business Support Services
Create a job alert for similar positions
Job alert turned off. You won’t receive updates for this search anymore.

Job description

We have an opportunity to impact your career and provide an adventure where you can push the limits of what's possible.


As a Lead Software Engineer at JPMorganChase within the Consumer and Community Banking, you are an integral part of an agile team that works to enhance, build, and deliver trusted market-leading technology products in a secure, stable, and scalable way. As a core technical contributor, you are responsible for conducting critical technology solutions across multiple technical areas within various business functions in support of the firm’s business objectives.


Job responsibilities


  • Lead the design, development, and maintenance of robust, scalable cloud-based data processing pipelines and infrastructure, ensuring adherence to engineering standards, governance frameworks, and industry best practices.
  • Architect and refine data models for large-scale datasets, optimizing for efficient storage, high-performance retrieval, and advanced analytics while upholding data integrity and quality.
  • Partner with cross-functional teams to translate complex business requirements into effective, scalable data engineering solutions that drive organizational value.
  • Champion a culture of innovation and continuous improvement, proactively identifying and implementing enhancements to data infrastructure, processing workflows, and analytics capabilities.
  • Define and execute data strategy, including the development of enterprise data models and the management of end-to-end data infrastructure—from design and construction to installation and ongoing maintenance of large-scale processing systems.
  • Drive data quality initiatives, ensure seamless data accessibility for analysts and data scientists, and maintain strict compliance with data governance and regulatory requirements.
  • Align data engineering practices with business objectives, ensuring solutions are both technically sound and strategically relevant.
  • Author, review, and approve technical requirements and architectural designs, and lead process re-engineering efforts to deliver cost-effective, high-impact business solution

Required qualifications, capabilities, and skills


  • Formal training or certification on software engineering concepts and 5+ years applied experience
  • Expert in at least one distributed data processing framework (Spark)
  • Expert in at least one cloud data Lakehouse platforms (AWS Data lake services or Databricks, if not Hadoop),
  • Expert in at least one scheduling/orchestration tools ( Airflow, alternatively AWS Step Functions or similar) 
  • Expert with relational and NoSQL databases.
  • Expert in data structures, data serialization formats (JSON, AVRO, Protobuf, or similar), and big-data storage formats (Parquet, Iceberg, or similar)
  • Proficiency in microservices architecture, serverless computing and distributed cluster computing tools such as Docker, Kubernetes etc.
  • Proficiency in one or more data modelling techniques (Dimensional, Data Vault, Kimball, Inmon, etc.)
  • Experience with test-driven development (TDD) or behavior-driven development (BDD) practices, as well as working with continuous integration and continuous deployment (CI/CD) tools.
  • Experience organizing and leading design workshops, coding sessions, and hackathons to promote a culture of excellence and innovation in data engineering.
  • Expertise in architecting reusable, future-ready design patterns that address diverse use cases across the organization.
  • Expertise in working with streaming platforms like Kafka, MQ etc.
Preferred qualifications, capabilities, and skills
  • Hands-on experience with Infrastructure as Code (IaC) tools, preferably Terraform; experience with AWS CloudFormation is also valued.
  • Proficiency in cloud-based data pipeline technologies such as Spinnaker  or similar platforms.
  • Strong working knowledge of the Snowflake data platform.
  • Experience in budgeting and resource allocation for data engineering projects.
  • Proven ability to manage vendor relationships effectively.

This job post has been translated by AI and may contain minor differences or errors.

You’ve reached the maximum limit of 15 job alerts. To create a new alert, please delete an existing one first.
Job alert created for this search. You’ll receive updates when new jobs match.
Are you sure you want to unapply?

You'll no longer be considered for this role and your application will be removed from the employer's inbox.