DescriptionWe have an opportunity to impact your career and provide an adventure where you can push the limits of whats possible.
As a Software EngineerIIIat JPMorgan Chase within the Employee Data Platform (Palmos) Group youare an integral part of an agile team that works to enhance build and deliver trusted marketleading technology products in a secure stable and scalable way. As a core technical contributor you are responsible for conducting critical technology solutions across multiple technical areas within various business functions in support of the firms business objectives.
Job responsibilities
- Executes software solutions design development and technical troubleshooting with ability to think beyond routine or conventional approaches to build solutions or break down technical problems
- Creates secure and highquality production code and maintains algorithms that run synchronously with appropriate systems
- Produces architecture and design artifacts for complex applications while being accountable for ensuring design constraints are met by software code development
- Gathers analyzes synthesizes and develops visualizations and reporting from large diverse data sets in service of continuous improvement of software applications and systems
- Proactively identifies hidden problems and patterns in data and uses these insights to drive improvements to coding hygiene and system architecture
- Contributes to software engineering communities of practice and events that explore new and emerging technologies
- Adds to team culture of diversity equity inclusion and respect
Required qualifications capabilities and skills
- Formal training or certification on software engineering concepts coupled with 3 years handson experience
- Experience in building data pipelines using Apache Spark with proficiency in Scala or Python utilizing platforms such as AWS EMR or Databricks
- Handson experience with IaC tools like Terraform
- Knowledge of open table formats such as Iceberg and Delta Lake along with expertise in AWS Glue Data Catalog and finegrained access control.
- Advanced in one or more programming language(s) (Python Java etc.)
- Proficiency in automation and continuous delivery methods
- Advanced understanding of agile methodologies such as CI/CD Application Resiliency and Security
- Demonstrated proficiency in software applications and technical processes within a technical discipline (e.g. cloud artificial intelligence machine learning mobile etc.)
- Hands on experience implementing data pipelines using Databricks including tools such as Unity Catalog Databricks Workflows and Databricks Live Table along with experience in data orchestration tools like Airflow technologies
Preferred qualifications capabilities and skills
- Extensive handson experience with AWS Cloud services including S3 IAM EMR Glue ECS/EKS and Athena.
- Certifications in AWS Databricks and Terraform are highly desirable.
- Exposure to cloud technologies