This inclusive employer is a member of myGwork – the largest global platform for the LGBTQ+ business community.
We have an exciting and rewarding opportunity for you to take your software engineering career to the next level.
As a Software Engineer II at JPMorgan Chase within the Employee Experience, Workforce Tech, you serve as a seasoned member of an agile team to design and deliver trusted market-leading technology products in a secure, stable, and scalable way. You are responsible for carrying out critical technology solutions across multiple technical areas within various business functions in support of the firm’s business objectives. A Data Reliability Engineer is responsible for ensuring the reliability, availability, and performance of data systems and infrastructure within an Organization. Their Primary focus is on the designing implementing, and maintaining data-related systems to support the organization’s data needs
Job Responsibilities
- Executes software solutions, design, development, and technical troubleshooting with ability to think beyond routine or conventional approaches to build solutions or break down technical problems
- Requires weekend support and late evening Weekdays support.
- Manage and Implement AWS Cloud infrastructure , ensuring scalable , highly available , and fault-tolerant systems.
- Develop and maintain data pipelines, leveraging Apache Spark , AWS Glue and other data processing frameworks to support efficient data ingestion , transformation and storage.
- Manage Data Storage solutions, including data lake , data lake House ensuring data integrity , security and accessibility .
- Collaborate cross-functional teams to integrate various data solutions into a cohesive and reliable data ecosystem.
- Monitor and optimize system performance, using advanced observability tools like Dynatrace , Grafana , etc. to ensure system reliability and performance metrics are met.
- Implement and monitor robust backup and disaster recovery solutions to ensure the app / data durability and high availability.
- Drive continuous improvement initiatives in the infrastructure, focusing on automation , cost optimization and process refinement .
- Ensure compliance with data governance and security polices , applying best practices in data management and cloud operations
- Proficient in Infrastructure as Code(IaC) tools like Terraform scripting.
- Strong understanding of observability and monitoring tools with the ability to implement solutions that enhance system readiness and performance .
- Excellent Problem-solving skills with ability to analyze complex data platform systems and implement best practices and solutions.
- Participate in on-call rotation for critical incident response and proactive system monitoring.
- Formal training or certification on software engineering concepts and 2+ years applied experience
- Experience with coding language Java, Python, Scala, Clojure.
- Experience working platform like Apache Spark, Apache Flink, Glue, Kafka
- Experience with SQL/ NoSQL platforms such as Redshift, DynamoDB, Tiger graph, Snowflake
- Experience with Orchestration tool like Alteryx, Terraform, Salt, CloudFormation
- Working Experience with Data platforms like Databricks
- Good knowledge of Cloud AWS, Azure, OCI (+ knowledge of CSP-specific services)
- Experience with incident management tools i.e. Jira/Service Now
- Experience with Observability or logging tools i.e. Dynatrace, Datadog, Splunk, Grafana, log4j, CloudWatch, ELK
- Experience with Cloud processing plane like Lambda, EKS, Kubernetes, Virtual machines.
- Overall knowledge of the Software Development Life Cycle
- Solid understanding of agile methodologies such as CI/CD, Applicant Resiliency, and Security
- Knowledge of reporting tools like Tableau, AWS QuickSight
- Familiarity with modern front-end technologies
- Cloud certification like AWS Solution architect , Terraform certified or Dynatrace certified.