ETL Developer

Company:  PETADATA
Location: Seattle
Closing Date: 17/10/2024
Hours: Full Time
Type: Permanent
Job Requirements / Description

POSITION : ETL Developer with Matillion Tool

LOCATION : Seattle, WA. - ONSITE

Experience: 10+ years

Work type: Fulltime ( W2 / C2C )


PETADATA is currently looking to hire for the ETL Developer role experience with Matillion tool for one of their clients.


Roles & Responsibilities:

  • The ideal candidate should have experience with both streaming and batch workflows will be essential in ensuring the efficient flow and processing of data to support our clients.
  • Collaborate with cross-functional teams to understand data requirements and design robust data architecture solutions.
  • Designing, develop, and implement scalable data processing solutions using Apache Spark
  • The ability to organize and to keep the projects well-arranged and structured.
  • Good knowledge on data analysis, design, and programming skills such as JavaScript, SQL and XML, and DOM.
  • Ensure data quality, integrity, and consistency throughout the ETL pipeline.
  • Integrate data from different systems and sources to provide a unified view for analytical purposes.
  • Collaborate with data analysts to implement solutions that meet their data integration needs.
  • Design and implement streaming workflows using PySpark Streaming or other relevant technologies.
  • Build strings of configured components inside a job and then run that job to accomplish a desired task such as loading or transforming data. using Matillion ETL tool
  • Develop batch processing workflows for large-scale data processing and analysis.
  • Has to analyze the business requirement to determine the volume of data extracted from different sources, data models, to ensure the quality of the data involved.
  • Should be able to figure out best storage medium required for the data warehouse needed.
  • Has to identify the data storage needs to determine the amount of data to deal with the company’s requirement.
  • Must ensure the data quality that everything is in place at the transformation stage to eliminate errors and fix unstructured and unorganized data extracted.
  • Must be responsible to ensure that the data is loaded into the warehouse system and meets the business needs and standards.
  • The Candidate must responsible for data flow validation, creating and building a secured database warehouse that meets a given company’s needs and standards.
  • Must be responsible for determining the storage needs of a business and the volume of data involved.


Required skills:

  • Implement ETL processes to extract, transform, and load data from various sources to ensure data quality, integrity, and consistency throughout the ETL pipeline.
  • The candidate should be expertise in Matillion ETL, Python, PySpark, ETL processes, CI/CD (Jenkins or GitHub).
  • Expertise in Python and PySpark to develop efficient data processing and analysis scripts.
  • Optimize code for performance and scalability, keeping up-to-date with the latest industry best practices.
  • Must load data and be proficient in valuable technical skills such as SQL, ⦁ JAVA, XML, and DOM, among others.
  • Extensive knowledge and Hands on Experience on Spark and its technologies.
  • Hands on Experience with Apache Spark framework, including the Spark SQL module for querying databases.
  • Familiar with various coding languages used in web development, including HTML, CSS and JavaScript, Python, Java, Scala, or R proficiency.
  • Applicants must be able to write clean code that’s free of bugs and reproducible by other developers.
  • Candidates should have experience managing SQL databases and organizing big data.
  • Hands on experience with ETL such as Matillion ETL, MS SQL, SSIS (Server Integration Services), Python / Perl, Oracle, SQL Server/ MySQL.
  • Solid understanding of Data warehousing Schemas, Dimensional modeling, implementing data storage solutions to supports efficient data retrieval and analysis.
  • The skill set necessary debugging ETL processes, optimizing data flows, and ensuring that the data pipeline is robust and error-free.



Educational Qualification:

Bachelor's/ Master’s degree in Computer Science, Engineering, or a related field.


We offer a professional work environment and are given every opportunity to grow in the Information technology world.


Note:

Candidates required to attend Phone/Video Call / In person interviews and after Selection of candidate (He/She) should go through all background checks on Education and Experience.


Please email your resume to:

Do NOT FORGET to mention your valid LinkedIn on Your resume


After carefully reviewing your experience and skills one of our HR team members will contact you on the next steps.

Apply Now
Share this job
PETADATA
  • Similar Jobs

  • ETL Developer

    Seattle
    View Job
  • Sr ETL Developer

    Seattle
    View Job
  • ETL/Informatica Senior Developer

    Seattle
    View Job
  • Teradata ETL Data Engineer

    Seattle
    View Job
  • Teradata ETL Data Engineer

    Seattle
    View Job
An error has occurred. This application may no longer respond until reloaded. Reload 🗙