Skip to content

 

McLean Virginia - 120-135K Full Time Posted: Wednesday, 15 May 2019
 
 
Applicants must be eligible to work in the specified location

Prestigious Enterprise Company is currently seeking a Sr. Big Data/Cloud ETL Engineer with strong Cloud Data Storage experience. Candidate will be part of a team responsible for supporting a wide range of internal customers. Candidate will analyze, design, and develop data storage and data analytic solutions using data lake patterns that help run more effective operations and make better business decisions.

Responsibilities:

  • Develop data filtering, transformational and loading requirements
  • Define and execute ETLs using Apache Sparks on Hadoop among other Data technologies
  • Determine appropriate translations and validations between source data and target databases
  • Implement business logic to cleanse & transform data
  • Design and implement appropriate error handling procedures
  • Develop project, documentation and storage standards in conjunction with data architects
  • Monitor performance, troubleshoot and tune ETL processes as appropriate using tools like in the AWS ecosystem.
  • Create and automate ETL mappings to consume loan level data source applications to target applications
  • Execution of end to end implementation of underlying data ingestion workflow.
  • Leverage and align work to appropriate resources across the team to ensure work is completed in the most efficient and impactful way
  • Understand capabilities of and current trends in Data Engineering domain

Qualifications:

At least 5 years of experience developing in Java, Python

  • Bachelors degree with equivalent work experience in statistics, data science or a related field.
  • Experience working with different Databases and understanding of data concepts (including data warehousing, data lake patterns, structured and unstructured data)
  • 3+ years experience of Data Storage/Hadoop platform implementation, including 3+ years of hands-on experience in implementation and performance tuning Hadoop/Spark implementations.
  • Implementation and tuning experience specifically using Amazon Elastic Map Reduce (EMR).
  • Implementing AWS services in a variety of distributed computing, enterprise environments.
  • Experience writing automated unit, integration, regression, performance and acceptance tests
  • Solid understanding of software design principles
  • Strong consultation and communication skills
  • Ability to work with and collaborate across the team and where silos exist
  • Deep curiosity to learn about new trends and how to do things better
  • Ability to use data to help inform strategy and direction

Preferred Skills:

Understanding of Apache Hadoop and the Hadoop ecosystem. Experience with one or more relevant tools (Sqoop, Flume, Kafka, Oozie, Hue, Zookeeper, HCatalog, Solr, Avro).

  • Deep knowledge on Extract, Transform, Load (ETL) and distributed processing techniques such as Map-Reduce
  • Experience with Columnar databases like Snowflake, Redshift
  • Experience in building and deploying applications in AWS (EC2, S3, Hive, Glue, EMR, RDS, ELB, Lambda, etc.)
  • Experience with building production web services
  • Experience with cloud computing and storage services
  • Knowledge of Mortgage industry

McLean Virginia, United States of America
IT
120-135K
Craig Johnson 
JSCJ-BDETL
5/15/2019 2:25:53 PM

About Request Technology - Craig Johnson
You may be just a keystroke away from fulfilling yourREQUEST for success...REQUEST is an executive search firm specializing in recruiting Information technology professionals for organizations in the...

We strongly recommend that you should never provide your bank account details to an advertiser during the job application process. Should you receive a request of this nature please contact support giving the advertiser's name and job reference.

Other jobs like this

Chevy Chase, MD
Chevy Chase, MD
Chevy Chase, MD