drjobs Remote SME Data Engineer

Remote SME Data Engineer

Employer Active

1 Vacancy
drjobs

Job Alert

You will be updated with latest job alerts via email
Valid email field required
Send jobs
Send me jobs like this
drjobs

Job Alert

You will be updated with latest job alerts via email

Valid email field required
Send jobs
Job Location drjobs

Ashburn, IL - USA

Monthly Salary drjobs

Not Disclosed

drjobs

Salary Not Disclosed

Vacancy

1 Vacancy

Job Description

Job Description:

Each day U.S. Customs and Border Protection (CBP) oversees the massive flow of people capital and products that enter and depart the United States via air land sea and cyberspace. The volume and complexity of both physical and virtual border crossings require the application of solutions to promote efficient trade and travel. Further effective solutions help CBP ensure the movement of people capital and products is legal safe and secure. CBP seeks capable qualified and versatile SME Data Engineers to help develop complex data analytical solutions for law enforcement personnel to assess risk of potential threats entering the country.

Responsibilities include but are not limited to:

  • Design develop and maintain scalable data pipelines and architectures to support data extraction transformation and loading (ETL/ELT) processes. Utilize strong SQL skills to perform complex data transformations and optimize database queries ensuring high performance and efficiency.
  • Building comprehensive datasets by aggregating data sourced from various relational databases facilitating data analysts and data scientists in creating machine learning models reports and dashboards.
  • Collaborate with crossfunctional teams (data analysts data scientists and business stakeholders) to understand business requirements and translate them into technical solutions.
  • Assist with the implementation of data migration/pipelines from onprem to cloud/nonrelational storage platforms.
  • Leverage distributed computing frameworks like Apache Spark to process large volumes of data efficiently.
  • Utilizing data analysis problemsolving investigative and creative thinking skills to handle extremely large datasets transforming them into various formats for diverse analytical products.
  • Respond to data queries/analysis requests from various groups within an organization. Create and publish regularly scheduled and/or ad hoc reports as needed.
  • Troubleshoot datarelated issues identify root causes and implement solutions to ensure data integrity and accuracy.
  • Implement best practices for data governance security and quality supporting the core business applications.
  • Responsible for data engineering source code control using GitLab.

Basic Qualifications:

  • Experience with relational databases and knowledge of query tools and/or BI tools like Power BI or OBIEE and data analysis tools.
  • Extensive experience with SQL and proficiency in writing complex queries.
  • Solid understanding of data warehousing concepts and platforms such as Oracle and cloudbased solutions.
  • Strong experience in automating ETL jobs via UNIX/LINUX shell scripts and CRON jobs.
  • Demonstrate a strong practical understanding of data warehousing from a production relational database environment.
  • Strong experience using analytic functions within Oracle or similar tools within nonrelational (MongoDB Cassandra etc. database systems.
  • Strong understanding of distributed computing principles and experience with frameworks like Apache Spark
  • Handsonexperience with data lake architectures and technologies in a cloud environment.
  • Experience with Atlassian suite of tools such as Jira and Confluence
  • Knowledge of Continuous Integration & Continuous Development tools (CI/CD)
  • Must be able to multitask efficiently and progressively and work comfortably in an everchanging data environment.
  • Must work well in a team environment as well as independently.
  • Excellent verbal/written communication and problemsolving skills; ability to communicate information to a variety of groups at different technical skill levels.

Preferred Qualifications:

  • 5 years of experience in developing maintaining and optimizing complex Oracle PL/SQL packages to aggregate transactional data for consumption by data science/machine learning applications.
  • 10 years of experience in working in data engineering with a focus on building and optimizing data pipelines and architectures. Must have full life cycle experience in design development deployment and monitoring.
  • Experience with one or more relational database systems such as Oracle MySQL Postgres SQL server with heavy emphasis on Oracle.
  • Extensive experience with cloud platforms (e.g. AWS Google Cloud etc) and cloud based ETL/ELT tools.
  • Experience with Amazon services such as S3 Redshift EMR and Scala.
  • Experience with migrating onprem legacy database objects and data to the Amazon S3 cloud environment.
  • Experience or familiarity with data science/machine learning and development experience for supervised and unsupervised learning with structure and unstructured datasets.
  • Certifications in relevant technologies (e.g. AWS Certified Big Data Google Professional Data Engineer) are a plus.

Employment Type

Full-Time

Company Industry

About Company

Report This Job
Disclaimer: Drjobpro.com is only a platform that connects job seekers and employers. Applicants are advised to conduct their own independent research into the credentials of the prospective employer.We always make certain that our clients do not endorse any request for money payments, thus we advise against sharing any personal or bank-related information with any third party. If you suspect fraud or malpractice, please contact us via contact us page.