drjobs Data Curator REMOTE

Data Curator REMOTE

Employer Active

1 Vacancy
drjobs

Job Alert

You will be updated with latest job alerts via email
Valid email field required
Send jobs
Send me jobs like this
drjobs

Job Alert

You will be updated with latest job alerts via email

Valid email field required
Send jobs
Job Location drjobs

Norwalk, CA - USA

Monthly Salary drjobs

Not Disclosed

drjobs

Salary Not Disclosed

Vacancy

1 Vacancy

Job Description

Job Title: Data Curator (REMOTE)

Location: Norwalk CT

Duration: 6 Months

Overview:

Our client a national insurance company is seeking several Data Curators to join their team! This is an exciting data engineering focused role where you will have the opportunity to leverage your expertise with Azure Data Factory and associated tools. This is a 6month contract role and 100% Remote. There is an opportunity to extend or potentially convert!

Job Summary:

Our Client is looking for a highly motivated and experienced Data Curator to join their team. This position will report to and support the Cloud Data Engineer. The Data Curators will be responsible for maintaining organizing and expanding and optimizing new and current data. The ideal candidate is an experienced data pipeline builder and data wrangler who enjoys optimizing data systems and building them from the ground up. The Candidate will focus on ensuring the data is accessible retrievable highquality and available within context. They must be selfdirected and comfortable supporting the data needs of multiple teams systems and products. The right candidate will be excited by the prospect of optimizing or even redesigning our companys data architecture to support our next generation of products and data initiatives.

Responsibilities:

  • Ensure technology practices and processes adhere to regulatory standards and best practices for data protection privacy information security and financial reporting. Ensure strong uptime performance and reliability of the Companys data systems and software including disaster recovery.
  • Create and maintain optimal data pipeline architecture
  • Assemble large complex data sets that meet functional / nonfunctional business requirements.
  • Identify design and implement internal process improvements: automating manual processes optimizing data delivery redesigning infrastructure for greater scalability etc.
  • Ensure longterm usability of data through proper preservation methods
  • Accurately track curation processes preserving raw data and adhering to quality standards
  • Work with stakeholders including the Executive Product Data and Design teams to assist with datarelated technical issues and support their data infrastructure needs.
  • Keep our data separated and secure across national and regional boundaries through multiple data centers and Azure regions.
  • Create data tools for analytics and data scientist team members that assist them in building and optimizing our product into an innovative industry leader.
  • Work with data and analytics experts to strive for greater functionality in our data systems.

Qualifications

  • Bachelors degree in computer science or related field preferred
  • 35 years experience in data curation or data developer roles
  • Insurance Data experience a plus
  • Demonstrated strategic business impact includes cost reductions enablement of new business insights new lines of business through technology improvements and/or significant scaling of the business
  • Knowledge of modern data engineering practices e.g. related to data orchestration data warehouses data observability and monitoring
  • Ability to balance short and longerterm customer product and datadriven priorities across the organization
  • Data Enrichment including using 3rd party sources to enrich data
  • Data Management including proper promotion of data through SDLC life cycle
  • Data classification building and adhering to sensitivity standards and imparting proper restrictions on data
  • Data Programming and ETL
  • Data Modelling
  • Ability to understand and communicate with technical and nontechnical stakeholders
  • Effective communication negotiation and interpersonal skills
  • Advanced working SQL knowledge and experience working with relational databases query authoring (SQL) as well as working familiarity with a variety of databases.
  • Experience performing root cause analysis on internal and external data and processes to answer specific business questions and identify opportunities for improvement.
  • Build processes supporting data transformation data structures metadata dependency and workload management.
  • A successful history of manipulating processing and extracting value from large disconnected datasets.
  • Strong project management and organizational skills.
  • Experience supporting and working with crossfunctional teams in a dynamic environment.
  • They should also have experience using the following software/tools:
    • Experience with relational SQL and NoSQL databases including Postgres and Cassandra.
    • Experience with data pipeline and workflow management tools
    • Experience with Azure cloud services: Fabric Azure Data Factory Azure Databricks Azure Synapse Pipelines Azure Synapse Azure Data Lake
    • Big Data Analytics. Preferably in Azure Synapse Analytics Azure Analysis Services
      Data Governance Data Catalog Master Data Management
    • Experience with streamprocessing systems: Storm SparkStreaming etc.
    • Experience with objectoriented/object function scripting languages: Python Java C Scala etc.
    • Experience with API Integrations.

Employment Type

Full Time

About Company

Report This Job
Disclaimer: Drjobpro.com is only a platform that connects job seekers and employers. Applicants are advised to conduct their own independent research into the credentials of the prospective employer.We always make certain that our clients do not endorse any request for money payments, thus we advise against sharing any personal or bank-related information with any third party. If you suspect fraud or malpractice, please contact us via contact us page.