Job Title: Data Curator (REMOTE)
Location: Norwalk CT
Duration: 6 Months
Overview:
Our client a national insurance company is seeking several Data Curators to join their team! This is an exciting data engineering focused role where you will have the opportunity to leverage your expertise with Azure Data Factory and associated tools. This is a 6month contract role and 100% Remote. There is an opportunity to extend or potentially convert!
Job Summary:
Our Client is looking for a highly motivated and experienced Data Curator to join their team. This position will report to and support the Cloud Data Engineer. The Data Curators will be responsible for maintaining organizing and expanding and optimizing new and current data. The ideal candidate is an experienced data pipeline builder and data wrangler who enjoys optimizing data systems and building them from the ground up. The Candidate will focus on ensuring the data is accessible retrievable highquality and available within context. They must be selfdirected and comfortable supporting the data needs of multiple teams systems and products. The right candidate will be excited by the prospect of optimizing or even redesigning our companys data architecture to support our next generation of products and data initiatives.
Responsibilities:
- Ensure technology practices and processes adhere to regulatory standards and best practices for data protection privacy information security and financial reporting. Ensure strong uptime performance and reliability of the Companys data systems and software including disaster recovery.
- Create and maintain optimal data pipeline architecture
- Assemble large complex data sets that meet functional / nonfunctional business requirements.
- Identify design and implement internal process improvements: automating manual processes optimizing data delivery redesigning infrastructure for greater scalability etc.
- Ensure longterm usability of data through proper preservation methods
- Accurately track curation processes preserving raw data and adhering to quality standards
- Work with stakeholders including the Executive Product Data and Design teams to assist with datarelated technical issues and support their data infrastructure needs.
- Keep our data separated and secure across national and regional boundaries through multiple data centers and Azure regions.
- Create data tools for analytics and data scientist team members that assist them in building and optimizing our product into an innovative industry leader.
- Work with data and analytics experts to strive for greater functionality in our data systems.
Qualifications
- Bachelors degree in computer science or related field preferred
- 35 years experience in data curation or data developer roles
- Insurance Data experience a plus
- Demonstrated strategic business impact includes cost reductions enablement of new business insights new lines of business through technology improvements and/or significant scaling of the business
- Knowledge of modern data engineering practices e.g. related to data orchestration data warehouses data observability and monitoring
- Ability to balance short and longerterm customer product and datadriven priorities across the organization
- Data Enrichment including using 3rd party sources to enrich data
- Data Management including proper promotion of data through SDLC life cycle
- Data classification building and adhering to sensitivity standards and imparting proper restrictions on data
- Data Programming and ETL
- Data Modelling
- Ability to understand and communicate with technical and nontechnical stakeholders
- Effective communication negotiation and interpersonal skills
- Advanced working SQL knowledge and experience working with relational databases query authoring (SQL) as well as working familiarity with a variety of databases.
- Experience performing root cause analysis on internal and external data and processes to answer specific business questions and identify opportunities for improvement.
- Build processes supporting data transformation data structures metadata dependency and workload management.
- A successful history of manipulating processing and extracting value from large disconnected datasets.
- Strong project management and organizational skills.
- Experience supporting and working with crossfunctional teams in a dynamic environment.
- They should also have experience using the following software/tools:
- Experience with relational SQL and NoSQL databases including Postgres and Cassandra.
- Experience with data pipeline and workflow management tools
- Experience with Azure cloud services: Fabric Azure Data Factory Azure Databricks Azure Synapse Pipelines Azure Synapse Azure Data Lake
- Big Data Analytics. Preferably in Azure Synapse Analytics Azure Analysis Services
Data Governance Data Catalog Master Data Management - Experience with streamprocessing systems: Storm SparkStreaming etc.
- Experience with objectoriented/object function scripting languages: Python Java C Scala etc.
- Experience with API Integrations.