Overview: Our client, a national insurance company, is seeking several Data Curators to join their team! This is an exciting data engineering focused role where you will have the opportunity to leverage your expertise with Azure Data Factory and associated tools.
This is a 6 month contract role and 100% Remote. There is an opportunity to extend or potentially convert!
Job Summary: Our client is looking for a highly motivated and experienced Data Curator to join their team. This position will report to and support the Cloud Data Engineer. The Data Curators will be responsible for maintaining, organizing, and expanding and optimizing new and current data. The ideal candidate is an experienced data pipeline builder and data wrangler who enjoys optimizing data systems and building them from the ground up.
The candidate will focus on ensuring the data is accessible, retrievable, high-quality, and available within context. They must be self-directed and comfortable supporting the data needs of multiple teams, systems and products. The right candidate will be excited by the prospect of optimizing or even re-designing our company’s data architecture to support our next generation of products and data initiatives.
Responsibilities
Ensure technology practices and processes adhere to regulatory standards and best practices for data protection privacy, information security and financial reporting. Ensure strong up-time performance and reliability of the Company’s data systems and software, including disaster recovery.
Create and maintain optimal data pipeline architecture,
Assemble large, complex data sets that meet functional / non-functional business requirements.
Identify, design, and implement internal process improvements: automating manual processes, optimizing data delivery, re-designing infrastructure for greater scalability, etc.
Ensure long-term usability of data through proper preservation methods
Accurately track curation processes, preserving raw data and adhering to quality standards
Work with stakeholders including the Executive, Product, Data and Design teams to assist with data-related technical issues and support their data infrastructure needs.
Keep our data separated and secure across national and regional boundaries through multiple data centers and Azure regions.
Create data tools for analytics and data scientist team members that assist them in building and optimizing our product into an innovative industry leader.
Work with data and analytics experts to strive for greater functionality in our data systems.
Qualifications
Bachelor’s Degree in Computer Science or related field preferred
3-5 years experience in data curation or data developer roles
Insurance Data experience a plus
Demonstrated strategic business impact includes cost reductions, enablement of new business insights, new lines of business through technology improvements and/or significant scaling of the business
Knowledge of modern data engineering practices, e.g. related to data orchestration, data warehouses, data observability and monitoring
Ability to balance short- and longer-term customer, product, and data-driven priorities across the organization
Data Enrichment, including using 3rd party sources to enrich data
Data Management, including proper promotion of data through SDLC life cycle
Data classification, building and adhering to sensitivity standards and imparting proper restrictions on data
Data Programming and ETL
Data Modelling
Ability to understand and communicate with technical and non-technical stakeholders
Effective communication, negotiation, and interpersonal skills
Advanced working SQL knowledge and experience working with relational databases, query authoring (SQL) as well as working familiarity with a variety of databases.
Experience performing root cause analysis on internal and external data and processes to answer specific business questions and identify opportunities for improvement.
Build processes supporting data transformation, data structures, metadata, dependency and workload management.
A successful history of manipulating, processing and extracting value from large disconnected datasets.
Strong project management and organizational skills.
Experience supporting and working with cross-functional teams in a dynamic environment.
They should also have experience using the following software/tools:
Experience with relational SQL and NoSQL databases, including Postgres and Cassandra.
Experience with data pipeline and workflow management tools
Experience with Azure cloud services: Fabric, Azure Data Factory, Azure Databricks, Azure Synapse Pipelines, Azure Synapse, Azure Data Lake
Big Data Analytics. Preferably in Azure Synapse Analytics, Azure Analysis Services
-Data Governance, Data Catalog, Master Data Management
Experience with stream-processing systems: Storm, Spark-Streaming, etc.
Experience with object-oriented/object function scripting languages: Python, Java, C++, Scala, etc.