Job Description
Job Title: Data Manager or a Data Steward
Location: Remote
Duration : 06 Months
Overview :
Our client is looking for a highly motivated and experienced Data Curator to join the team. This position will report to and support the Cloud Data Engineer. The Data Curator will be responsible for maintaining, organizing, and expanding and optimizing new and current data. The ideal candidate is an experienced data pipeline builder and data wrangler who enjoys optimizing data systems and building them from the ground up. The candidate will focus on ensuring the data is accessible, retrievable, high-quality, and available within context. They must be self-directed and comfortable supporting the data needs of multiple teams, systems and products. The right candidate will be excited by the prospect of optimizing or even re-designing our company's data architecture to support our next generation of products and data initiatives.
Responsibilities
- Ensure technology practices and processes adhere to regulatory standards and best practices for data protection privacy, information security and financial reporting. Ensure strong up-time performance and reliability of the Company's data systems and software, including disaster recovery.
- Create and maintain optimal data pipeline architecture,
- Assemble large, complex data sets that meet functional / non-functional business requirements.
- Identify, design, and implement internal process improvements: automating manual processes, optimizing data delivery, re-designing infrastructure for greater scalability, etc.
- Ensure long-term usability of data through proper preservation methods
- Accurately track curation processes, preserving raw data and adhering to quality standards
- Work with stakeholders including the Executive, Product, Data and Design teams to assist with data-related technical issues and support their data infrastructure needs.
- Keep our data separated and secure across national and regional boundaries through multiple data centers and Azure regions.
- Create data tools for analytics and data scientist team members that assist them in building and optimizing our product into an innovative industry leader.
- Work with data and analytics experts to strive for greater functionality in our data systems.
Qualifications
- Bachelor's Degree in Computer Science or related field preferred
- 3-5 years experience in data curation or data developer roles
- Insurance Data experience a plus
- Demonstrated strategic business impact includes cost reductions, enablement of new business insights, new lines of business through technology improvements and/or significant scaling of the business
- Knowledge of modern data engineering practices, e.g. related to data orchestration, data warehouses, data observability and monitoring
- Ability to balance short- and longer-term customer, product, and data-driven priorities across the organization
- Data Enrichment, including using 3 rd party sources to enrich data
- Data Management, including proper promotion of data through SDLC life cycle
- Data classification, building and adhering to sensitivity standards and imparting proper restrictions on data
- Data Programming and ETL
- Data Modelling
- Ability to understand and communicate with technical and non-technical stakeholders
- Effective communication, negotiation, and interpersonal skills
- Advanced working SQL knowledge and experience working with relational databases, query authoring (SQL) as well as working familiarity with a variety of databases.
- Experience performing root cause analysis on internal and external data and processes to answer specific business questions and identify opportunities for improvement.
- Build processes supporting data transformation, data structures, metadata, dependency and workload management.
- A successful history of manipulating, processing and extracting value from large disconnected datasets.
- Strong project management and organizational skills.
- Experience supporting and working with cross-functional teams in a dynamic environment.
- They should also have experience using the following software/tools:
- Experience with relational SQL and NoSQL databases, including Postgres and Cassandra.
- Experience with data pipeline and workflow management tools
- Experience with Azure cloud services: Fabric, Azure Data Factory, Azure Databricks, Azure Synapse Pipelines, Azure Synapse, Azure Data Lake
- Big Data Analytics. Preferably in Azure Synapse Analytics, Azure Analysis Services
-Data Governance, Data Catalog, Master Data Management - Experience with stream-processing systems: Storm, Spark-Streaming, etc.
- Experience with object-oriented/object function scripting languages: Python, Java, C++, Scala, etc.
- Experience with API Integrations
Job Tags
Temporary work, Remote job,