We are seeking an experienced Lead Data Integration Specialist to join our team.
The ideal candidate has a minimum of 5 years of experience in data management, storage, modeling, analytics, migration, and database design. In this role as a Tech Lead/Team Lead, you will be responsible for creating innovative data integration solutions, shaping cloud solution architecture, and guiding a high-performing team to maintain best practices. Your extensive expertise in cloud environments, data warehousing solutions, and data security will be crucial for the success of our initiatives.
Responsibilities
- Lead a team of engineers and provide technical guidance and mentorship
- Design data integration solutions, model databases, and build scalable data platforms
- Utilize both classic data technologies and modern cloud or hybrid data solutions
- Shape cloud solution architecture and serve as an example for the team
- Collaborate with product and engineering teams to define technical needs and recommend architectural approaches
- Build effective relationships with architects and stakeholders across the organization
- Conduct analysis of business problems and technical environments to design quality solutions
- Review code and test implementations, ensuring compliance with best practices
- Foster a high-performance engineering culture within the team
- Prepare and maintain project documentation, including technical specs and use cases
Requirements
- At least 5 years of experience in data management, storage, analytics, migration, and database design
- Proven leadership abilities to manage technical teams and support engineer development
- Expertise in cloud environments (AWS, Azure, GCP) and data warehousing (Redshift, Snowflake, Azure Synapse Analytics, Google BigQuery)
- Familiarity with data integration tools, such as Azure Data Factory, AWS Glue, GCP Dataflow, Talend, Informatica, Apache NiFi, KNIME, SSIS
- Competency in relational databases (MS SQL Server, Oracle, MySQL, PostgreSQL) and production coding using tools like SQL, Python, SparkSQL, PySpark, R, Bash, Scala
- Proficiency in data security, data modeling, and integration methodologies, including OLAP, OLTP, ETL, DWH, Data Lake, Delta Lake, and Data Mesh
- Background in integration patterns, CDC methods, micro-batching, delta extracts, and housekeeping processes
- Understanding of data lineage, metadata management, data traceability, and the ability to produce high-quality design documentation
- Familiarity with compliance standards such as PI, GDPR, HIPAA, and the flexibility to use professional English effectively
- Background in direct customer communication and the ability to present technical solutions clearly
Technologies
- Cloud Platforms (AWS, Azure, GCP): Storage, Compute, Networking, Identity & Security, Data Warehousing (Redshift, Snowflake, BigQuery, Azure Synapse)
- Data Integration Tools (Azure Data Factory, AWS Glue, GCP Dataflow, Talend, Informatica, Apache NiFi, KNIME, SSIS, etc.)
- Programming Languages: SQL, Python, SparkSQL, PySpark, R, Bash, Scala
- Relational Databases (MS SQL Server, Oracle, MySQL, PostgreSQL)
- Dataflow orchestration, replication, and data preparation tools
- Version Control Systems (Git, SVN)
- Testing: Component, Integration Testing, and Data Reconciliation