We are seeking a Senior Data Integration Engineer (Airflow, Snowflake) to design and implement a robust data factory that powers AI-driven investment intelligence. In this role, you will architect scalable data integration solutions, build advanced graph-based data layers and ensure secure, high-quality data management across the platform.
Responsibilities
- Identify and integrate structured, unstructured and semi-structured data sources
- Implement role-based access control (RBAC) and develop logical data models, ensuring compliance and data lineage
- Document business transformations and establish data quality approaches and rules
- Create a comprehensive data catalog and register data assets with metadata, focusing on sensitivity and security
- Translate and implement the architectural blueprint for the Data Factory, encompassing acquisition, preparation and consumption
- Ensure robust data management and quality to support automated investment processes and AI-driven analytics
- Implement a Neo4j-powered Knowledge Graph as the central data layer for investment intelligence, linking research, strategies and portfolio actions into a lineage graph for decision traceability
- Develop and maintain a multi-dimensional asset ontology for flexible categorization and cross-dimensional analysis
- Integrate temporal versioning to support time-aware queries and historical context preservation
- Incorporate GraphRAG for transforming unstructured documents into queryable graph entities with source attribution
- Implement property-based access control for secure, multi-tenant data visibility and query-level protection
- Oversee real-time data ingestion pipelines and the implementation of scalable data warehousing solutions using Snowflake
Requirements
- 3+ years of experience in data integration and engineering
- Strong proficiency in Snowflake and Airflow
- Deep knowledge of GraphRAG, Graph Data Structures and Data Sharding concepts
- Background in Data Architecture/Design with experience across various market data sources
- Expertise in integrating with heavy AI-driven and assisted architecture
- Capability to share knowledge with EPAM and client teams on Cypher, Graph and RAG traversals
- English proficiency at B2 level or higher
Nice to have
- Familiarity with AWS for cloud-based solutions