The project demanded implementing complex ETL processes for a diverse mix of data objects including both tables and views from multiple source systems. Each source had its own schema conventions, data types, and update frequencies, requiring careful mapping and transformation logic to unify into a consistent target model.
Managing schema changes and data type conversions across multiple environments (UAT, preproduction, and production) added significant coordination complexity. Each deployment needed to be validated independently, and slowly changing dimension (SCD) type validation had to be implemented to ensure historical data integrity.
The integration of real-time market data from Thomson Reuters alongside batch data from internal systems required an event-driven architecture that could handle both patterns reliably. Performance optimisation for large-scale data processing was critical, as trading decisions depended on timely, accurate data availability.


