Description
As a leading financial services and healthcare technology company based on revenue, SS&C is headquartered in Windsor, Connecticut, and has 27,000+ employees in 35 countries. Some 20,000 financial services and healthcare organizations, from the world's largest companies to small and mid-market firms, rely on SS&C for expertise, scale, and technology.
Job Description
Principal Data Platform Engineer / Data Pipeline Lead OverviewWe are looking for a Principal Data Platform Engineer to lead the development of batch and real-time data pipelines on top of a modern, evolving data platform.
This is a hands-on technical leadership role where you will design and build scalable ingestion and transformation pipelines while mentoring a small team of engineers.
The core data platform foundation—including storage, compute engines, and shared services—has already been established by a dedicated platform engineering team. You will work closely with that team to build pipelines on the platform and help guide its evolution based on real-world data integration needs.
Our environment combines modern streaming and lakehouse technologies with complex legacy data sources, including DB2 replication, fixed-width files, CSV extracts, and APIs.
The team is currently small but expected to grow, providing opportunities for increased leadership responsibility and career advancement as the platform and organization expand.
What You'll DoBuild data pipelines
Design and develop batch and real-time data pipelines
Implement CDC pipelines using Debezium and Kafka
Build streaming pipelines using Kafka and Apache Flink
Develop transformation workflows using Python, Spark / PySpark, and Airflow
Integrate complex data sources
Ingest data from DB2 replication streams
Process legacy fixed-width and CSV data feeds
Integrate API-based data sources
Work with modern data platforms
Store and manage data using Apache Iceberg and Parquet
Enable analytics through Trino and StarRocks
Lead and grow the team
Mentor and guide a small team of data engineers
Establish best practices for pipeline architecture, testing, and reliability
Help recruit and grow the team as the platform expands
Remain deeply hands-on in system design and development
Collaborate with platform engineering
Work closely with the team responsible for the underlying data platform
Provide input into the ongoing evolution of the platform
8+ years building data platforms or large-scale data pipelines
Strong programming experience in Python
Experience with Spark / PySpark
Experience building pipelines with Apache Airflow
Experience with Kafka-based streaming architectures
Experience implementing CDC pipelines (Debezium or similar)
Experience with Apache Flink or other streaming frameworks
Experience with Parquet and modern table formats such as Apache Iceberg
Experience with distributed query engines such as Trino, Presto, or StarRocks
Experience integrating data from heterogeneous or legacy systems
Experience leading or mentoring engineers
Python
Apache Spark / PySpark
Apache Flink
Apache Airflow
Debezium
Kafka
Apache Iceberg
Parquet
Trino
StarRocks
Work on complex legacy-to-modern data integration problems
Build streaming and batch data pipelines at scale
Help shape the evolution of a modern open data platform
Lead and grow a small, high-impact engineering team
Opportunities for increased leadership scope as the team expands
Stay hands-on with modern distributed data systems
Unless explicitly requested or approached by SS&C Technologies, Inc. or any of its affiliated companies, the company will not accept unsolicited resumes from headhunters, recruitment agencies, or fee-based recruitment services.
SS&C Technologies is an Equal Employment Opportunity employer and does not discriminate against any applicant for employment or employee on the basis of race, color, religious creed, gender, age, marital status, sexual orientation, national origin, disability, veteran status or any other classification protected by applicable discrimination laws.
Apply on company website