The Sr. Data Engineer for streaming will be responsible for supporting and enhancing our Kafka and Kinesis based streaming data collection platforms. The ideal candidate will have strong experience with Kafka (open-source or Confluent) and AWS Kinesis, as well as cloud platforms. The candidate should also be strong with AWS and Kafka components (Topics, Producer/Consumer, KStream, KTable, and KSQL), Kinesis components including streams, firehose and data analytics, and be experienced working with streaming systems that handle very high volume and velocity data. This will require a strong understanding of streaming architectures, and experience with AWS, CI/CD tools, and infrastructure as code.
Working as part of the Data Engineering team responsible for managing data feeds and collecting analytical streaming data from our client software. In this fast-paced environment, you will deal with high volume and high velocity data, and demonstrate your strong analytical, problem solving, organizational, and prioritization skills. This important role has a wide range of responsibilities, including:
This is a critical role with a wide range of responsibilities, including:
We believe the right individual will have the following skills and experience in order to be successful in this role: