We are looking for a Senior Data Engineer to join our Platform Data Engineering team. The hire will be responsible for building platform level data engineering frameworks and tools. In addition, designing and implementing the data pipelines that feed data into the data warehouse platform across clouds (AWS + Azure).
Responsibilities
- Architect, design and develop data pipeline for scale and maintainability
- Lead in the design, implementation, and deployment of successful systems and services
- Ensure the quality of architecture and design of systems
- Build the infrastructure required for optimal extraction, transformation, and loading of data from a wide variety of data sources using SQL and Big Data technologies.
- Perform design and code reviews
- Functionally decompose complex problems into simple, straight-forward solutions
- Cross-training peers and mentoring teammates
- Document HLD/LLD for easy knowledge sharing and future scaling
- Create data tools for analytics, data scientist team members that assist them in building and optimizing our product
- Create data apis for application team members that assist them in building and optimizing our product
- 4+ years of experience in one high performance programming language - Java and/or Python preferred
- 3 Normal & Dimensional modeling experience
- ETL/ELT pipelines loading into data warehouses / data lakes
- Working knowledge of message queuing, stream processing, and highly scalable ‘big data’ data stores.
- Experience with relational SQL and NoSQL databases, including Postgres and Cassandra.
- Experience with stream-processing systems: Storm, Spark-Streaming, etc.
- Experience with big data tools: Hadoop, Spark, Kafka, etc.
- Experience designing and developing microservices
- Good Linux/UNIX systems knowledge
- Experience productionizing Machine Learning models
- Experience with feature engineering, feature stores, and feature pipeline development
- Experience helping data science teams with ML training data preparation and validation
- Knowledge of ML data versioning tools (like DVC, MLflow, etc.)
- Familiarity with data quality and validation frameworks for ML pipelines
- Experience with AWS/Azure cloud technologies and their ML-specific services
About FourKites
FourKites®, the leader in AI-driven supply chain transformation for global enterprises and pioneer of advanced real-time visibility, turns supply chain data into automated action. FourKites’ Intelligent Control Tower™ breaks down enterprise silos by creating a real-time digital twin of orders, shipments, inventory and assets. This comprehensive view, combined with AI-powered digital workers, enables companies to prevent disruptions, automate routine tasks, and optimize performance across their supply chain. FourKites processes over 3.2 million supply chain events daily — from purchase orders to final delivery — helping 1,600+ global brands prevent disruptions, make faster decisions and move from reactive tracking to proactive supply chain orchestration.
Working with us
We provide competitive compensation with stock options, outstanding benefits and a collaborative culture for all employees around the globe, including:
- 5 global recharge days, in addition to standard holidays, and a hybrid, flexible approach to work.
- Parental leave for all parents, an annual wellness stipend and volunteer days also provide you with time and resources for self care and to care for others.
- Opportunities throughout the year to learn and celebrate diversity.
- Access to leading AI tools and foundation models, with the freedom to experiment and find creative ways to be more effective in your role