Day-to-Day
We are looking for a fully remote Senior Data Engineer to join a client in the rail industry. In this role, you will be responsible for expanding and optimizing our data and data pipeline architecture, as well as optimizing data flow and collection for cross functional teams. The ideal candidate is an experienced data pipeline builder and data wrangler who enjoys optimizing data systems and building them from the ground up. The Data Engineer will develop and support a broad range of software capabilities including building data pipelines, managing the ETL/ELT processes, receiving, and delivering data through various interfaces, and processing significant amounts of data related to railcar movements, railcar liability, and financial calculations. The right candidate will be excited by the prospect of optimizing or even re-designing our company's data architecture to support our next generation of products and data initiatives. A successful candidate has experience working with Agile teams and developing data engineering solutions, Design and develop generic, scalable data pipelines in Azure Data factory and Databricks with python for on-prem and cloud data sources, as well as leverage your curiosity for solving unstructured data problems and ability to manipulate and optimize large data sets to advance business problem-solving. We are looking for a highly technical individual who is at the expert-level in coding with Python and has a background in data warehousing.
Must-haves
- Bachelor's degree in computer science, computer engineering, a related field, or equivalent experience.
- 5+ years of data engineering or equivalent experience.
- Expert-level technical experience with Apache Spark / Azure Databricks.
- Expert-level coding experience with PySpark and Python.
- Proficient in using and designing solutions on Azure Cloud infrastructure (particularly Azure Data Factory) and Azure DevOps.
- Proficient with core business intelligence and data warehousing technology.
- Proficient designing and developing data integration solutions using ETL tools such as Azure Data Factory and/or SSIS.
- Proficient with software development practices such as Agile, TDD, and CI/CD.
- 5+ years of hands-on experience in developing and deploying data architecture strategies or engineering practices.
- 5+ years of experience with complex SQL queries and knowledge of database technologies.
Plusses
- Experience with graph databases or graph libraries
- Kafka or other streaming technologies
- Elastic Search
- Snowflake
- Prior experience in rail industry or other transportation-based commodities.