About The Role
Our Data Platform group at Crowdstrike is unique among its kind for being uncommonly customer-focused. In line with classic Data Platform groups, we build and operate systems to centralize all of the data from Falcon Sensors and 3rd Party sources derived from trillions of events/day. We also drive industry-leading innovation on a hyper scale security data lake that helps find bad actors and stop breaches.
As an E6 Engineer on the team - what would commonly be called "Senior Engineer" at most companies, SDE3 at Crowdstrike - you will contribute to the full spectrum of our systems, from foundational processing and data storage to scalable pipelines, frameworks, tools and applications that make that data available to other teams and systems.
One of your core responsibilities will be building and maintaining Self-service Data Platforms based upon Spark as the runtime, making it easy for customers to transform and access our data for analytics, machine learning, and threat hunting. Ultimately, you're empowering them to predictively build (among other things) automations that defend against such threats 'before' they appear in their own environments.
Another key value add for you will be contributions to a new graph database, which you will have a significant hand in building out.
Your primary toolset in your work will be Java/Spark, Kubernetes and AWS native tooling.
What You'll Do
- Write highly fault-tolerant Java code within Apache Spark to produce platform products used by our customers to query our data for insight into active threat trends and related analytics
- Design, develop, and maintain ultra-high-scale data platforms that process petabytes of data
- Participate in technical reviews of our products and help us develop new features and enhance stability
- Continually help us improve the efficiency and reduce latency of our high-performance services to delight our customers
- Research and implement new ways for both internal stakeholders as well as customers to query their data efficiently and extract results in the format they desire
What You'll Need
- 6+ years' experience combined between backend and data platform engineering roles
- 4+ years of experience building data platform product(s) or features with (one of) Apache Spark, Flink or Iceberg, or with comparable tools in GCP
- 3+ years of experience programming with Java, Scala or Kotlin.
- Proven experience owning robust feature/product design end to end, yourself, especially with vaguely defined problem statements or only 'loose' specs leading the way.
- Proven expertise with algorithms, distributed systems design and the software development lifecycle
- Experience building large scale data/event pipelines
- Expertise designing solutions with relational SQL and NoSQL databases, including Postgres/MySQL, Cassandra, DynamoDB
- Good test driven development discipline
- Reasonable proficiency with Linux administration tools
- Proven ability to work effectively with remote teams
Experience With The Following Is Desirable
- Go
- Pinot or other time-series/OLAP-style database
- Iceberg
- Kubernetes
- Jenkins
- Parquet
- Protocol Buffers/GRPC
PandoLogic.