Twenty is seeking a Staff Data Engineer to build the data infrastructure powering our cyber operations applications and capabilities. We're looking for someone with 8+ years of experience in data engineering and architecture with mastery-level expertise in ETL pipeline development, data lake architecture, and schema design for complex datasets, plus proven leadership experience mentoring engineers and driving technical initiatives.
About the Company
At Twenty, we're taking on one of the most critical challenges of our time: defending democracies in the digital age. We develop revolutionary technologies that operate at the intersection of cyber and electromagnetic domains, where the speed of operations exceeds human sensing and complexity transcends conventional boundaries. Our team doesn't just solve problems – we deliver game-changing outcomes that directly impact national security. We're pragmatic optimists who understand that while our mission of protecting America and its allies is challenging, success is possible.
What you’ll do
Lead the development and operation of a data lake of cyber operations and intelligence data.
Collaborate with our engineers and intelligence analysts to design query patterns for economical access to petabyte-sized datasets.
Who you are
You might have previous roles like “Data Engineer”, “Data Scientist” or have worked as a Software Engineer with a focus on infrastructure and/or data engineering.
MUST HAVE
Strong focus on schema and index design. Ability to analyze access patterns and determine optimal partitioning, indexing, and clustering strategies
Experience designing and managing data lake systems (example technologies: Apache Iceberg, Delta Lake, Apache Hive) and systems for querying them (example technologies: Apache Trino, Apache Presto, AWS Athena, Apache Spark)
Experience with a column oriented database (example technologies: Clickhouse, AWS AWS Redshift, Google BigQuery, etc.)
Experience with setting up ETL pipelines (example technologies: Airflow, ClickPipe, AWS Glue, NiFi, etc)
Experience building data systems from scratch, not just operating existing ones
NICE TO HAVE
Experience with a Key-Value datastore (relevant technologies: HBase, Google BigQuery, AWS DynamoDB, Redis, etc.)
Experience with streaming and message queue technologies (like: Kafka, RabbitMQ, NATS, AWS Kinesis, etc)
Experience with a graph database technology (like: Neo4J, AWS Neptune, Memgraph, Apache AGE)
Experience working with internet networking related datasets (scan data, DNS data, netflow, certificates).
Interest working on intelligence and national security problems.
Security Requirements
Must possess and be able to maintain a TS/SCI security clearance with appropriate polygraph
U.S. citizenship required
Ability to work on-site in Arlington, VA with occasional travel to Fort Meade, MD
Sponsored