About the Role
Data is at the core of our business, providing insights into the effectiveness of our products and enabling the technology that powers them. We build and operate the platform used by the rest of the company for streaming and batch computation and to train ML models. We’re building an ecosystem where consumers and producers of data can depend on each other safely. We thrive to build high quality systems we can be proud to open source and an amazing experience for our users and ourselves. We regard culture and trust highly and are looking forward to welcoming your contribution to the team.
If you’re passionate about building large scale data processing systems, and you are motivated to make an impact in creating a robust and scalable data platform used by every team, come join us. You will jump into an early stage team that builds the data transport, collection and orchestration layers. You will help shape the vision and architecture of WeWork's next generation data infrastructure, making it easy for developers to build data-driven products and features. You are responsible for developing a reliable infrastructure that scales with the company’s incredible growth. Your efforts will allow accessibility to business and user behavior insights, using huge amounts of WeWork data to fuel several teams such as Analytics, Data Science, Sales, Revenue, Product, Growth and many others as well as empowering them to depend on each other reliably. You will be a part of an experienced engineering team and work with passionate leaders on challenging distributed systems problems.
- Building and operating large scale data infrastructure in production (performance, reliability, monitoring)
- Designing, implementing and debugging distributed systems
- Thinking through long-term impacts of key design decisions and handling failure scenarios
- Building self-service platforms to power WeWork’s Technology
- Experience with one or more of the following technologies:
- Distributed logging systems (Kafka, Pulsar, Kinesis, etc)
- Stream processing. Flink, Spark, Storm, Beam, etc
- Batch processing: Spark, Hadoop, …
- IDL: Avro, Protobuf or Thrift
- MPP databases (Redshift, Vertica, …)
- Query execution (Columnar storage, push downs): Hive, Presto, Parquet, ...
- Workflow management (Airflow, Oozie, Azkaban, ...)
- Cloud storage: S3, GCS, ...
- Understanding of distributed systems concepts and principles (consistency and availability, liveness and safety, durability, reliability, fault-tolerance, consensus algorithms)
- Eager to learn new things and passionate about technology
- Experience with contributing to open source software
- Experience with the following Cassandra, DynamoDB, RocksDB/LevelDB, Graphite, StatsD, CollectD
- You're focused on team over individual achievements.
- You building software incrementally and make consistent progress.
- You love to learn. mentor and teach others.
- You're empathetic, you build long-lasting relationship characteristic of highly efficient teams.
- You keep up-to-date with the latest developments in the field.
WeWork Technology is bridging the gap between physical and digital platforms, providing a delightful, flawless & powerful experience for members and employees. We build software and hardware that enables our members to connect with each other and the space around them like never before.
We augment our community and culture teams through the tools we build. We believe there’s a macro shift toward a new way of working—one focused on a movement towards meaning and purpose. WeWork Technology is proud to be shaping this movement.
We are a team of passionate, fearless and collaborative problem-solvers distributed globally with one goal in mind - to humanize technology across the world.
We are an equal opportunity employer and value diversity in our company. We do not discriminate on the basis of race, religion, color, national origin, gender, sexual orientation, age, marital status, veteran status, or disability status.