Desired Skills and Experience
- Python – our primary infrastructure language
- Cassandra
- Docker (in production!)
- Splunk, Spark, Hadoop, and PrestoDB
- AWS
- Python and Fabric for automation and our CD pipeline
- Jenkins for builds and task execution
- Linux (CentOS and Ubuntu)
- DataDog for metrics and alerting
- Puppet
- Python development experience
- Experience working on projects in the Hadoop ecosystem
- Experience in Streaming data platforms, (Kafka, Storm)
- Familiarity with querying a SQL interface
- Familiarity with Cassandra, MySQL, and Elasticsearch
- Experience in AWS services like Kinesis, IAM, EMR, Redshift and S3
- Familiarity with using and supporting analytics systems like Hive, Redshift, Presto, Tableau and similar tools.
- Familiarity with performance debugging and tuning at the OS, JVM and cluster (MapReduce, Hive, Spark jobs) levels.
- Proficient in high level scripting languages - Python preferred
- Experience developing solutions leveraging Docker
- Experience managing Linux (Centos, Ubuntu) systems
- Configuration management experience with Puppet, Chef, or Ansible
- Continuous integration, testing, and deployment using Git, Jenkins
- Exceptional communication and troubleshooting skills.
- Experience managing Linux (Centos, Ubuntu) systems
- Experience with relational databases (MySQL)
- Bonus points for deploying/operating large-ish Hadoop clusters in AWS/GCP and use of EMR, DC/OS, Dataproc.
- Unlimited paid vacation days. Choose how your time is spent
- Never go hungry! We provide weekly Grubhub/Seamless credit
- Regular in-office social events, including happy hours, wine tastings, karaoke, bingo with prizes and more
- Company-Wide Initiatives encouraging innovation, continuous learning and cross-department connections