Desired Skills and Experience

  • Python – our primary infrastructure language
  • Cassandra
  • Docker (in production!)
  • Splunk, Spark, Hadoop, and PrestoDB
  • AWS
  • Python and Fabric for automation and our CD pipeline
  • Jenkins for builds and task execution
  • Linux (CentOS and Ubuntu)
  • DataDog for metrics and alerting
  • Puppet
  • Python development experience
  • Experience working on projects in the Hadoop ecosystem
  • Experience in Streaming data platforms, (Kafka, Storm)
  • Familiarity with querying a SQL interface
  • Familiarity with Cassandra, MySQL, and Elasticsearch
  • Experience in AWS services like Kinesis, IAM, EMR, Redshift and S3
  • Familiarity with using and supporting analytics systems like Hive, Redshift, Presto, Tableau and similar tools.
  • Familiarity with performance debugging and tuning at the OS, JVM and cluster (MapReduce, Hive, Spark jobs) levels.
  • Proficient in high level scripting languages - Python preferred
  • Experience developing solutions leveraging Docker
  • Experience managing Linux (Centos, Ubuntu) systems
  • Configuration management experience with Puppet, Chef, or Ansible
  • Continuous integration, testing, and deployment using Git, Jenkins
  • Exceptional communication and troubleshooting skills.
  • Experience managing Linux (Centos, Ubuntu) systems
  • Experience with relational databases (MySQL)
  • Bonus points for deploying/operating large-ish Hadoop clusters in AWS/GCP and use of EMR, DC/OS, Dataproc.
  • Unlimited paid vacation days. Choose how your time is spent
  • Never go hungry! We provide weekly Grubhub/Seamless credit
  • Regular in-office social events, including happy hours, wine tastings, karaoke, bingo with prizes and more
  • Company-Wide Initiatives encouraging innovation, continuous learning and cross-department connections