Data Engineer


Who are we?

Stuff™ is a frontier-tech company based in NY, with offices in Tel Aviv and Ramallah. For the past 3.5 years, we’ve been building our own vision for the future of work. Still in stealth mode, we’re developing and refining a managed marketplace for on-demand work that perfectly captures the advantages of the human agent (gumption) and AI (consistency). Here at Stuff™, we take our vision seriously. Second only to the most important thing we have: People

What will you be doing?

  • Architect, Design, and implement infrastructures and tools to enable AI research and deployment within the AWS ecosystem
  • Develop batch and streaming pipelines that fuel machine learning services
  • Architect complicated jobs and orchestrate them on managed services
  • Install and update disaster recovery procedures to safeguard our data
  • Integrate techniques to constantly improve data reliability and quality

Who you are?

A seasoned software engineer who’s experienced in all layers of the data hierarchy – from database design to data collection and storage techniques, to a deep understanding of data transformation tools and methodologies, to provisioning and managing of analytical databases, to building infrastructures that bring machine learning capabilities into production.

What should you have?

  • 2+ years of experience developing real-time stream processing solutions using Apache Kafka, Amazon Kinesis or Apache Flink
  • 2+ years of experience developing batch processing solutions using AWS EMR or alike
  • 3+ years of experience writing production-grade Python code and working with both relational and non-relational databases
  • 2+ years of experience writing production-grade SQL
  • 2+ years of experience developing REST or PRC-based APIs
  • Solution orientation and ‘can do’ attitude – with a sense of ownership and accountability
  • Bachelor’s Degree in Computer Science, Engineering or a similar computational discipline
  • High English skills

Advantage :

  • Experience developing infrastructures that bring machine learning services to production using MLFlow or Amazon Sagemaker
  • Experience orchestrating containerized applications in the AWS ecosystem using AWS ECS and ECR
  • Experience working with graph databases such as Neo4j or Amazon Neptune
  • Experience with workflow management frameworks, such as apache airflow.

Apply for this job

    By submitting your application you accept our Privacy Policy and Terms of Use.