Senior Data Engineer (Python/Spark) - Fintech

  • Location

    New York

  • Sector:

    Data Science

  • Job type:


  • Salary:

    Circa $250,000

  • Contact:


  • Contact email:


  • Job ref:

    AAJR 02393

  • Startdate:


  • Consultant:

    Jason Rumney

My client is a well-funded New York Fintech startup who are well backed by a large investment firm. With a team of over 20 across Engineering my client are looking to expand across a number of key areas.

Your key responsibilities will include:

  • Responsibility for design and development of application on Cloudera or Hortonworks platform.
  • You will implement complex algorithms in a scalable fashion. Core data processing skills are highly important with SQL tools like HIVE/Impala/LLAP

  • Writing MapReduce jobs or Spark jobs for implementation. Ability to write Java based middle layer orchestration between various components on Hadoop/Spark stack. (example, taking a model definition in Java and scoring a dataset using Spark RDD on a cluster and storing output in a downstream application.

  • You will work closely with product and Analytic managers, user interaction designers, and other software engineers to develop new product offerings and improve existing ones.


  • Number of years of experience is not a requirement for this role, but you must have demonstrated expertise and a record of having created and implemented scaled up platforms from scratch

  • Core competency in Big Data and distributed systems, including data storage and management, searching, scheduling, logging, serialization, and modeling technologies

  • Experience working across the Fintech or Finance related industries is essential
  • Proficiency in Core Java or Scala and Spark or Hadoop environments

  • Familiarity with dependency injection, Concurrency, Guice/Spring

  • Familiarity with JDBC API / Databases like MySQL, Oracle, Hadoop

  • Knowledge of graph databases(like OrientDb) and traversal

  • Experience in using file formats like XML, JSON, Avro, Parquet, Sequence, ORC and text files

  • Knowledge of SOLR/ElasticSearch, real-time data processing using Flume or Kafka or Cloud based deployment would be preferred but not required.

  • Have experience in agile development methodologies, e.g. Scrum, Kanban.

  • Excellent logical and analytical reasoning capabilities combined with strong knowledge of core computer science concepts like algorithms, data structures, software design patterns and OOP etc.

  • Experience in building data and analytics products and a demonstrated aptitude for analytics is a plus

My client are unable to sponsor visas at this stage. They also have a budget of up to $250,000 for the right candidate.

Please get in touch to find out more.