AWS and Python Data Engineer - job id 33183

Your Way To Work™

AWS and Python Data Engineer

Up to $170,000 - Per Year F/T Employee


How to Apply


Jason Vu


(646) 876-9536


(212) 616-4800 ext-290

A Full-time position at company that owns exchanges for financial and commodity markets, and operates 23 regulated exchanges and marketplaces..

Pay Options: F/T Employee.

Contact Jason Vu. call (646)876-9536 / (212)616-4800 ext.290 or email with the Job Code JV33183 or Click the Apply Now button ().

Location: Downtown.

Skills required for the position: AWS, ORACLE, PYTHON, RESTFUL.

Optional (not required):IMPALA, snowflake, redshift

Detailed Info:

A career making opportunity for a Data Engineer with any Oracle and Python experience. Any exposure to big data (Hadoop/MapReduce) and/or AWS is a plus.

Will be responsible for delivering solution for the data engineering group while working with cloud based database engines Impala and Amazon Redshift, Snowflake and Oracle.

The position requires the ability to quickly debug problems with systems/applications, background with Python, AWS, the ability to create RESTful API's and being able to adapt different languages.


  • Build massive reservoirs for Big data

  • Design, Develop, construct, test and maintain architectures such as large-scale data processing systems

  • Tool selection and POC analysis

  • Gather and process raw data at scale that meet functional / non-functional business requirements (including writing scripts, REST API calls, SQL Queries, etc.)

  • Develop data set processes for data modeling, mining and production

  • Integrate new data management technologies and software engineering tools into existing structures

  • Create custom software components (e.g. specialized UDFs) and analytics applications

  • Employ a variety of languages and tools (e.g. scripting languages) to marry systems together

  • Install and update disaster recovery procedures

  • Recommend ways to improve data reliability, efficiency and quality

  • Collaborate with data architects, modelers and IT team members on project goals

  • Support our business users with ad-hoc analysis and reports

  • Build high-performance algorithms, prototypes, predictive models and proof of concepts

  • Research opportunities for data acquisition and new uses for existing data

Knowledge and Experience

  • Masters degree in Computer Science, software/computer Engineering, applied math, physic, statistics or related field (or relevant work experience and technical expertise)

  • Hadoop based technologies (e.g. hdfs, Spark). Spark Experience is must

  • Strong SQL skills on multiple platform (preferred MPP systems)

  • Database Architectures

  • Data Modeling tools (e.g. Erwin, Visio)

  • 5+ years of Programming experience in Python, and/or Java

  • Experience with Continuous integration and deployment

  • Strong Unix/Linux skills

  • Experience in petabyte scale data environments and integration of data from multiple diverse sources

  • Kafka, Cloud computing, machine learning, text analysis, NLP & Web development experience is a plus

  • NoSQL experience a plus (HBase, Cassandra)

  • Finance experience, most notably in Equities and Options trading and reference data is a plus

Development/Computing Environment: .

The position offers competitive compensation package.

Job Id: 33183