Course: “IT Recruitment. Practice. Only. Course. Second edition” Tickets are always available for sale How: online 13 modules Registration: bit.ly/3iTcMz7 Data Science UA started in 2016 with the first Data Science and AI conference in Kyiv and for now, we’ve already organized 9 International Data Science UA Conferences.
21 березня 2022

Python Data Engineer (вакансія неактивна)

Київ, віддалено

Adthena’s data team is a market leader in developing complex ETL and machine learning solutions. With published authors and award winning data scientists who contribute some of the major machine learning and distributed data technologies such as Apache Spark, we are a friendly, passionate group of engineers making a career out of building great software for our customers.

The Adthena data team is dealing with hundreds of millions of data points every day, generated from over two thousand data processes running through workflows, huge distributed computations in spark, streaming data coming in twenty-four hours a day at hundreds of times a second.

Our engineering culture is underpinned by sharing knowledge, coaching and growing together. You will have the opportunity to explore / innovate new technologies, mentor engineers and lead Technology initiatives. You will enjoy this role if you love writing code, learning cutting edge new technologies, solving problems and winning as a team.

As a Data Engineer, you will be working across our entire stack, so a real passion to drive the product and technology forward is something that we value. Your responsibilities will include helping with a vision for the future architecture of this complex data system, adding innovative ideas that use the latest cutting edge technology. You will work closely with Web and Data Science teams to deliver user-centric solutions to our customers and become an expert in developing high quality technical solutions.

Responsibilities

  • Design, build and maintain analytics solutions and Python bases platforms and services.
  • Collaborate with teams to identify, profile and validate potential sources of data to meet reporting requirements
  • Work with Analyst teams to ensure requirements are understood and translated into a development build
  • Input into the organisation’s understanding of data utilization, i.e. canonical data models, data lineage, data dictionaries, user metric guides, and liaise accordingly with stakeholders
  • Write high quality, extensible and testable code by applying good engineering practices (TDD, SOLID) using Adthena’s

Engineering Practices

  • Understand and apply modern technologies, data structures and design patterns to solve real problems efficiently
  • Understand the Adthena’s data architecture and use appropriate design patterns and designs complex database tables
  • Proactively identify and help fixing production and non-production application issues
  • Develop reports, documents and dashboards as per requirements
  • Provide high level estimate of effort and risk regarding the expected deliverables
  • Work as part of Agile team to maintain, improve, monitor Adthena’s data collection processes
  • Take ownership and pride in the products we build and always make sure they are of the highest standard
  • Be empathetic towards team members and customers

Preferred qualifications

  • Bachelor degree in Computer Science, similar technical field of study or equivalent practical experience.
  • Commercial experience building data solutions using Python
  • Commercial experience with data models and data warehouses
  • Good experience in SQL, database management, and performance tuning
  • Knowledge working cloud technologies (AWS)
  • Building complex ETL workflows (Luigi, Airflow)
  • Excellent communication skills
  • Experience in data processing using traditional and distributed systems (Hadoop, Spark)
  • Web frameworks (flask / django / react)
  • Messaging technologies (rabbitMQ / kafka)
  • Basic Linux system administration
  • Continuous Integration technologies

Technologies we use across all teams

  • Languages: Python, Java, Scala, JavaScript (React, Backbone), SQL
  • Frameworks: DropWizard, React, Akka and Play Framework (Scala)
  • Databases: PostgreSQL, AWS(S3), Redshift, Redis, MongoDB, Cassandra
  • Technologies: RabbitMQ (messaging), Quartz scheduling, Docker and Kubernetes, Maven
  • CI/CD: TeamCity, Jenkins
  • Source Control: Git (GitHub)
  • Other Tools: IntelliJ IDEA, Jira, Grafana