Cloud (regular) NoSQL (regular) Data (regular) ML (advanced) Databases (advanced) ETL tools (advanced) ETL (advanced) Python (advanced) Together with our partner ToolsGroup we want to invite a Data Engineer for a cooperation in a long-term project.ToolsGroup is the world’s most experienced provider of market-driven demand analytics, supply chain optimization, and retail demand management software. Having been high performing for more than 20 years, they have a loyal global customer base of more than 300 customers worldwide in more than 44 countries and an excellent reputation for delivering high-quality products and superior results. ToolsGroupis recognized in multiple Gartner Magic Quadrants and Nucleus Value Matrices. About the role:As a Data Engineer, you will be responsible for assisting our Data Scientists and Machine Learning engineers in designing and executing experiments and in industrializing advanced AI & ML features. You will also be involved in helping software engineers in implementing data structures and workflows for real-time analytics at scale.What you will take on:● Designing reliable and scalable ETL/ELT pipelines right from ingestion of the data to the delivery of the end product;● Building custom solutions to automate workflows and supporting architecture;● Understanding business logic, allowing you and other team members to make informed decisions about design changes and improvements;● Taking ownership of existing products and pipelines, making sure they are delivered on time;● Implementing upgrades and optimizations to existing processes, including migration of the code to newer versions of libraries/packages/infrastructure/etc.;● Communicating with different stakeholders to set expectations and figure out priorities;● Ensuring quality, reliability and uptime for critical automated processes, including helping the datascience team diagnose and resolve issues in the pipeline and in the data;● Designing environments within the Toolsgroup platform to house the data in a secure manner andenabling clients and internal teams to access it appropriately.What you already have:● At least 3 years of commercial experience programming in languages typically employed in AI/ML such as Python, Java, Scala, GoLang or R;● At least 2 years experience working with distributed systems such as Hadoop, Spark in cloud computing platforms such as AWS, GCP and Azure (AWS Glue, Databrick, Azure Data Factory);● At least 1 year of experience maintaining a client-facing data pipeline in a production environment;● Experience in working in an Agile environment at a software company a plus;● Great communication and teamwork skills;● Experience with both structured and unstructured databases, SQL, No-SQL, Relational Databases, Datalakes and Lakehouses. Bonus points for having experience with Graph Databases and Knowledge Graphs;● Knowledge of engineering best practices.Preffered skills:● Good understanding of Data Engineering, NoSQL databases and database design, distributed systems and/or information retrieval;● Knowledge of Apache Airflow;● Familiarity with common Azure services;● Experience writing high-performance queries in SQL and unstructured or graph query languages;● Experience with SaaS highly distributed systems, distributed cloud storage, event queues and data streaming using Apache Kafka.● Ability to plan and collect requirements for projects, and interact with the analyst and data scienceteams.In exchange for your skills, we offer:● Supportive relationships, built on transparency and a flat structure, in a diverse and multinational team;● Office in the center of historical Kraków, where your dog is always welcome;● Flexibility when it comes to working from the office or home;● Perks, incl Multikafeteria, Group Life Generali insurance, Luxmed, Multisport, language lessons 1:1;● Sport and other events, including, weekly running, squash and team lunches on the house;● Tea, coffee, and all-you-can-eat fruits and nuts in the office.Educational Requirements:● Undergraduate degree in Computer Science, Computer Information Systems, Information Systems, or other related fields.The recruitment process includes:● HR screening 30min zoom call;● 1h technical interview;● General call with the team leader;● Making an offer meeting.
Remote Senior Data Engineer in Constanţa
Datele de contact vor fi vizibile dupa ce veti aplica!