Data Engineer (Databricks)
15 120 - 21 000 PLN/ mies.B2B (netto)
MidFull-time·B2B
#331100·Dodano 12 dni temu·22
Źródło: nofluffjobs.comTech Stack / Keywords
PythonSQLETLAzureDatabricksSparkDockerCI/CDKubernetesKafkaPower BIAirflowDagsterdbt
Firma i stanowisko
Addepto is a leading AI consulting and data engineering company that builds scalable, ROI-focused AI solutions for large enterprises and startups, including Rolls Royce, Continental, Porsche, ABB, and WGU. The company focuses exclusively on Artificial Intelligence and Big Data, helping organizations unlock the potential of their data for measurable business impact and growth. Addepto is part of KMS Technology, a US-based global technology group, combining AI specialization with enterprise-scale delivery capabilities.
Wymagania
- At least 3 years of commercial experience implementing, developing, or maintaining Big Data systems.
- Strong programming skills in Python including clean code and OOP design.
- Strong SQL skills including performance tuning, query optimization, and experience with data warehousing solutions.
- Experience designing and implementing data governance and data management processes.
- Deep expertise in Big Data technologies including Databricks, Spark, Apache Airflow, and other modern data orchestration and transformation tools.
- Experience implementing and deploying solutions in cloud environments, preferably Azure.
- Knowledge of building and deploying Power BI reports and dashboards for data visualization.
- Excellent understanding of dimensional data and data modeling techniques.
- Consulting experience with ability to guide clients through architectural decisions, technology selection, and best practices.
- Ability to work independently and take ownership of project deliverables.
- Master’s or Ph.D. in Computer Science, Data Science, Mathematics, Physics, or related field.
Nice to have:
- Experience with CI/CD, Kubernetes, Kafka, Power BI, Airflow, Dagster, dbt.
Obowiązki
- Design and optimize scalable data processing pipelines for streaming and batch workloads using Big Data technologies such as Databricks, Apache Airflow, and Dagster.
- Architect and implement end-to-end data platforms ensuring high availability, performance, and reliability.
- Lead development of CI/CD and MLOps processes to automate deployments, monitoring, and model lifecycle management.
- Develop and maintain applications for aggregating, processing, and analyzing data from diverse sources ensuring efficiency and scalability.
- Collaborate with Data Science teams on Machine Learning projects including text/image analysis, feature engineering, and predictive model deployment.
- Design and manage complex data transformations using Databricks, DBT, and Apache Airflow ensuring data integrity and consistency.
- Translate business requirements into scalable and efficient technical solutions while ensuring optimal performance and data quality.
- Ensure data security, compliance, and governance best practices are followed across all data pipelines.
Oferta
- Work in a supportive team of AI & Big Data enthusiasts.
- Engage with top-tier global enterprises and startups on international projects.
- Flexible work arrangements with options to work remotely or from modern offices and coworking spaces.
- Professional growth through career paths, knowledge-sharing initiatives, language classes, and sponsored training or conferences including Databricks certifications.
- Choice of cooperation form: B2B or contract of mandate with 20 fully paid days off.
- Participation in team-building events and integration budget.
- Celebrations of work anniversaries, birthdays, and milestones.
- Access to medical and sports packages, eye care, psychotherapy, and coaching.
- Full work equipment including laptop and necessary devices.
- Support to boost personal brand via speaking at conferences, blogging, and meetups.
Elastyczne godziny
Płatny urlop
Imprezy teamowe
Kursy językowe
Budżet konferencyjny
Dofinansowanie szkoleń
Opieka zdrowotna
Karta sportowa
Addepto
50 aktywnych ofert