Data Engineer

Details of the offer

Addepto is a leading consulting and technology company specializing in AI and Big Data, helping clients deliver innovative data projects. We partner with top-tier global enterprises and pioneering startups, including Rolls Royce, Continental, Porsche, ABB, and WGU. Our exclusive focus on AI and Big Data has earned us recognition by Forbes as one of the top 10 AI companies.
As a Data Engineer, you will have the exciting opportunity to work with a team of technology experts on challenging projects across various industries, leveraging cutting-edge technologies. Here are some of the projects we are seeking talented individuals to join:
Design and development of a universal data platform for global aerospace companies. This Azure and Databricks-powered initiative combines diverse enterprise and public data sources. The data platform is in the early stages of development, covering the design of architecture and processes and allowing freedom for technology selection.
Design and development of the data platform for managing electric and hybrid vehicle data. This project involves building a data pipeline for electric vehicle data, processing thousands of signals efficiently through streaming and batch services. The data powers IoT applications for business intelligence, customer support, maintenance, and AI insights, offering a chance to work with cutting-edge technology in electric mobility.
Design of the data transformation and following data ops pipelines for a global car manufacturer. This project aims to build a data processing system for both real-time streaming and batch data. We'll handle data for business uses like process monitoring, analysis, and reporting, while also exploring LLMs for chatbots and data analysis. Key tasks include data cleaning, normalization, and optimizing the data model for performance and accuracy.
Discover our perks and benefits:Work in a supportive team of passionate enthusiasts of AI & Big Data.Engage with top-tier global enterprises and cutting-edge startups on international projects.Enjoy flexible work arrangements, allowing you to work remotely or from modern offices and coworking spaces.Accelerate your professional growth through career paths, knowledge-sharing initiatives, language classes, and sponsored training or conferences.Choose from various employment options: B2B, employment contracts, or contracts of mandate.Make use of 20 fully paid days off available for B2B contractors and individuals under contracts of mandate.Participate in team-building events and utilize the integration budget.Celebrate work anniversaries, birthdays, and milestones.Access medical and sports packages, eye care, and well-being support services, including psychotherapy and coaching.Get full work equipment for optimal productivity, including a laptop and other necessary devices.With our backing, you can boost your personal brand by speaking at conferences, writing for our blog, or participating in meetups.Experience a smooth onboarding with a dedicated buddy, and start your journey in our friendly, supportive, and autonomous culture.In this position, you will:Design and develop scalable data management architectures, infrastructure, and platform solutions for streaming and batch processing using Big Data technologies like Apache Spark, Hadoop, Databricks, Snowflake.Design streaming pipelines using Apache Spark and Kafka.Design and implement data management and data governance processes and best practices.Contribute to the development of CI/CD and MLOps processes.Develop applications to aggregate, process, and analyze data from diverse sources.Collaborate with the Data Science team on Machine Learning projects, including text/image analysis and predictive model building.Develop and organize data transformations using DBT and Apache Airflow.Translate business requirements into technical solutions and ensure optimal performance and quality.What you'll need to succeed in this role:5+ years of proven commercial experience in implementing, developing, or maintaining Big Data systems.Strong programming skills in Python: writing clean code, OOP design.Experience in designing and implementing data governance and data management processes.Familiarity with Big Data technologies like Spark, Cloudera, Kafka, Airflow, NiFi, Docker, Kubernetes, Iceberg, Trino or Hudi.Proven expertise in implementing and deploying solutions in cloud environments (with a preference for Azure and AWS).Experience with Databricks and/or Snowflake is an asset.Excellent understanding of dimensional data and data modeling techniques.Excellent communication skills and consulting experience with direct interaction with clients.Ability to work independently and take ownership of project deliverables.Master's or Ph.D. in Computer Science, Data Science, Mathematics, Physics, or a related field.
#J-18808-Ljbffr


Source: Grabsjobs_Co

Requirements

Paid Product Tester

Compensation: Varies per assignment. Up to $500 per week. Location: Remote (USA) Company: ProductReviewJobs Thank you for your interest in becoming a Pai...


From Product Review Jobs - Arizona

Published 6 days ago

Cyber Security Engineer - Prisma Cloud

Innova Solutions has a client that is immediately hiring for a Cyber Security Engineer - Prisma Cloud Position type: Full-time Contract(W2) Duration: 8 Month...


From Innova Solutions - Arizona

Published 6 days ago

Ct Tech - Travel - $2,806 Per Week

AlliedTravelCareers is working with Triage Staffing LLC to find a qualified CT Tech in Sun City West, Arizona, 85375! Pay Information $2,806 per week About T...


From Alliedtravelcareers - Arizona

Published 6 days ago

Ct Tech - Travel - $2,468 To $2,736 Per Week

AlliedTravelCareers is working with LRS Healthcare to find a qualified CT Tech in Phoenix, Arizona, 85004! Pay Information $2,468 to $2,736 per week About Th...


From Alliedtravelcareers - Arizona

Published 6 days ago

Built at: 2024-11-05T15:18:53.898Z