Internship in Data Engineering | Adaltas

Work Description

Facts is a useful enterprise asset. Some phone it the new oil. The knowledge engineer collects, remodel and refine ​​raw details into data that can be used by enterprise analysts and knowledge scientists.

As aspect of your internship, you will be qualified in the distinct facets of the info engineer actions. You will construct a authentic-time, conclusion-to-stop data streaming ingestion pipeline combining metric collections, data cleansing and aggregation, storage to multiple info warehouses, (near) true-time examination by publicity crucial metrics in a dashboard, and the usage of device learning types used to the prediction and detection of weak indicators.

You will participate in the software architecture and the implementation of the pipeline with the aim of likely into production. You will be a part of an agile team led by a Significant Knowledge professional.

In addition, you will attain at the close of the internship a certification from a Cloud company, and a Databricks certification.

Firm presentation

Adaltas specializes in the processing and storage of information. We work on-premise and in the cloud to operate Huge Information platforms and fortify our clients’ groups in the areas of architecture, operations, data engineering, information science and DevOps. Companion with Cloudera and Databricks, we are also open up source contributors. We invite you to browse our website and our lots of specialized publications to learn far more about Adaltas.

Obligations

  • Gathering technique and application metrics
  • Providing a dispersed info warehouse with OLAP-sort column storage
  • Cleansing, enrichment, aggregation of knowledge flows
  • True-time investigation in SQL
  • Dashboards creation
  • Placing machine studying models into production in an MLOps cycle
  • Deployment in an Azure cloud infrastructure and on-premise

Expected skills

  • Engineering faculty, finish of scientific tests internship
  • Analytical and structured
  • Autonomous and curious
  • You are an open-minded particular person who enjoys sharing, communicating and learning from other people
  • Excellent know-how of Python, Spark and Linux systems

You will be in demand of creating the specialized architecture. We are wanting for a individual who masters or who will produce abilities on the pursuing resources and remedies:

All complementary activities are useful.

Further information and facts

  • Area: Boulogne Billancourt, France
  • Languages: French or English
  • Start out: February 2022
  • Length: 6 months
  • Teleworking: probability of doing the job 2 times a week remotely

Out there hardware

A laptop computer with the pursuing properties:

  • 32GB RAM
  • 1TB SSD
  • 8c/16t CPU

A cluster built up of:

  • 3x 28c/56t Intel Xeon Scalable Gold 6132
  • 3x 192TB RAM DDR4 ECC 2666MHz
  • 3x 14 SSD 480GB SATA Intel S4500 6Gbps

A Kubernetes cluster and a Hadoop cluster.

Remuneration

  • Salary 1200 € / thirty day period
  • Restaurant tickets
  • Transportation go
  • Participation in a person global convention

In the previous, the conferences which we attended consist of the KubeCon organized by the CNCF foundation, the Open Resource Summit from the Linux Foundation and the Fosdem.

For any request for extra info and to submit your software, make sure you contact David Worms:

You may also like