Job Summary
This is a Junior Data Engineer position where you will play a crucial role in the development, deployment, and maintenance of observability and monitoring solutions deployed alongside High Performance Computing (HPC) clusters and/or Generative Artificial Intelligence (GenAI) platforms. You will be responsible for improving available code base, troubleshooting issues related to hardware, software, and network connectivity, creating and maintaining documentation, working directly with clients, and collaborating with the R&D team.
About the Role
This role requires expertise in basic ETL (Extract / Transform / Load) workflows and the technologies that underpin them. You will work on a balanced mix of R&D work to continue the development on the solutions and support/deploy tasks to deliver and support customers. Key Responsibilities include:
* Solution Development: Improve available code base by adding support for new CPU or accelerator architectures, fixing bugs, improving usability, and adding new features in agreement with the Team Leader and Senior Developers.
* Troubleshooting: Identifying and resolving issues related to hardware, software, and network connectivity.
* Documentation: Creating and maintaining documentation related to deployment operations, configurations, and delivery processes.
* Client Interaction: Working directly with clients to understand carry out proof of concept analysis and providing technical support during and after installation.
* Collaboration: Working with the R&D team using the solutions to investigate the performance and reliability of new hardware being validated in the laboratory.
Requirements
We are looking for someone with:
* Basic programming experience with Python
* Fundamental understanding of SQL and database concepts
* Knowledge of REST APIs and HTTP methods concepts
* Basic understanding of ETL processes
PREFERRED SKILLS
The ideal candidate will have:
* Experience with data processing frameworks (e.g., Pandas, PySpark)
* Familiarity with cloud platforms (AWS, GCP, or Azure)
* Basic understanding of ML/DL/AI Frameworks (scikit-learn, PyTorch)
* Knowledge of data warehousing concepts
* Familiarity with other programming languages (Go, Rust, C, C++, Java)
* Experience with data visualization tools (Grafana)
* Basic understanding of Linux/Unix environments
Qualifications
To apply for this role, you should have:
* Bachelor's degree in Computer Science, Information Technology, Information Engineering, or related fields in a STEM area with equivalent field experience.
* Experience with AWS, Azure, or Google Cloud.
About Us
E4 Computer Engineering designs and implements very high-tech solutions for HPC Clusters, Cloud, Data Analytics, Artificial Intelligence, and Hyper-Converged Infrastructure for Academic and Enterprise markets. We collaborate with leading research centers and are involved in national and European-level projects in HPC and AI. Our company continuously explores future scenarios to find practical and innovative solutions to complex computational demands and new application areas.