We are looking for an outstanding individual with expertise in training large-scale AI models, including Large Language Models (LLMs) or equivalent architectures. This person will play a key role in the development and optimization of AI workloads, collaborating closely with hardware and software engineers in a co-design environment.
Requirements
- Design, train, and optimize large-scale AI models, including LLMs and similar architectures.
- Collaborate with hardware and software engineers to ensure efficient model deployment on custom silicon solutions.
- Develop new techniques for improving training efficiency, model accuracy, and hardware utilization.
- Implement distributed training strategies for large-scale models across high-performance computing infrastructures.
- Strong programming skills in Python, (Go, C++,…is considered a plus)
- Explore and implement advanced AI/ML techniques, such as sparsity, quantization, and knowledge distillation.
- Work with state-of-the-art frameworks such as TensorFlow, PyTorch, JAX, and Hugging Face Transformers.
- Conduct performance benchmarking and profiling of ML workloads on custom accelerators.
- Stay up to date with advancements in deep learning and AI hardware to bring innovative ideas into Openchip’s AI strategy.
- Participate in technical discussions and contribute to Openchip’s AI research and development roadmap.
Benefits
- Join an innovative team and experience company growth.
- We believe in investing in our employees and providing them with opportunities for growth and career development.
- Work in a hybrid environment with flexible scheduling.
- We offer a remuneration that values your experience.
- A chance to work on one of the most transformative AI and silicon engineering companies in Europe.
- The position will be based in Barcelona (Spain)