Applied Research Engineer - Multimodal LLMs

Apply Now

Company: Apple

Location: Seattle, WA 98115

Description:

Summary
Are you excited with the amazing potential of foundation models, LLMs and multimodal LLMs?

We are looking for someone who thrives on collaboration and wants to push the boundaries of what is possible today! The Video Computer Vision org is a centralized applied research and engineering organization responsible for developing real-time on-device Computer Vision and Machine Perception technologies across Apple products. We balance research and product to deliver Apple quality, state-of-the-art experiences, innovating through the full stack, and partnering with HW, SW and ML teams to influence the sensor and silicon roadmap that brings our vision to life.

Description
We are seeking a highly motivated and skilled Applied Research Engineer to join our team. The ideal candidate will have a strong background in developing and exploring multimodal large language models that integrate various types of data such as text, image, video, and audio. You will work on cutting-edge research projects to advance our AI and computer vision capabilities, contributing to both foundational research and practical applications
Conduct research and development on multimodal large language models, focusing on exploring and utilizing diverse data modalities
Design, implement, and evaluate algorithms and models to enhance the performance and capabilities of our AI systems
Collaborate with cross-functional teams, including researchers, data scientists, software engineers, to translate research into practical applications
Stay up-to-date with the latest advancements in AI, machine learning, and computer vision, and apply this knowledge to drive innovation within the company

Similar Jobs