CVPR 2025 Career Opportunities
Here we highlight career opportunities submitted by our Exhibitors, and other top industry, academic, and non-profit leaders. We would like to thank each of our exhibitors for supporting CVPR 2025.
Search Opportunities
Location RI, Carnegie Mellon University, Pittsburgh, USA
Description
Location USA, WA, Seattle USA, VA, Arlington USA, NY, New York USA, CA, Palo Alto
Description The Sponsored Products & Brands team is looking for a Sr. Principal Scientist to help our millions of shoppers intuitively navigate our vast inventory by harnessing the power of GenAI and large language models.
Amazon Advertising is one of Amazon's fastest growing and most profitable businesses. As a core product offering within our advertising portfolio, SPB helps merchants, retail vendors, and brand owners succeed via advertising, which grows the incremental sales of their products sold through Amazon. The SPB team's primary goals are to help shoppers discover new products they love, be the most efficient way for advertisers to meet their business objectives, and build a sustainable business that continuously innovates on behalf of customers. Our products and solutions are strategically important to enable our Retail and Marketplace businesses to drive long-term growth. We deliver billions of ad impressions and millions of clicks and break fresh ground in product and technical innovations every day!
As a Senior Principal Scientist in Sponsored Products, you will have deep subject matter expertise in the area of large language models and generative AI across various modalities. You will work with multiple teams of scientists and engineers to translate business and functional requirements into concrete deliverables. You will invent new product experiences that enable our shoppers to easily navigate our vast inventory either via search queries, multi-turn conversations, or other modes of input. You will use your expertise to process shopper behavior and product catalog information to generate accurate shopper representations and use it to accurately predict shoppers propensity to engage with our products. You will have the opportunity to invent new approaches that help our advertisers achieve better performance using natural language as the interface. Your inputs will shape how our marketplace understands the shopper and advertiser context to present delightful discovery opportunities to our shoppers.
You will liaise with internal Amazon partners and work on bringing state-of-the-art LLM/GenAI models to production. You will stay abreast of the latest developments in the field of GenAI and identify opportunities to improve the efficiency and productivity of the team. You will define a long-term science vision for our advertising business, driven by our customer’s needs, and translate it into actionable plans for our team of of applied scientists, and engineers. Finally, you will work with academic partners to support our in-house talent with direct access to cutting edge research and mentoring.
San Diego preferred
Qualcomm's Multimedia R&D and Standards Group is seeking a candidate for the position of Video Compression Research Engineer with a focus on machine learning for video compression. You join a world-renowned team of video compression experts who develop algorithms, hardware architectures, and systems for state-of-the-art applications of classical and machine learning methods in video compression, video processing, point cloud coding and processing, AR/VR and computer vision use cases.
The ideal candidate will be a highly self-directed individual with strong creative and analytic skills, and a passion for machine learning and video compression technologies. Your work will involve developing innovative applications of Neural Networks in video compression to enhance state-of-the-art video codecs. This role offers the opportunity to contribute to groundbreaking advancements in video compression technology, working alongside a team of experts in a dynamic and collaborative environment.
We are considering candidates with various levels of experience. We are flexible on location and open to hiring anywhere, with preferred locations in USA, Germany and Taiwan.
Responsibilities: - Contribute to the conception, development, implementation, and optimization of new Neural Networks based algorithms allowing improved video compression. - Represent Qualcomm in the related standardization forums: JVET, MPEG Video, and ITU-T/VCEG. - Document and present new algorithms and implementations in various forms, including standards contributions, patent applications, conference papers and presentations, and journal publications, etc. - Ideal candidate would have the skills/experience below: - Knowledge of Neural Networks based data compression, and the theory, algorithms, and techniques used in video and image coding. - Experience in video compression standards, such as VVC/H.266 or HEVC/H.265, is a significant benefit. - Track record of successful research accomplishments demonstrated through published papers at leading conferences, and/or patent applications in the field of applications of Machine - Learning to image or video compression. - Excellent programming skills including Python and C/C++ combined with knowledge of at least one machine learning framework such as PyTorch. - Strong written and verbal English communication skills, great work ethic, and ability to work in a team environment to accomplish common goals. - PhD degree with relevant work experience or publications in the areas of video compression, video/image processing algorithms, or machine learning.
Qualifications: PhD or Masters degree in Electrical Engineering, Computer Science, Physics, Mathematics, or similar fields. 1+ years of experience with programming language such as C, C++, MATLAB, etc.
Minimum Qualifications: Bachelor's degree in Engineering, Information Systems, Computer Science, Mathematics, Physics or related field and 4+ years of Software Engineering, Hardware Engineering, Systems Engineering, or related work experience. OR Master's degree in Engineering, Information Systems, Computer Science, Mathematics, Physics or related field and 3+ years of Software Engineering, Hardware Engineering, Systems Engineering, or related work experience. OR PhD in Engineering, Information Systems, Computer Science, Mathematics, Physics or related field and 2+ years of Software Engineering, Hardware Engineering, Systems Engineering, or related work experience.
For other positions and to register please see: https://qualcomm.eightfold.ai/events/candidate/landing?plannedEventId=bkyZbLR19
Location San Jose, US / Barcelona, Spain
Description AutoX is a California-based startup in Self-Driving Car technology, and AGI in a broader context, driven by our mission to "Democratize Autonomy". With a world-class team of engineers and R&D centers across the U.S. and beyond, AutoX continues to innovate. We have multiple openings for both software and hardware engineers. The full list of available positions and their requirements can be found in "https://www.autox.ai/en/careers.html"
For those who are interested in perception, motion prediction, imitation learning, reinforcement learning, simulation and end-to-end deep learning techniques, you may also contact the hiring manager, Dr. Zheng Wu (zwu@autox.ai) for a quick update. Dr. Wu will also be on site in CVPR for in-person communication.
Los Altos, CA
At Toyota Research Institute (TRI), we’re on a mission to improve the quality of human life. We’re developing new tools and capabilities to amplify the human experience. To lead this transformative shift in mobility, we’ve built a world-class team in Human-Centered AI, Human Interactive Driving, Energy and Materials, Machine Learning, and Robotics.
The Human Interactive Driving team seeks to accelerate the path to building a much more intelligent vehicle that places humans at the center of a more evocative and safer driving experience. Under the Driving-Sensei concept, we research AI-based interactions to unlock a person’s full driving capability while simultaneously making driving safer and more enjoyable.
We are looking for a Senior Human-Computer Interaction Researcher to work on our Driving-Sensei project, focusing on designing user studies, implementing prototype software and hardware, conducting user studies, analyzing data, and reporting findings through writing papers. You will also be involved in transferring research prototypes from performance driving applications to real-world teen driving scenarios both in a driving simulator as well as real vehicle platforms.
Responsibilities
-
Design and conduct user studies to evaluate the effectiveness of HCI systems in improving driving skills and enjoyment.
-
Develop and implement prototype HCI systems that involve real-time interactions reacting to sensor inputs from a vehicle.
-
Transfer existing research prototypes from performance driving domain to teen driving application domain.
-
Analyze data collected from user studies (Python/R) and present findings through internally and externally published research papers and presentations.
-
Debug hardware and software interfaces from both software and hardware perspectives.
-
Train and manage research assistants to aid in the execution of research.
-
Capture intellectual property generated through the research process in the form of invention disclosures.
-
Specify, budget, and manage recruitment activities performed by external agencies.
-
Work on-site at our Los Altos office and travel monthly to a race track for vehicle testing.
Qualifications
-
PhD in Computer Science, Robotics, or equivalent.
-
3+ years of experience in HCI/HMI research or human-factor-related projects in the automotive field.
-
Hands-on approach to prototyping and coding (e.g., Python, C++, Arduino, Raspberry Pi).
-
Strong programming skills in Python, C++, and development tools in Linux.
-
Experience implementing interactions using machine learning algorithms, including large language models and computer vision.
-
Ability to communicate complex concepts clearly across different audiences.
-
Have a high level of initiative and self-motivation, work without direct supervision.
Bonus Qualifications
-
Working knowledge of software development using ROS2.
-
Basic understanding of vehicle control and/or wheeled robotics.
-
Experience in software development in a production/commercial environment.
-
Experience in researching shared control systems.
-
Experience working in the field of artificial intelligence including autonomous and semi-autonomous driving systems or robotics.
-
Experience working with various input and output modalities (i.e. touch, speech, haptics, displays, etc.).
San Diego or Santa Clara
We are seeking Machine Learning Engineers to develop behavioral foundation models and drive the future of autonomy across all current and future vehicle generations. You will be part of a lean, boundary-less team with access to extensive, diverse real-world datasets from various geographies and sensor setups.
Minimum Requirements - Proficient in Python programming with expertise in machine learning libraries like PyTorch. - Solid understanding of deep learning, including the underlying principles and mechanisms of modern deep learning algorithms. - Experience working with, modifying, and creating advanced algorithms. - Analytical and scientific mindset, with the ability to solve complex problems. - Excellent written and verbal communication skills, with the ability to work with a cross-functional team.
Preferred Qualifications - Ph.D. + 1 year of industry experience in the autonomous driving or robotics domain. - Proficient in deep learning models and generative models, including transformers and diffusion models. - Experience in reinforcement learning, including areas such as offline RL, reward modeling, and other related techniques. - Experience working with simulation environments and real-world data for model validation and performance benchmarking. - Track record of publications at top-tier conferences like NeurIPS, CVPR, ICRA, ICLR, CoRL, etc. - Familiarity with self-driving technologies, sensor data processing, and real-time decision-making algorithms. - Experience with large-scale machine learning systems, distributed training, and deploying models in production environments.
Roles/Responsibilities - Design and refine behavioral models using advanced deep learning techniques to ensure safe and efficient autonomous driving and to navigate complex driving scenarios effectively. - Train and integrate deep learning models with real-world data and simulation frameworks to enhance model accuracy and reliability. - Collaborate with researchers and engineers to push the boundaries of AI, driving innovation in autonomous vehicle technology. - Requires verbal and written communication skills to convey complex information. May require negotiation, influence, tact, etc. - Has a great degree of influence over key organizational decisions. - Tasks do not have defined steps; planning, problem-solving, and prioritization must occur to complete the tasks effectively. - Provides leadership and supervision/guidance to other team members.
Minimum Qualifications: • Bachelor's degree in Computer Science, Electrical Engineering, Mechanical Engineering, or related field and 6+ years of Systems Engineering or related work experience. OR Master's degree in Computer Science, Electrical Engineering, Mechanical Engineering, or related field and 5+ years of Systems Engineering or related work experience. OR PhD in Computer Science, Electrical Engineering, Mechanical Engineering, or related field and 4+ years of Systems Engineering or related work experience.
For other positions and to register see: https://qualcomm.eightfold.ai/events/candidate/landing?plannedEventId=bkyZbLR19
University of Surrey, Guildford, UK
Warehouses represent complex, dynamic environments requiring efficient navigation and task execution by autonomous robots, with minimal downtime and human intervention. Modern SLAM (Simultaneous Localization and Mapping) techniques can provide a robust mapping of the warehouse itself. However, there are also a huge number of dynamically interacting entities (including other robots, human operators, vehicles, etc). This, coupled with significant variations across regions and customer requirements, can lead to complex emergent situations at deployment that were not considered during development. This project seeks to address these challenges by developing flexible generative world models which are capable of rapidly simulating diverse warehouse environments using varying sensor loadouts and metadata.
By leveraging generative diffusion models, multimodal generative models, and large language models, the project will create adaptive, scalable world models that can simulate and predict warehouse dynamics. This enables robots to train, test, and adapt in virtual environments before deployment, significantly reducing real-world testing requirements and robustness. This digital twinning capability also makes it possible to safely test resilience to rare emergency scenarios, such as hardware failures or human obstacles.
Academic Institution
This is an opportunity to join the Centre for Vision, Speech and Signal Processing at the University of Surrey. This is the largest such UK institute, and is ranked 1st for Computer vision research in the UK and 3rd in Europe (csrankings.org). The supervisory team includes award-winning and world-renowned academics, with the applicant joining a large and tightly knit research team of 10+ peers working on various related topics (http://personalpages.surrey.ac.uk/s.hadfield/).
Industrial Sponsorship and Research Impact
This project is funded at a 50% rate by Locus Robotics. Locus are world-leaders in Warehouse Automation. They have a fleet of more than 50,000 active robots in daily operation around the world, and supply warehouse logistics systems for more than 50 internationally recognisable customers including DHL, UPS and Boots. In addition to funding studentship costs, Locus are also providing access to their hardware, warehouses, and engineer time to support the project. There is also an opportunity to undertake a paid internship in industry, helping build impact for your research activities and getting real-world research experience.
Eligibility criteria Open to any UK or international candidates. Up to 30% of our UKRI funded studentships can be awarded to candidates paying international rate fees.
You will need to meet the minimum entry requirements for our PhD programme.
Location USA, WA, Seattle USA, VA, Arlington USA, NY, New York
Description Our team drives innovation for the onsite retail shopping site publisher. We collaborate and influence the advertising science community at Amazon across the entire Ads stack— sourcing and ranking ads, bidding into and pricing in the ads auction, experimentation design to measure causal impact of showing ads on shoppers and Amazon vendors, creative development with GenAI, and more. If you’re interested in joining a unique, highly respected advertising group with a relentless focus on the customer, you’ve come to the right place. As we continue building to out our team, we are looking for a Principal Applied Scientist, to lead research, design, experiment and implementation of cutting edge algorithms for complex Stores Display publisher use cases. You will collaborate with Amazon demand programs and front-end teams to influence how the entire advertising stack works end-to-end. You will work with business and finance teams to design experiments that inform tradeoffs across advertiser and shopper goals. You will work collaboratively with software engineers, data engineers, and other scientists on the team to deploy production algorithms that support the billions of ad auctions run daily and set a high bar for science and engineering excellence. This is a rare opportunity to be a foundational member with huge potential for impact and innovating new user experiences at Amazon.
Los Altos, CA
At Toyota Research Institute (TRI), we’re on a mission to improve the quality of human life. We’re developing new tools and capabilities to amplify the human experience. To lead this transformative shift in mobility, we’ve built a world-class team in Energy & Materials, Human-Centered AI, Human Interactive Driving, Large Behavioral Models, and Robotics.
The Mission
Make general-purpose robots a reality via large-scale embodied AI.
The Challenge
We envision a future where robots assist with household chores and cooking, aid the elderly in maintaining their independence, and enable people to spend more time on the activities they enjoy most. To achieve this, robots need to be able to operate reliably in messy, unstructured environments. Our mission is to answer the question “What will it take to create truly general-purpose robots that can accomplish a wide variety of tasks in settings like human homes with minimal human supervision?”. To answer this, we are gathering large datasets of physical interaction from a variety of sources (including robots and people) and training large generative foundation models on this physical interaction data along with language, video, audio, and other rich modalities.
The Team
Our goal is to revolutionize the field of robotic manipulation, enabling long-horizon dexterous behaviors to be efficiently taught, learned, and improved over time in diverse, real world environments.
Our team has deep cross-functional expertise across simulation, perception, controls, language, vision, multimodal learning, and generative modeling. Success is measured by the advancement of robot capabilities and we’re strong believers in open research. Our north star is fundamental technological advancement in building robots that can flexibly perform a wide variety of tasks in diverse environments with minimal human supervision. Come join us and let’s make general-purpose robots a reality!
We operate a fleet of robots, and robot-embodied teaching and deployment are key parts of our strategy. Some of our previous work is highlighted here.
The Opportunity
We’re looking for a driven Staff/Senior Software Engineer comfortable working on large integrated machine learning systems. Experience with robots or other embodied systems (such as autonomous vehicles) is a bonus. You will play a pivotal role in designing, building, and optimizing pipelines that support large behavior models. You will collaborate with a team of researchers and engineers to develop resilient, scalable, and impactful systems. This role requires technical leadership, hands-on development, and strategic thinking.
If our mission of revolutionizing robotics through machine learning resonates with you, get in touch and let’s talk about how we can create the next generation of AI-powered capable robots together!
Responsibilities
-
Architect and implement large-scale software systems for behavior modeling and decision-making in robotics.
-
Develop tools, frameworks, and pipelines to train, evaluate, and deploy models efficiently.
-
Mentor junior engineers and provide guidance on best practices in software development and machine learning.
-
Contribute to technical strategy and decision-making for the LBM team.
-
Partner with researchers to translate cutting-edge algorithms into production-quality systems.
-
Work cross-functionally with simulation, hardware, perception, and control teams to integrate LBM capabilities into embodied systems.
-
Improve the scalability, robustness, and efficiency of large behavior models for real-world applications.
-
Identify and address bottlenecks in data pipelines and model performance.
Location: Los Altos
Description
At Toyota Research Institute (TRI), we’re on a mission to improve the quality of human life. We’re developing new tools and capabilities to amplify the human experience. To lead this transformative shift in mobility, we’ve built a world-class team in Energy & Materials, Human-Centered AI, Human Interactive Driving, Large Behavioral Models, and Robotics.
TRI is assembling a world-class team to develop and integrate innovative solutions that enable a robot to perform complex, human-level mobile manipulation tasks, navigate with and among people, and learn and adapt over time. The team will develop, deploy, and validate systems in real-world environments, in and around homes.
The team will be focused on heavily leveraging machine learning to marry perception, prediction, and action to produce robust, reactive, coordinated robot behaviors, bootstrapping from simulation, leveraging large amounts of data, and adapting in real world scenarios.
TRI has the runway, roadmap, and expertise to transition the technology development to a product that impacts the lives of millions of people. Apply to join a fast moving team that demands high-risk innovation and learning from failures, using rigorous processes to identify key technologies, develop a robust, high quality system, and quantitatively evaluate performance. As part of the team, you will be surrounded and supported by the significant core ML, cloud, software, and hardware expertise at TRI, and be a part of TRI's positive and diverse culture.
Responsibilities
- Develop, integrate, and deploy algorithms linking perception to autonomous robot actions, including manipulation, navigation, and human-robot interaction.
- Invent and deploy innovative solutions at the intersection of machine learning, mobility, manipulation, human interaction, and simulation for performing useful, human-level tasks, in and around homes.
- Invent novel ways to engineer and learn robust, real-world behaviors, including using optimization, planning, reactive control, self-supervision, active learning, learning from demonstration, simulation and transfer learning, and real-world adaptation.
- Be part of a team that fields systems, performs failure analysis, and iterates on improving performance and capabilities.
- Follow software practices that produce maintainable code, including automated testing, continuous integration, code style conformity, and code review.