Jointly developed by open-source initiative Agentica and San Francisco-based start-up Together AI, DeepSWE was trained on the Qwen3-32B large language model (LLM) – part of Alibaba...
Reinforcement learning with human feedback (RLHF) is the standard method for aligning large language models (LLMs) with human preferences — such as the preferences for nontoxic...
Code generation — automatically translating natural-language specifications into computer code — is one of the most promising applications of large language models (LLMs). But the more...
From left to right: David Luan, VP of Autonomy and head of Amazon’s AGI SF Lab, and Pieter Abbeel, Amazon Scholar, Robotics. Today, we’re excited to...
Amazon’s papers at the International Conference on Machine Learning (ICML) lean — like the conference as a whole — toward the theoretical. Although some papers deal...
The Conference on Neural Information Processing Systems (NeurIPS) takes place this week, and the Amazon papers accepted there touch on a wide range of topics, from...
Lihong Li, a senior principal scientist in Amazon Ads, has won the 2023 Seoul Test of Time award for the 2010 paper “A Contextual-Bandit Approach to...