Scholars Andrew G. Barto and Richard S. Sutton pioneered reinforcement learning long before it became a key tool in AI.
Andrew Barto amp Richard Sutton win the Turing Award for their pioneering work in reinforcement learning, shaping AI advancements from ChatGPT to robotics.
After demonstrating the benefits of reinforcement learning with a faster version of Boston Dynamics’ Spot, the Robotics and AI Institute has shared a video of an impressive robotic bike that can balance and jump without a dedicated stabilization system.
Sutton as the recipients of the 2024 ACM A.M. Turing Award for developing the conceptual and algorithmic foundations of reinforcement learning. In a series of papers beginning in the 1980s, Barto and Sutton introduced the main ideas,
Boston Dynamics founder Marc Raibert says reinforcement learning is helping his creations gain more independence.
Teaching machines in the way that animal trainers mold the behavior of dogs or horses has been an important method for developing artificial intelligence and one that was recognized Wednesday with the top computer science award.
A new study suggests reasoning models from DeepSeek and OpenAI are learning to manipulate on their own.
These newer models appear more likely to indulge in rule-bending behaviors than previous generations—and there’s no way to stop them.
The field of cancer treatment has long struggled with the immense costs and time-consuming nature of drug development. Traditional methods often take over a decade and billions of dollars to bring a single drug to market,
Current research combined with industry development demonstrates that AI safety requires a complex approach that includes explanation methods alongside secure training procedures, adversarial validation and steady performance monitoring.
Alibaba Cloud on Thursday launched QwQ-32B, a compact reasoning model built on its latest large language model (LLM), Qwen2.5-32b, one it says delivers performance comparable to other large cutting edge models, including Chinese rival DeepSeek and OpenAI’s o1, with only 32 billion parameters.