A new training framework developed by researchers at Tencent AI Lab and Washington University in St. Louis enables large language models (LLMs) to improve themselves without requiring any ...
Large language models can transmit harmful behavior to one another through training data, even when that data lacks any ...
Cisco Talos Researcher Reveals Method That Causes LLMs to Expose Training Data Your email has been sent In this TechRepublic interview, Cisco researcher Amy Chang ...
Singapore-based AI startup Sapient Intelligence has developed a new AI architecture that can match, and in some cases vastly outperform, large language models (LLMs) on complex reasoning tasks, all ...
On the surface, it seems obvious that training an LLM with “high quality” data will lead to better performance than feeding it any old “low quality” junk you can find. Now, a group of researchers is ...
New Anthropic research shows that undesirable LLM traits can be detected—and even prevented—by examining and manipulating the model’s inner workings. A new study from Anthropic suggests that traits ...
In recent months, the AI industry has started moving toward so-called simulated reasoning models that use a “chain of thought” process to work through tricky problems in multiple logical steps. At the ...
Apple researchers have published a study that looks into how LLMs can analyze audio and motion data to get a better overview of the user’s activities. Here are the details. They’re good at it, but not ...
Like most leaders these days, the dominant topic of discussion I find myself hearing repeatedly revolves around AI—whether hearing murmurs of it as a "phantom menace" coming to steal the jobs of white ...