US Military Backs Research Into AI ‘Common Sense’

The US military is backing artificial intelligence research into areas including systems with “common sense” and networks that can spot images manufactured by other AIs.

Darpa, a US Defence Department research agency best known for backing early work on the internet and self-driving vehicles, has said it is working with outside researchers on AI systems that could have the ability to adapt to unexpected circumstances, said programme manager David Gunning.

The research could result in machines with “common sense”, he said, leading to greater flexibility and adaptability and the ability to communicate with people in a more natural way.

He said current AI systems are “brittle” and can’t take on problems outside the narrow range they were built for, according to the Financial Times.

AI ‘common sense’

AI has recently become a popular way to automate complex or repetitive tasks, such as spotting patterns in large amounts of data, but there’s been a recent revival in research focused on giving machines an intuitive awareness of the world.

Microsoft co-founder Paul Allen doubled investment into his own AI research institute earlier this year in order to focus on this idea.

Darpa typically backs a range of commercial and academic groups to carry out research on its behalf, something that recently caused an internal backlash at Google over the company’s military work on image recognition.

Gunning said the agency called together third-party researchers earlier this year for a “brainstorming” session on AI “common sense” and is now putting together a formal proposal for the project.

Spotting ‘deepfakes’

In a separate project, Darpa is bringing together forensic experts this summer to test technologies aimed at spotting artificially generated images.

Called “deepfakes”, because they rely on “deep learning” techniques, they typically involve projecting one individual’s face onto the body of another, and can be surprisingly realistic.

Of particular concern is a new technique that involves the use of generative adversarial networks, or GANs, which are specifically engineered to bypass automated forgery-detection techniques.

Concerned that GANs could be used for misinformation, Darpa wants to develop more advanced detection methods, Gunning told the MIT Technology Review.

At the event this summer, the agency wants experts to compete at producing the most convincing AI-generated fake audio, imagery and video, as well as automatically spotting the counterfeits.

Google is one of the companies at the forefront of developing natural-seeming AIs, having produced a system earlier this year that carries out automated telephone conversations.

Put your knowledge of artificial intelligence (AI) to the test. Try our quiz!

Matthew Broersma

Matt Broersma is a long standing tech freelance, who has worked for Ziff-Davis, ZDnet and other leading publications

Recent Posts

Intel To Invest More Than $28 Billion In Ohio Chip Factories – Report

Troubled chip giant Intel will invest more than $28 billion to construct two new chip…

2 days ago

Apple Returns To Top 5 Smartphone Ranks In China, Amid Tim Cook Visit

In Q3 Apple rejoins ranks of top five smartphone makers in China, as government welcomes…

2 days ago

Apple Cuts Orders iPhone 16, Says Analyst

Industry supply chain analyst says Apple cut orders for the iPhone 16 for Q4 2024…

2 days ago

LinkedIn Fined €310m By Irish Data Protection Commission

Heavy fine for LinkedIn, after Irish data protection watchdog cites GDPR violations with people's personal…

3 days ago

CMA Begins Probe Into Alphabet Partnership With Anthropic

UK competition regulator begins phase one investigation into Alphabet's partnership with AI startup Anthropic

3 days ago