Should I Worry About Artificial Intelligence?
Insight from CeDAR Director, Thomas Strohmer is featured in UC Davis Magazine on The Big Question of whether we should be concerned about Artificial Intelligence.
Written by Andy Fell
It’s the oldest plot in science fiction: A scientist creates something that can think for itself; the creation does just that, and turns on its creator. Despite this common trope, the term “artificial intelligence,” or AI, is recently everywhere from health care to travel bookings. If you’ve interacted with an automated phone or text message system, you’ve spoken to an AI. It is managing investments, assisting in medical diagnosis and placing ads on your social media. Soon, it could be driving cars and trucks.
So is a machine uprising on the horizon? Not really, said Thomas Strohmer, professor of mathematics and director of the Center for Data Science and Artificial Intelligence Research (CeDAR) at UC Davis.
“Currently there’s no ‘I’ in AI,” Strohmer said. “A true AI is very, very far away.”
But other issues need attention if AI is to bring the widest possible benefits to society, Strohmer said.
CeDAR was established in 2019 to foster groundbreaking multidisciplinary research in data science. The center provides seed funding and acts as a hub for several other data science and AI-related projects on campus, in fundamental research, agricultural sciences, health and medicine, and veterinary medicine.
Big data and algorithms
What is called AI is actually a way to combine data (often very large amounts of data) with algorithms (basically, a set of rules for calculation) to solve problems. In traditional computing, these rules are mostly set in advance (if A happens, do B). Thanks to ever-larger datasets and advances in algorithms and computing power, modern AI algorithms can learn more and more of their rules from the data itself. They are “trained” on existing data. A more accurate, if less eye-catching name is machine learning.
Because of their ability to handle a lot of data, AI/ML systems may see patterns that human minds would miss. But they are learning in a prescribed way, Strohmer said. They are limited by their programming and the data available to them.
“It doesn’t come up with its own ideas,” Strohmer said. AIs have been programmed to write poems or make digital artwork. But no AI could spontaneously decide to paint a sunset or jot down a sonnet.
Trust and accountability
The real issues around AI technology, Strohmer said, are in trust, accountability and responsibility.
“When you use machine learning to solve problems in the real world, you have to realize that not all problems are simply logic problems,” he said. “Human elements need to be incorporated.”