AI is EVERYWHERE

But
is
it
right
for
your
business,
lab
or
classroom?

AI is deeply integrated and evolving how and when we work.
Using artificial intelligence, it took only a few minutes for the Douglas team to create this robo-beer-making image.

Between the time most people wake up and go to bed, they’ll use artificial intelligence half a dozen times, often without even knowing it. It probably made life either easier in some small way, or more complicated and cluttered, depending on perspective. Whether it’s facial recognition to open a smartphone, a relentless stream of curated content on social feeds, a digital voice assistant for a quick weather update or a navigational app to find a lunch date, AI permeates all aspects of life. According to a recent survey by software maker Pegasystems, 84 per cent of the company’s customers use at least one AI-powered device or service, but only a third of them are aware that they’re using AI.

In other words, AI is embedded into day-to-day life whether we realize it or not. 

However, when Open Source AI unleashed ChatGPT in November 2022, it felt like the dawn of a new era. Discussions and debates lit up around the dinner table about AI and machine learning — algorithms that teach computers to think like humans. Everybody, it seemed, was suddenly talking about this chat bot and its uncanny ability to quickly generate a poem customized for a friend’s birthday, churn out a 1,000-word paper on electric vehicles for a Grade 10 student or craft a legal argument in a fraction of the time it would take an articling lawyer. Or, as the team at Victoria’s Whistle Buoy Brewery demonstrated last year, how you can lean on ChatGPT to craft a recipe for a bestselling hazy pale ale, which they dubbed Robo Beer, and also write the marketing copy and help design artwork for the brew.       

Photo by: Whistle Buoy Brewing

Two months after its launch, ChatGPT had become the fastest growing consumer software app in history with more than 100 million users. Its meteoric rise fuelled OpenAI’s current valuation of US $80 billion. Similarly, the stock value of semiconductor maker Nvidia Corporation surged past US $1 trillion last year, driven by the burgeoning AI market.     

Yet, many questions remain unanswered about AI’s impact. Will it relieve workers of boring, rote tasks, liberating them for more creative endeavours, or just put people out of work?  

Is AI reliable? Failures with Tesla’s autopilot technology are linked to nearly 500 accidents, including 13 fatalities. It forced Elon Musk’s automaker to recall two million vehicles last year and proved AI’s application to autonomous driving still needs refinement.

Can we trust it? AI is able to rapidly scour the internet for vast amounts of data, some of it private and without consent, allowing an individual’s movements and behaviours to be tracked like never before. Not only do AI’s surveillance capabilities pose an Orwellian threat to privacy, its deployment in the defence sector is also extremely troubling from an ethical standpoint.     

It’s why University of Montreal computer scientist Yoshua Bengio, world renowned for pioneering work on artificial neural networks and deep learning, spends as much time championing AI as he does raising red flags. Named recently to Time magazine’s 2024 list of the 100 most influential people, Begio notably led discussions about AI’s potential for negative social impact while helping to shape the 2017 Montreal Declaration on Responsible AI.

The voluntary declaration’s noble aim is “to put AI development at the service of the well-being of all people.”

Against the backdrop of this rapidly developing AI world, Douglas sat down with some Victoria thinkers, innovators and leaders to hear how they have embraced and accommodated AI into their workflows.

AI in the Boardroom

What’s in a word? Potentially a lot. But can you tell the difference between words crafted by a human and those generated by AI? If you’ve noticed lately that superlative-laden real estate listing descriptions are making all Realtors sound like poets, you’re not alone. 

Deirdre Campbell, president of tartanbond, works in the world of communications, marketing and public relations largely for clients in the tourism sector. Authenticity, tone and voice are everything. As powerful as AI is, it makes mistakes and that can be disastrous for a media campaign. Artificial intelligence is only as good as the digital data it mines; if that data is biased, wrong, inaccurate or lacks diversity, AI will churn our dubious results and predictions.

In
meetings
AI
can
be
the
note-taker
and
help
identify
“next
steps
or
‘to
dos,’
”
enabling
everyone
to focus
on
the
meeting
without
someone
having
to
be
the
note-taker.

— Deirdre Campbell, 
president of tartanbond

Campbell and her team are embracing AI as another tool, albeit with caveats. For example, in meetings AI can be the note-taker and help identify “next steps or 
‘to dos,’ ” enabling everyone to focus on the meeting without someone having to be the note-taker, says Campbell.  

“It also has a lot of value for research. AI can pull out information from a large chunk of material that we have on file and that can save a lot of time and money,” she says. “But you have to be aware of its limitations. From a tourism perspective, the biggest concern is AI-generated photography versus the real thing. A picture is worth a thousand words, unless that picture is AI-generated.”

In 2023, a team of University of California, Berkeley and Stanford University researchers conducted a study that poked holes in the fundamental assumption of AI: that machines can learn like humans and get better over time. According to their results, ChatGPT actually got dumber and less accurate over time when it was asked to do certain tasks. For example, the chatbot’s accuracy rate on solving a certain math problem dropped more than 30 per cent over a three-month period.           

“At the end of the day, people want to do business with people,” Campbell says. “If someone feels like you’re being talked to by a machine, that’s a big ‘no.’ ”         

AI in the Lab

Scientists and tech workers are leveraging AI to find solutions, make discoveries and explore worlds that not long ago were beyond the reach of human bandwidth. AI’s ability to quickly scan, analyze and draw inferences from huge data sets is opening up exciting new avenues of applied science.

Douglas magazine 10 To Watch winner Ocean AID is harnessing AI and machine learning to help tackle ocean pollution. Andrew Polanyi, a University of Victoria computer science graduate, has had a longstanding interest in ocean health. Two years ago, his Ocean AID cofounder Archit Kumar suggested applying AI to the environmental challenge of ocean plastic. After some business strategizing, they shifted to focus on ghost fishing gear — lost and abandoned fishing gear that floats around, entangling and endangering marine mammals. 

Local Victoria companies use AI
From left: Andrew Polanyi and Archit Kumar, the founders of Ocean AID, have married AI and sonar technology to detect underwater objects — like ghost fishing gear — quickly and inexpensively. Photo by: Jeffrey Bosdet/Douglas magazine

Current methods use sonar to locate, track and retrieve underwater objects like fishing nets. It’s slow and expensive, requiring a human to either monitor the system in real time or analyze the data afterwards. Either way, it’s tedious and time-consuming.

By marrying AI with sonar technology, Ocean AID is a potential game changer.   

“AI can automatically detect objects based on machine learning from thousands of known data points,” Polanyi says. “We want to provide real-time, onboard detection that’s faster and cheaper.”

Polanyi and Kumar are in the process of leveraging a round of angel and family investment to grow the business. So far the young entrepreneurs have worked with Coastal Restoration Society, the BC Shellfish Growers Association and many of the 100 or so organizations engaged in ghost fishing gear recovery. Recently, Ocean AID completed a project with Bear River First Nation in Nova Scotia, using its AI-driven platform and data pipeline to locate and map more than 1,000 lost lobster traps.

Another UVic academic is literally making waves with AI-assisted research. Oceanographer Johannes Gemmrich recently co-published a paper with collaborators from the University of Copenhagen demonstrating how AI can help predict monster rogue waves. The researchers used AI to examine data from more than a billion waves, then the algorithm produced an equation. To their surprise, it contradicted prevailing wisdom about rogue waves that says they form when one wave slowly extracts energy from another wave to become one giant wave.

Instead, thanks to AI and machine learning, Gemmrich and his colleagues learned that rogue waves are caused by something else called linear superposition, when two or more wave systems briefly align to create massive waves.

According to the research team, they now have hard data to back what mariners have known anecdotally for centuries, giving the shipping industry a tool for avoiding dangerous waves that put crew and cargo at risk. 

Researchers at the University of British Columbia have also developed an AI model that can analyze an oncologist’s notes and flag whether or not a cancer patient will need mental health support during their treatment. Instead of a human poring over a doctor’s notes looking for telltale words and subtle nuance, AI can make predictions with better than 70 per cent accuracy, according to the UBC study.    

“AI is blowing up everywhere. Pretty much any mundane or challenging task that can be done by AI, will be done by AI,” says Ocean AID’s Polanyi.

AI in the Classroom

Some historical perspective is helpful to fully comprehend the monumental task facing schools and educators in this brave new artificial intelligence world. In 2007, Apple launched its flagship iPhone, 13 years after IBM’s prototypical Simon Personal Communicator hit the market. Yet three decades later, our education system is still struggling to develop policy around smartphone use in the classroom. Should there be system-wide bans, or should it be left to the discretion of individual teachers, schools or districts?  

In a media release last year, provincial Minister of Education and Child Care Rachna Singh said, “The ministry is continuing to research AI and how best to support schools and teachers as they navigate this topic.”

It’s a tall order. Chances are today’s students, who have grown up in the age of smartphones, are already more fluent in ChatGPT than many of their teachers ever will be. 

AI in schools

“The idea of blocking out AI is not an option,” says Michael Paskevicius, assistant professor in educational technology at UVic. “I don’t ban AI from my classrooms. It’s a tool that I encourage my students to explore, understand and use.”

Paskevicius sees time-saving potential for teachers leveraging AI to develop lesson plans or tailoring a lesson to meet individual student needs.           

Software companies are flooding the market with AI-driven platforms aimed at students, teachers, parents and schools. For example, Classcraft, according to its maker, is a game-based classroom management tool that uses AI to recognize patterns in student learning while providing real-time suggestions to teachers. Happy Numbers provides individualized math education while giving teachers insights about student growth and learning. Just how many teachers will have the bandwidth to adopt and scale these AI tools in the classroom is an open question.

“Do
your
own
work
and
check
your
sources

even
if
it
is
AI.

AI
doesn’t
provide
citations,
it
just
gives
you
a
block
of
text.”


— Michael Paskevicius, assistant professor in educational technology at UVic.

There has been plenty of hand-wringing about kids using AI to “cheat” on papers and projects, but Paskevicius believes existing policy around academic integrity suffices.

“It’s not to say that we won’t have to revisit policy, but the same rules apply. Do your own work and check your sources even if it is AI,” he says. “AI doesn’t provide citations, it just gives you a block of text.” 

Data privacy is always a concern, but especially in a school setting involving minors. Consider the fact that any information a child inputs to an AI-driven educational tool automatically becomes part of that algorithm’s generative machine learning process. Is that information secure, and who’s watching? Those are big questions.

As is often the case, regulations struggle to keep pace with the increasing sophistication of AI. In April 2024, a group of American newspapers including the Chicago Tribune and New York Daily News launched a lawsuit against Microsoft and OpenAI for using their articles to train their AI systems. It likely won’t be the last time a charge of copyright infringement is levelled against AI tech.      

Adam Dubé, associate professor in McGill University’s Department of Educational and Counseling Psychology, said educators and administrators are being “bombarded with endless advice on how to make our classrooms and institutions GenAI ready.”

“Advice ranges from the Pollyannaish rapid adoption of the latest GenAI tools to scaremongering, ineffective bans of all things algorithmic. The future of GenAI in education is neither so promising nor so bleak,” Dubé said in an interview for McGill.ca (“Experts: The impact of AI on education”).

According to Dubé, what’s needed is a “middle path” approach, one that recognizes both the benefits and risks of AI.

Just what that middle path will look like is anybody’s guess.