Welcome to the forefront of conversational AI as we explore the fascinating world of AI chatbots in our dedicated blog series. Discover the latest advancements, applications, and strategies that propel the evolution of chatbot technology. From enhancing customer interactions to streamlining business processes, these articles delve into the innovative ways artificial intelligence is shaping the landscape of automated conversational agents. Whether you’re a business owner, developer, or simply intrigued by the future of interactive technology, join us on this journey to unravel the transformative power and endless possibilities of AI chatbots.
Expert says the law is trying to catch up to technology
You can save this article by registering for free here. Or sign-in if you have an account.
The fear of Artificial Intelligence (AI) is both rational and irrational.
Subscribe now to read the latest news in your city and across Canada.
Subscribe now to read the latest news in your city and across Canada.
Create an account or sign in to continue with your reading experience.
Create an account or sign in to continue with your reading experience.
Don’t have an account? Create Account
Irrational due to our tendency to fear what we don’t understand, and rational in that no one seems to really understand what AI will bring to us for good or bad, and that includes its creators.
Is AI really intelligent? Can AI be guilty of a crime?
That question is being tested in court in Florida, where the FSU shooter asked ChatGPT what weapons would be best and when the student union would be busiest. ChatGPT apparently provided answers, without throwing up any warning flags for authorities.
The AFP reported, “Now Attorney General James Uthmeier wants to know whether that makes OpenAI a criminal.
“’If the thing on the other side of the screen was a person, we would charge it with homicide,’ he said, announcing a criminal investigation into ChatGPT maker OpenAI and leaving open the possibility of charges against the company or its employees.’”
Gavin Tighe, senior partner at Gardiner Roberts LLP in Toronto, says this is a case of the law trying to catch up to technology.
Your noon-hour look at what’s happening in Toronto and beyond.
By signing up you consent to receive the above newsletter from Postmedia Network Inc.
A welcome email is on its way. If you don’t see it, please check your junk folder.
The next issue of Your Midday Sun will soon be in your inbox.
We encountered an issue signing you up. Please try again
Interested in more newsletters? Browse here.
“What jail would you send AI to? I do think it is a problem to fit the technology into the law, as opposed to the people or companies that own the technology, that they may be responsible for what it does, just like people who own any type of technology are responsible when it commits harm to others.”
The Florida case is not dissimilar to the Tumbler Ridge shooting in Canada where the victims are bringing a civil suit against the owners of ChatGPT, which had identified a potential shooter, but the humans behind it did not warn police.
But Tighe says, “AI is just responding to questions asked. It is very difficult to put the round peg into the square hole of law which is human law for humans.”
The Los Angeles Times reports, “Adam Raine, a California teenager, used ChatGPT to find answers about everything.
“But his conversations with a chatbot took a disturbing turn when the 16-year-old sought information from ChatGPT about ways to take his own life before he died by suicide in April.
“Now the parents of the teen are suing OpenAI, the maker of ChatGPT.”
Francis Syms, Associate Dean of Information and Communications Technology at Humber Polytechnic, says that some people form a relationship with a chatbot that goes beyond getting answers to questions. It becomes a conversation as if with a real person.
Syms said, “These stories are becoming so common that the premier of Manitoba, Wab Kinew is looking at banning AI chatbots in the classroom. People become addicted and they can’t give it up.”
“I think what is really scary is that people are using ChatGPT to talk about work, and then over time it sounds like a companion and then it becomes your friend or maybe potentially your special friend.”
Could AI, a “friend” with no moral compass, send us in the wrong direction?
“They are sycophantic,” Syms says, “in that they reinforce what you feed them, so if you have negative ideas about who you are, it reinforces that. It mostly tells you that the thing you are thinking is the right thing to think.”
We are living in a real-life dystopian novel, and these court cases will be hugely consequential.
Postmedia is committed to maintaining a lively but civil forum for discussion. Please keep comments relevant and respectful. Comments may take up to an hour to appear on the site. You will receive an email if there is a reply to your comment, an update to a thread you follow or if a user you follow comments. Visit our Community Guidelines for more information.
365 Bloor Street East, Toronto, Ontario, M4W 3L4
© 2026 Toronto Sun, a division of Postmedia Network Inc. All rights reserved. Unauthorized distribution, transmission or republication strictly prohibited.
This website uses cookies to personalize your content (including ads), and allows us to analyze our traffic. Read more about cookies here. By continuing to use our site, you agree to our Terms of Use and Privacy Policy.
You can manage saved articles in your account.
and save up to 100 articles!
You can manage your saved articles in your account and clicking the X located at the bottom right of the article.