• Admiral Patrick@dubvee.org
    link
    fedilink
    English
    arrow-up
    62
    ·
    9 months ago

    Such a good and sane outcome. Not only am I happy for the affected customers, I’m also going to file this away in my bookmarks for later use.

    One of the main concerns I bring up every time one of the managers wants to throw “AI” at something because it’s trendy is “who’s responsible when it just makes something up?”.

    I know this is a Canadian ruling instead of a US one, but at least I can point to it and say “probably us”.

  • circuscritic@lemmy.ca
    link
    fedilink
    arrow-up
    44
    ·
    9 months ago

    This guy was only saved because he saved the screenshots.

    This may seem like overkill, but whenever I’m dealing with customer service for a significant financial concern, I always take detailed notes: time, date, name, what was discussed, and screenshots when relevant.

    Just a few years ago a large appliance company tried to dick me around on a faulty product that was several thousand dollars.

    It took weeks of back and forth, but when I was able to make contact with a layer of management just outside and above their customer service channel with my very long and detailed log of their bullshit, they authorized a check the next day.

    • TheEntity@kbin.social
      link
      fedilink
      arrow-up
      12
      ·
      9 months ago

      Are screenshots even still considered evidence? They should be absolutely trivial to manipulate.

      • Otter@lemmy.ca
        link
        fedilink
        English
        arrow-up
        25
        ·
        9 months ago

        It wouldn’t be direct evidence, but it might put it on the company to provide their logs. If they didn’t retain them, then that could affect the ruling

      • specialdealer@lemm.ee
        link
        fedilink
        arrow-up
        13
        ·
        9 months ago

        If you’re willing to commit outright fraud, lots of crime is easy. But also the penalty for ever being discovered can be severe.

        They also didn’t deny it happened. They would check logs before paying out.

      • mozz@mbin.grits.dev
        link
        fedilink
        arrow-up
        4
        ·
        edit-2
        9 months ago

        Yes, but submitting a manipulated screenshot in court is a whole different type of fraud and can carry whole different types of penalties than just claiming X happened when it was Y, so very reasonably puts the onus back on the company to look through their logs and try to prove what’s in your screenshots didn’t happen that way.

  • Computerchairgeneral@kbin.social
    link
    fedilink
    arrow-up
    21
    ·
    9 months ago

    Good. I’m sure the chatbot will be back up and running soon, but anything that reminds companies there are risks to replacing humans with “AI-enhanced” chatbots is good. Unfortunately, I’m sure the lesson companies are going to take away from this is to include a disclaimer that the chatbot isn’t always correct. Which kind of defeats the whole point of using a chatbot to me. Why would I want to use something to try and solve a problem that you just told me could give me inaccurate information?

    • bedrooms@kbin.social
      link
      fedilink
      arrow-up
      1
      ·
      edit-2
      9 months ago

      Expedia does it right. Just stop stupid customers at the brigade, and connect people with the real need immediately to real people.

      Amazon is similar but the real people there are useless as fuck in my country. They’re foreign part timers barely speaking the language if my country… Can’t do anything specific.

  • intrepid@lemmy.ca
    link
    fedilink
    arrow-up
    4
    ·
    9 months ago

    There are two disturbing tendencies being demonstrated here:

    1. Using useless AI to engage and disperse complaining customers. The AI can’t make meaningful solutions to many customer complaints. But companies use it to annoy the customers into giving up, so that they can save the cost of real customer support.
    2. Either blaming the AI or insisting that it’s right, when it makes a mistake. AI by nature is biased and unpredictable. But that doesn’t stop the companies from saying ‘the computer says so’.

    These companies need a few high profile hefty penalties as a motivation to avoid such dirty tricks.

    • RickRussell_CA@beehaw.org
      link
      fedilink
      English
      arrow-up
      4
      ·
      edit-2
      9 months ago

      \3. Asserting that their IT system is a “separate legal entity” and that they are not responsible for the accuracy of the system. They are eating legal loco weed.

  • AutoTL;DR@lemmings.worldB
    link
    fedilink
    English
    arrow-up
    4
    arrow-down
    1
    ·
    9 months ago

    🤖 I’m a bot that provides automatic summaries for articles:

    Click here to see the summary

    On the day Jake Moffatt’s grandmother died, Moffat immediately visited Air Canada’s website to book a flight from Vancouver to Toronto.

    In reality, Air Canada’s policy explicitly stated that the airline will not provide refunds for bereavement travel after the flight is booked.

    Experts told the Vancouver Sun that Moffatt’s case appeared to be the first time a Canadian company tried to argue that it wasn’t liable for information provided by its chatbot.

    Last March, Air Canada’s chief information officer Mel Crocker told the Globe and Mail that the airline had launched the chatbot as an AI “experiment.”

    “So in the case of a snowstorm, if you have not been issued your new boarding pass yet and you just want to confirm if you have a seat available on another flight, that’s the sort of thing we can easily handle with AI,” Crocker told the Globe and Mail.

    It was worth it, Crocker said, because “the airline believes investing in automation and machine learning technology will lower its expenses” and “fundamentally” create “a better customer experience.”


    Saved 81% of original text.

    • A1kmmA
      link
      fedilink
      English
      arrow-up
      5
      ·
      9 months ago

      Ironically the bot summary missed the crucial point that Air Canada’s chatbot gave inaccurate information.

  • jarfil@beehaw.org
    link
    fedilink
    arrow-up
    3
    ·
    9 months ago

    Air Canada essentially argued that “the chatbot is a separate legal entity that is responsible for its own actions,”

    Another step back for the AI Liberation Front… can’t file patents, can’t own copyrights, can’t be a legal entity, can’t incorporate… what’s next, denying AI sentience? This dehumanizing and discrimination against AIs needs to stop. 🤡