Email or username:

Password:

Forgot your password?
Top-level
Ben Evans

@KathleenC @petrosyan @publictorsten The point is that no human did this - the AI tool that the social media person used did it "automatically".

It's an example of the harmful biases ("pictures of women should be sexually suggestive") encoded into every one of these damn things, and why none of them are fit for purpose.

14 comments
Yann 不停 Heurtaux :antifa:

@kittylyst @KathleenC @petrosyan @publictorsten The Social Media person was left, not trained enough, not supervised enough. Fire them, and most of all their Manager.

Dave "Wear A Goddamn Mask" Cochran :donor:

@shalf @kittylyst @KathleenC @petrosyan @publictorsten I understand where you're coming from, and you are right to be upset. This was not an act of malice, it was addressed professionally and responsibly by management once they became aware of the situation, and - based solely on the original post since that's all the context I have - the injured party appears satisfied with the outcome.

If we fired people for not noticing things that IT tools do, every time it happened, there would literally not be anyone left working in IT. No one in their right mind thinks an "expand image" plugin is going to sexualize the image's contents; would you punish someone for printing flyers that have hidden metadata from the printer (snopes.com/fact-check/househol) in them?

The correct course of action, imo, is to start exactly as they've done by removing the offending image (and apologizing to the injured party), and to then use that as an educational moment so that it doesn't happen a second time.

Now, if it DOES happen a second time, fuck 'em. Unless/until, though, let people learn from their mistakes.

@shalf @kittylyst @KathleenC @petrosyan @publictorsten I understand where you're coming from, and you are right to be upset. This was not an act of malice, it was addressed professionally and responsibly by management once they became aware of the situation, and - based solely on the original post since that's all the context I have - the injured party appears satisfied with the outcome.

Yann 不停 Heurtaux :antifa:

@dave_cochran Alright. Train better the SoMe worker after a solid warning. Fire their Manager.
To be clearer: reputation and legal risk are way underestimated in this industry. It has to change. Yes, that means people at responsibility and strategy level thinking AI is harmless and allowing or promoting its unsupervised use, mostly for cost reasons (and ignoring externalities), losing their jobs.

Dave "Wear A Goddamn Mask" Cochran :donor:

@shalf have you seen this report? nngroup.com/articles/computer-

admittedly, it's a few years old now but I'd bet all the money in all of my pockets against all the money in all of yours that the data would be essentially the same if it was done again today.

The short version is that people, as a group, are WAY worse at using computers than people, as a group, think. Like, by virtue of having and using a Mastodon account at all, you are probably in the top 20% or so of computer users worldwide.

The point being, punishing people for things they cannot reasonably be expected to know about or even to know that there's ANYTHING TO LEARN about them, is counterproductive at best, and actively harmful at worst.

It's probably worth keeping in mind that we're talking about the folks running a conference here, and not, like, Facebook. If anything, this could become a phenomenal talk at the very conference it happened at since they're talking about UI/UX stuff (if I'm remembering the OP correctly)!

I just don't think that people acting in good faith should be penalized for things that they had absolutely no reason to think would cause harm, y'know?

@shalf have you seen this report? nngroup.com/articles/computer-

admittedly, it's a few years old now but I'd bet all the money in all of my pockets against all the money in all of yours that the data would be essentially the same if it was done again today.

The short version is that people, as a group, are WAY worse at using computers than people, as a group, think. Like, by virtue of having and using a Mastodon account at all, you are probably in the top 20% or so of computer...

Yann 不停 Heurtaux :antifa:

@dave_cochran “Absolutely no reason to think would cause harm”, in 2024, in a design-oriented communications and events community and business. Alright. Let me elaborate: it was the Manager’s job to a. have heard and learned from Monteiro e.g. 11y ago (see Webstock 2013 iirc), b. have heard and learned about e.g. Gebru et al. stories and work. Defending workers is right, hence my leaning towards your point. Managers are paid to know better. About time for some accountability in those fields.

Yann 不停 Heurtaux :antifa:

@dave_cochran To be honest and fair, let in charge, I would fire producers of indoors conferences without neither masks mandates nor best-in-line air renewing and filtering infrastructures and practices too for being stupid and having learned nothing, to explain how I approach taking care of people attending events, or producing/volunteering/speaking there. YMMV.

Yann 不停 Heurtaux :antifa: replied to Dave "Wear A Goddamn Mask" Cochran :donor:

@dave_cochran Then you cannot not see the issue with a woman speaker being sexualized through automation and edition of a picture she provided but never consented to be edited, never being shown the edited version, and nothing internally preventing that to a. happen, b. not be thrown away in due time, c. ever be published.

How is it taking good care of your speaker community? How can it not show failure at a high level of core biz competencies? How can someone in charge not be fired?

Kathleen

@dave_cochran @shalf @kittylyst @petrosyan @publictorsten

Nope. No excuses for creating and posting a SEXUALIZED photo of this individual without permission. No excuses. This isn't "not noticing every single piddly thing IT tools ever do," you straw man you, this is altering a the professional photo of a woman to SEXUALIZE IT.

Wake up.

Kathleen

@dave_cochran @shalf @kittylyst @petrosyan @publictorsten Yeah because "hidden metadata" is exactly the same as showing a faked image they pretended were her BREASTS.
Do you even listen to yourself?

jelte

@kittylyst @KathleenC @petrosyan @publictorsten the change itself is such an example, yes, but the mistake is using the tool in the first place.

Rich Felker

@kittylyst @KathleenC @petrosyan @publictorsten A human did do it. A human decided to use the shit "AI" tool, and uncritically accepted the result that came out of it for use in the material they were posting.

What YOU are doing is exactly the reason bad people love "AI". You are failing to account for the human agency of how the tool was used, and laundering the blame onto the tool that can't be held accountable.

Amoshias

@kittylyst @KathleenC @petrosyan @publictorsten a human chose to use the tool.

You can't wash bad decisions by saying "an AI tool did it."

Karsten Johansson

@kittylyst @KathleenC @petrosyan @publictorsten The person who "did this" was the person who accepted the AI output even with the most obvious difference from the original.

AI is not to blame. Blindly accepting what it produces is to blame. There is a whole team that should have caught that at some point, but the one who specifically did the action should definitely be raked over the coals.

We've taught an entire generation to check their Wikipedia-found data before using it. Why did this change when it comes to AI output?

@kittylyst @KathleenC @petrosyan @publictorsten The person who "did this" was the person who accepted the AI output even with the most obvious difference from the original.

AI is not to blame. Blindly accepting what it produces is to blame. There is a whole team that should have caught that at some point, but the one who specifically did the action should definitely be raked over the coals.

Go Up