Create AI policies and procedures now!

Policies and procedures
Summary: Publishers should establish guidelines for internal AI use within their organizations. It suggests key considerations, such as whether AI can be used, preferences for specific AI models, disclosure of AI usage, and addressing bias and inaccuracies. Additionally, it highlights the need for a user’s guide to report and improve AI-related issues. The article suggests that these policies should involve input from corporate, legal, and AI users to effectively integrate AI within the publishing industry.

It’s guidelines now or lawsuits later.

I’ve advocated external-facing guidelines for artificial intelligence so that publishers can protect themselves against bots that scrape their content, train their models on it, and then use their own content to compete with them. But that’s not today’s topic.

Female robot reading paperPublishers also need internal guidelines about how AI can be used within their organization.

Bo Sacks distributed an article by Pierre de Villiers that makes this point pretty well. I’ll link to it below.

I’m a big fan of checklists, so I started to dream up a list of topics that such a policy ought to address.

  • Can AI be used at all? I know some companies that prohibit it altogether.
  • If AI can be used, is there a preference for one over another — e.g., ChatGPT over Bard. Or, rather, should there be a requirement that the results from one large language model be checked against another?
  • If an article does rely on AI, does that need to be disclosed to the fact-checker, the editor, or to the reader?
  • There are different levels of use of AI. Should these be defined and distinguished? For example, it might be okay to use AI for background, but not okay to use the text verbatim.
  • Are there special procedures to deal with AI’s known bias and inaccuracies? Along those lines, this quote from the article made me laugh. “If you’re The New York Times, you cannot publish things that are not checked for biases.” I guess that means the correct biases.

In addition to policies, publishers need a user’s guide.

  • As inaccuracies and biases are discovered, they should be reported to the people who use the technology so they can keep an eye out for them.
  • As people find ways to avoid such inaccuracy and bias, they should be reported. For example, it’s often a good idea to ask ChatGPT if what it just told you is true. It’s remarkable how many times it will catch its own errors.
  • Employees should share prompts and techniques and that work well for specific needs.

The user’s guide sounds like a user’s group, or a bulletin board for everyone on staff so they can collectively learn how to use AI more effectively.

Who should create these AI policies? I really liked this quote from the article.

“The key to an effective AI strategy, … is to combine the managerial ability of a company’s top structure with the creativity of those on the shop floor trying to find the best way to make AI work for them.
“You want to set a clear direction within your company, but you also want innovation and clever use cases and that seldom comes from the top of the organisation ….” “They come from the people using the technologies.”

A good policy, in my opinion, will need input from at least three groups. Corporate, legal, and the people who actually use the AI.

Links

Steffen Damborg: Publishers must urgently draw up AI guidelines

Leave a Reply

Your email address will not be published. Required fields are marked *