Wikibooks:Artificial Intelligence: Difference between revisions

From Wikibooks, open books for an open world
Jump to navigation Jump to search
Content deleted Content added
→‎Text generation: text criticism requires thought, so LLMs should not be used for this
→‎Text generation: outline generation is too close to content generation and could produce issues
Line 7: Line 7:
{{w|Large language model}}s (LLMs), often referred to as "AI chatbots" in certain contexts, can be beneficial. However, like human-generated text, machine-generated text can also contain errors, flaws, or even be entirely useless. In particular, requesting a language model to write a book or an essay can sometimes cause the production of complete fabrications, including fictitious references. The output may be biased, libel living people, infringe on copyrights, or simply be of poor quality. As such, LLMs '''may not be used to generate original material''' or ideas at Wikibooks, and their sources should not be blindly trusted.
{{w|Large language model}}s (LLMs), often referred to as "AI chatbots" in certain contexts, can be beneficial. However, like human-generated text, machine-generated text can also contain errors, flaws, or even be entirely useless. In particular, requesting a language model to write a book or an essay can sometimes cause the production of complete fabrications, including fictitious references. The output may be biased, libel living people, infringe on copyrights, or simply be of poor quality. As such, LLMs '''may not be used to generate original material''' or ideas at Wikibooks, and their sources should not be blindly trusted.


Because they are just language models, LLMs may only be used to '''assist''' with editing, adhering to the following guidelines:
Because they are simply language models, LLMs may only be used to '''assist''' with editing, adhering to the following guidelines:
# You may use LLMs as writing advisors, such as asking for outlines, asking how to improve a paragraph, etc. However, you should be aware that the information provided by the LLM can be unreliable or entirely incorrect. Exercise due diligence and common sense when deciding whether to incorporate the LLM's suggestions.
# You may use LLMs as writing advisors, such as asking how to improve a paragraph, etc. However, you should be aware that the information provided by the LLM can be unreliable or entirely incorrect. Exercise due diligence and common sense when deciding whether to incorporate the LLM's suggestions.
# You are ultimately responsible for the content generated by your use of an LLM. LLMs should not be used for tasks with which the editor does not have substantial familiarity, and their outputs should be rigorously scrutinized for both quality and compliance with all applicable policies. Editors who are not fully aware of the risks associated with LLMs and who are not able to overcome the limitations of these tools may not use them.
# You are ultimately responsible for the content generated by your use of an LLM. LLMs should not be used for tasks with which the editor does not have substantial familiarity, and their outputs should be rigorously scrutinized for both quality and compliance with all applicable policies. Editors who are not fully aware of the risks associated with LLMs and who are not able to overcome the limitations of these tools may not use them.
# You must document your use of an LLM and its purpose in the edit summary and on the talk page (see disclosure section below).
# You must document your use of an LLM and its purpose in the edit summary and on the talk page (see disclosure section below).

Revision as of 17:59, 24 May 2024

The following draft policy outlines the Wikibooks community's perspective on the use of artificial intelligence-generated content on this site.

Text generation

Large language models (LLMs), often referred to as "AI chatbots" in certain contexts, can be beneficial. However, like human-generated text, machine-generated text can also contain errors, flaws, or even be entirely useless. In particular, requesting a language model to write a book or an essay can sometimes cause the production of complete fabrications, including fictitious references. The output may be biased, libel living people, infringe on copyrights, or simply be of poor quality. As such, LLMs may not be used to generate original material or ideas at Wikibooks, and their sources should not be blindly trusted.

Because they are simply language models, LLMs may only be used to assist with editing, adhering to the following guidelines:

  1. You may use LLMs as writing advisors, such as asking how to improve a paragraph, etc. However, you should be aware that the information provided by the LLM can be unreliable or entirely incorrect. Exercise due diligence and common sense when deciding whether to incorporate the LLM's suggestions.
  2. You are ultimately responsible for the content generated by your use of an LLM. LLMs should not be used for tasks with which the editor does not have substantial familiarity, and their outputs should be rigorously scrutinized for both quality and compliance with all applicable policies. Editors who are not fully aware of the risks associated with LLMs and who are not able to overcome the limitations of these tools may not use them.
  3. You must document your use of an LLM and its purpose in the edit summary and on the talk page (see disclosure section below).

Translation

LLMs may not be used for translation of content. Please see Wikibooks:Content translation for further information.

Media

Most AI tools have the capacity to create media, particularly images, from prompts. If you are interested in uploading this media, please be aware of relevant policies on licensing at our sister project Wikimedia Commons or our local policy on images, depending on whether you are uploading it here or there.

Required disclosure

All content made with the help of an LLM must be explicitly marked as such in both the edit summary and the page's discussion page. The following information must be provided:

This applies to every instance of using AI content. If you create new prompts and incorporate them into a page multiple times, each instance must be documented, including on the talk page. Any user basing content on artificial intelligence suggestions who fails to include this explicit notice is subject to warnings and subsequent blocking.

Detection and enforcement

As of this policy's creation, there are no reliable, high-quality tools capable of detecting AI-generated materials. Instead, editors will have to be on the lookout for regular issues of quality in text, such as:

  • Syntactical or grammatical errors
  • Illogical or meaningless sentences
  • Word changes that inappropriately change the meaning of a sentence
  • Citations or sources that do not match a claim
  • Low-quality or flawed images

Use of copyright violation detectors (e.g. Earwig's Copyvio Detector) can be used to help identify text copied verbatim from online sources.

Because of the challenges in definitively identifying generative AI use, editors who contribute the above issues without obvious AI use should first be referred to other applicable policies, such as Wikibooks:Copyrights. It is also best to engage in good faith discussion first to determine and resolve the cause of the problematic content. If good faith discussion and guidance fails, or if blatant violation of this policy is found, problematic editors may be subject to warning and subsequent editing restrictions.