A team of Wikipedia volunteers has been quietly working in the background for years, cataloging examples of "AI tells" or common patterns that can be used to identify when an article has been written by an artificial intelligence. These patterns include overused phrases, overly formal language, and inconsistencies in grammar and syntax.
Now, a tech entrepreneur named Siqi Chen has developed a plugin called "Humanizer" for Anthropic's Claude Code AI assistant. This plugin takes the list of "AI tells" compiled by Wikipedia editors and turns it into a set of instructions that can be fed to the AI model. The goal is to instruct the AI to avoid using these patterns, making its writing sound more like human-written text.
The Humanizer plugin has gained over 1,600 stars on GitHub, indicating a high level of interest from developers and researchers in the field of artificial intelligence. Chen notes that the plugin is "really handy" because it allows users to tell their LLMs (large language models) to "not do that" when it comes to writing like an AI model.
However, some experts caution that while the Humanizer plugin can help with detecting AI-generated text, it's not a foolproof solution. Language models don't always perfectly follow instructions, and there are cases where using such plugins can actually harm coding ability or produce lower-quality output.
One of the main challenges in detecting AI-generated text is that human writing can be just as "chatbot-like" as machine-written text. This means that even if a piece of writing has been generated by an AI model, it may still pass through various quality checks and detection tools without being flagged as suspicious.
As a result, some researchers are advocating for a more nuanced approach to detecting AI-generated text, one that takes into account not just the surface-level patterns and phrasing used in AI-written content, but also the underlying factual accuracy and substance of the writing itself. This approach acknowledges that even high-quality human writing can sometimes be indistinguishable from machine-written content, especially when it comes to certain types of topics or styles.
In the end, the Humanizer plugin represents an important step forward in the ongoing cat-and-mouse game between AI developers and those who seek to detect and counteract AI-generated content. By providing a standardized set of instructions that can be used by language models, Chen's plugin has the potential to help make written communication more transparent and trustworthy β at least when it comes to identifying the source of the writing in question.
Now, a tech entrepreneur named Siqi Chen has developed a plugin called "Humanizer" for Anthropic's Claude Code AI assistant. This plugin takes the list of "AI tells" compiled by Wikipedia editors and turns it into a set of instructions that can be fed to the AI model. The goal is to instruct the AI to avoid using these patterns, making its writing sound more like human-written text.
The Humanizer plugin has gained over 1,600 stars on GitHub, indicating a high level of interest from developers and researchers in the field of artificial intelligence. Chen notes that the plugin is "really handy" because it allows users to tell their LLMs (large language models) to "not do that" when it comes to writing like an AI model.
However, some experts caution that while the Humanizer plugin can help with detecting AI-generated text, it's not a foolproof solution. Language models don't always perfectly follow instructions, and there are cases where using such plugins can actually harm coding ability or produce lower-quality output.
One of the main challenges in detecting AI-generated text is that human writing can be just as "chatbot-like" as machine-written text. This means that even if a piece of writing has been generated by an AI model, it may still pass through various quality checks and detection tools without being flagged as suspicious.
As a result, some researchers are advocating for a more nuanced approach to detecting AI-generated text, one that takes into account not just the surface-level patterns and phrasing used in AI-written content, but also the underlying factual accuracy and substance of the writing itself. This approach acknowledges that even high-quality human writing can sometimes be indistinguishable from machine-written content, especially when it comes to certain types of topics or styles.
In the end, the Humanizer plugin represents an important step forward in the ongoing cat-and-mouse game between AI developers and those who seek to detect and counteract AI-generated content. By providing a standardized set of instructions that can be used by language models, Chen's plugin has the potential to help make written communication more transparent and trustworthy β at least when it comes to identifying the source of the writing in question.