
Researchers from AIRI, the Higher School of Economics (HSE), and Innopolis University have uncovered a fascinating potential use for profane language within Artificial Intelligence. Specifically, it turns out that neural networks can be trained to significantly condense text by substituting extensive phrases with strong expletives, all while retaining the core message.
The team from these various Russian institutions were investigating how AI handles diverse content. They observed that in natural conversation, a single coarse term frequently substitutes an entire sentence, packing substantial meaning into a small package. Consequently, they decided to test whether this principle could be leveraged for text compression via neural networks, aiming to conserve computational power. They developed a tailored lexicon mapping standard phrases to their blunt equivalents and used this to train their model.
It is explicitly noted in the press release that the objective of this research is not to teach the AI to swear. The current iteration of the model refrains from employing obscenities when processing regular texts.
Looking ahead, the researchers intend to concentrate future efforts on curating higher-quality, more contextually rich datasets so that the model can distinguish these substitution scenarios with greater accuracy.
Discuss
Source: AIRI