Powered by OpenAIRE graph
Found an issue? Give us feedback
image/svg+xml art designer at PLoS, modified by Wikipedia users Nina, Beao, JakobVoss, and AnonMoos Open Access logo, converted into svg, designed by PLoS. This version with transparent background. http://commons.wikimedia.org/wiki/File:Open_Access_logo_PLoS_white.svg art designer at PLoS, modified by Wikipedia users Nina, Beao, JakobVoss, and AnonMoos http://www.plos.org/ ZENODOarrow_drop_down
image/svg+xml art designer at PLoS, modified by Wikipedia users Nina, Beao, JakobVoss, and AnonMoos Open Access logo, converted into svg, designed by PLoS. This version with transparent background. http://commons.wikimedia.org/wiki/File:Open_Access_logo_PLoS_white.svg art designer at PLoS, modified by Wikipedia users Nina, Beao, JakobVoss, and AnonMoos http://www.plos.org/
ZENODO
Conference object
Data sources: ZENODO
addClaim

Unsmurfed: How an LLM Interprets the Smurfs' Distributional Language

Authors: Escouflaire, Louis;

Unsmurfed: How an LLM Interprets the Smurfs' Distributional Language

Abstract

We use the French model CamemBERT to replace every occurrence of the word schtroumpf- ("smurf-") with its contextually most probable word in a corpus of 300 pages (five albums) of the Belgian Smurfs comics. Our multimodal pipeline consists of comic-centered OCR, image captioning and automated token prediction. By generating ten versions of all speech bubbles in about 3000 panels (top-1 to 10 most likely predictions), our experiment exposes how an LLM performs the very task that enables humans to naturally understand the Smurfs’ playful language: inferring meaning from context.

Powered by OpenAIRE graph
Found an issue? Give us feedback