logoalt Hacker News

dev_hugepagesyesterday at 9:00 AM3 repliesview on HN

memorized: https://www.asciiart.eu/cartoons/spongebob-squarepants


Replies

irthomasthomasyesterday at 9:16 AM

Naturally. That's how LLMs work. During training you measure the loss, the difference between the model output and the ground-truth and try to minimize it. We prize models for their ability to learn. Here we can see that the large model does a great job at learning to draw bob, while the small model performs poorly.

show 4 replies
ricardobeatyesterday at 10:43 AM

For the model to have memorized the entire sequence of characters precisely, this must appear hundreds of times in the training data?

ginkoyesterday at 9:13 AM

Conveniently removed the artist's signature though.

show 2 replies