this post was submitted on 23 Dec 2025
34 points (94.7% liked)

TechTakes

2338 readers
34 users here now

Big brain tech dude got yet another clueless take over at HackerNews etc? Here's the place to vent. Orange site, VC foolishness, all welcome.

This is not debate club. Unless it’s amusing debate.

For actually-good tech, you want our NotAwfulTech community

founded 2 years ago
MODERATORS
all 16 comments
sorted by: hot top controversial new old
[–] Jayjader@jlai.lu 2 points 3 days ago

On the nth day of Christmas, my true love gave to meeeee–

An LLM in a pear tree?

[–] Soyweiser@awful.systems 10 points 4 days ago (1 children)

Talked to somebody who is really into chatbot roleplay (of the 'longer term stories with new fantasy characters' type), and he mentioned that he needs to take his characters stories and archetypes to different models every now and then as a sort of refresh, as the models tend to eventually converge into certain stuck patterns. First clue of this seems to be that the replies seem to start to become a similar pattern of text organization. Sorry if this is vague as it is second hand, but the main point is, text based LLMs prob also do this.

[–] dgerard@awful.systems 7 points 4 days ago (3 children)

oh yeah, Suno does the same, it has about 12 songs

[–] flaviat@awful.systems 11 points 4 days ago

clanker's dozen

[–] Soyweiser@awful.systems 7 points 4 days ago (1 children)

Wonder if this is some sort of pre model collapse sign.

[–] corbin@awful.systems 3 points 15 hours ago (1 children)

Nah, it's more to do with stationary distributions. Most tokens tend to move towards it; only very surprising tokens can move away. (Insert physics metaphor here.) Most LLM architectures are Markov, so once they get near that distribution they cannot escape on their own. There can easily be hundreds of thousands of orbits near the stationary distribution, each fixated on a simple token sequence and unable to deviate. Moreover, since most LLM architectures have some sort of meta-learning (e.g. attention) they can simulate situations where part of a simulation can get stuck while the rest of it continues, e.g. only one chat participant is stationary and the others are not.

[–] Soyweiser@awful.systems 1 points 11 hours ago
[–] pikesley@mastodon.me.uk 3 points 4 days ago

@dgerard @Soyweiser the Randy Newman record?

[–] blakestacey@awful.systems 13 points 5 days ago (3 children)

sports and action imagery (cluster 0), formal interior spaces (cluster 1), maritime lighthouse scenes (cluster 2), urban night scenes with atmospheric lighting (cluster 3), gothic cathedral interiors (cluster 4), pompous interior design (cluster 5), industrial and vintage themes (cluster 6), rustic architectural spaces (cluster 7), domestic scenes and food imagery (cluster 8), palatial interiors with ornate architecture (cluster 9), pastoral and village scenes (cluster 10), and natural landscapes and animals with dramatic lighting (cluster 11).

Revealed: World's shittiest "tag yourself" meme

[–] V0ldek@awful.systems 1 points 1 day ago

domestic scenes and food imagery (sitting on my ass at the PC ingesting industrial amounts of crisps)

[–] swlabr@awful.systems 6 points 4 days ago

sigh. 10 sun, 7 moon, 8 rising

[–] dgerard@awful.systems 4 points 4 days ago

honestly, whomst amongst us isn't Pompous Interior Design

[–] fullsquare@awful.systems 6 points 4 days ago (2 children)

so after putting together text to image and image to text idiot boxes, there appears to be small number of approximate sort of eigenvalues in there. does that even mean anything or has any consequences?

[–] blakestacey@awful.systems 10 points 4 days ago
[–] dgerard@awful.systems 9 points 4 days ago

as i said this is a completely unsurprising result, but it's amusing to know what the twelve templates actually are

I got an email from the author, he says the paper was a passing observation and he's surprised it's got as much attention as it has