this post was submitted on 15 Apr 2025
25 points (100.0% liked)

Technology

38565 readers
194 users here now

A nice place to discuss rumors, happenings, innovations, and challenges in the technology sphere. We also welcome discussions on the intersections of technology and society. If it’s technological news or discussion of technology, it probably belongs here.

Remember the overriding ethos on Beehaw: Be(e) Nice. Each user you encounter here is a person, and should be treated with kindness (even if they’re wrong, or use a Linux distro you don’t like). Personal attacks will not be tolerated.

Subcommunities on Beehaw:


This community's icon was made by Aaron Schneider, under the CC-BY-NC-SA 4.0 license.

founded 3 years ago
MODERATORS
 
all 10 comments
sorted by: hot top controversial new old
[–] haverholm@kbin.earth 14 points 1 week ago (1 children)

FFS, our first verbal contact with nonhuman persons is going to cause a diplomatic crisis 🤦

[–] HurlingDurling@lemm.ee 8 points 1 week ago* (last edited 1 week ago) (1 children)

Something something tarifs, something something antisemitism

[–] haverholm@kbin.earth 6 points 1 week ago

Something speciesist Flipper jokes something.

[–] drspod@lemmy.ml 14 points 1 week ago (1 children)

It's just going to hallucinate bullshit. Because we have so much training data of conversations between humans and dolphins don't we?

[–] jarfil@beehaw.org 7 points 1 week ago (1 children)

This group has been studying dolphins since 1985 using a non-invasive approach to track a specific community of Atlantic spotted dolphins. The WDP creates video and audio recordings of dolphins, along with correlating notes on their behaviors.

Just the 40 years, for that specific group of dolphins.

[–] drspod@lemmy.ml 7 points 1 week ago (1 children)

So imagine the language model can produce grammatically correct and semantically meaningful dolphin language, how does it translate that to a human language?

The reason LLMs can do this for human languages is that we have an enormous corpus of Rosetta stones for every language that allow the model to correlate concepts in each language. The training data for human to dolphin is going to be just these “behavioural notes.”

So the outcome is that the bullshitting machine will bullshit the scientists that it knows what they’re saying when it’s actually just making stuff up.

It’s a big problem with LLMs that they very rarely answer, “I don’t know.”

[–] jarfil@beehaw.org 1 points 1 week ago

LLMs use a tokenizer stage to convert input data into NN inputs, then a de-tokenizer at the output.

Those tokens are not limited to "human language", they can as well be positions, orientations, directions, movements, etc. "Body language", or the flight pattern of a bee, are as tokenizable as any other input data.

The concepts a dolphin language may have, no matter what they are, could then be described in a human language, and/or matched to human words for the same description.

[–] n3m37h@lemmy.dbzer0.com 12 points 1 week ago

So long and thanks for all the fish

...but dolphins don't all speak the same language?