This account is being kept for the posterity, but it won’t see further activity past February.

If you want to contact me, I’m at /u/lvxferre@mander.xyz

  • 1 Post
  • 156 Comments
Joined 3 years ago
cake
Cake day: April 9th, 2021

help-circle


  • The source that I’ve linked mentions semantic embedding; so does further literature on the internet. However, the operations are still being performed with the vectors resulting from the tokens themselves, with said embedding playing a secondary role.

    This is evident for example through excerpts like

    The token embeddings map a token ID to a fixed-size vector with some semantic meaning of the tokens. These brings some interesting properties: similar tokens will have a similar embedding (in other words, calculating the cosine similarity between two embeddings will give us a good idea of how similar the tokens are).

    Emphasis mine. A similar conclusion (that the LLM is still handling the tokens, not their meaning) can be reached by analysing the hallucinations that your typical LLM bot outputs, and asking why that hallu is there.

    What I’m proposing is deeper than that. It’s to use the input tokens (i.e. morphemes) only to retrieve the sememes (units of meaning; further info here) that they’re conveying, then discard the tokens themselves, and perform the operations solely on the sememes. Then for the output you translate the sememes obtained by the transformer into morphemes=tokens again.

    I believe that this would have two big benefits:

    1. The amount of data necessary to “train” the LLM will decrease. Perhaps by orders of magnitude.
    2. A major type of hallucination will go away: self-contradiction (for example: states that A exists, then that A doesn’t exist).

    And it might be an additional layer, but the whole approach is considerably simpler than what’s being done currently - pretending that the tokens themselves have some intrinsic value, then playing whack-a-mole with situations where the token and the contextually assigned value (by the human using the LLM) differ.

    [This could even go deeper, handling a pragmatic layer beyond the tokens/morphemes and the units of meaning/sememes. It would be closer to what @njordomir@lemmy.world understood from my other comment, as it would then deal with the intent of the utterance.]


  • Not quite. I’m focusing on chatbots like Bard, ChatGPT and the likes, and their technology (LLM, or large language model).

    At the core those LLMs work like this: they pick words, split them into “tokens”, and then perform a few operations on those tokens, across multiple layers. But at the end of the day they still work with the words themselves, not with the meaning being encoded by those words.

    What I want is an LLM that assigns multiple meanings for those words, and performs the operations above on the meaning itself. In other words the LLM would actually understand you, not just chain words.


  • Complexity does not mean sophistication when it comes to AI and never has and to treat it as such is just a forceful way to make your ideas come true without putting in the real effort.

    It’s a bit off-topic, but what I really want is a language model that assigns semantic values to the tokens, and handles those values instead of directly working with the tokens themselves. That would be probably far less complex than current state-of-art LLMs, but way more sophisticated, and require far less data for “training”.








  • It’s less complicated than it looks like. The text is just a poorly written mess, full of options (Fedora vs. Ubuntu, repo vs. no repo, stable vs. beta), and they’re explaining how to do this through the terminal alone because the interface that you have might be different from what they expect. And because copy-pasting commands is faster.

    Can’t I just download a file and install it? I’m on Ubuntu.

    Yes, you can! In fact, the instructions include this option; it’s under “Installing the app without the Mullvad repository”. It’s a bad idea though; then you don’t get automatic updates.

    A better way to do this is to tell your system “I want software from this repository”, so each time that they make a new version of the program, yours get updated.

    but I have no idea what I’m doing here.

    I’ll copy-paste their commands to do so, and explain what each does.

    sudo curl -fsSLo /usr/share/keyrings/mullvad-keyring.asc https://repository.mullvad.net/deb/mullvad-keyring.asc
    echo "deb [signed-by=/usr/share/keyrings/mullvad-keyring.asc arch=$( dpkg --print-architecture )] https://repository.mullvad.net/deb/stable $(lsb_release -cs) main" | sudo tee /etc/apt/sources.list.d/mullvad.list
    sudo apt update
    sudo apt install mullvad-vpn
    

    The first command boils down to “download this keyring from the internet”. The keyring is a necessary file to know if you’re actually getting your software from Mullvad instead of PoopySoxHaxxor69. If you wanted, you could do it manually, and then move to the /usr/share/keyrings directory, but… it’s more work, come on.

    The second command tells your system that you want software from repository.mullvad.net. I don’t use Ubuntu but there’s probably some GUI to do it for you.

    The third command boils down to “hey, Ubuntu, update the list of packages for me”.

    The fourth one installs the software.


  • Lvxferre@lemmy.mltoTechnology@lemmy.mlTransparent Aluminium
    link
    fedilink
    arrow-up
    97
    arrow-down
    7
    ·
    edit-2
    8 months ago

    Misleading name, on the same level as calling water “non-explosive hydrogen”. That said the material looks promising, as a glass replacement for some applications (the text mentions a few of them, like armoured windows).

    (It is not a metal; it’s a ceramic, mostly oxygen with bits and bobs of aluminium and nitrogen. Interesting nonetheless, even if I’m picking on the name.)


  • Lvxferre@lemmy.mltoLinux@lemmy.ml*Permanently Deleted*
    link
    fedilink
    arrow-up
    10
    arrow-down
    7
    ·
    8 months ago

    The first response seems reasonable for me; it’s informative and replying to an ambiguous comment, as you can’t quite know if “isn’t there” refers to his individual needs or in general.

    The second response is however passive aggressive garbage. Fl4ppers clarified that he was talking about his individual needs; notjustforhackers failed to take it into account, and his response sounds a lot like “I’m just sayin lol lmao… you liar”.


  • I don’t know (…or care, really) about USA so I’ll speak on more general grounds.

    There’s a lot of stuff in social media that makes it a great soapbox for social manipulation:

    • low cost, wide reaching: it’s easy to be heard
    • decontextualisation: it gives more room for assumers¹ to do their shit, and make an incorrect context out of nowhere.
    • virality: it’s easy to start a witch hunt. Cue to the pitchfork emporium / Twitter MC of the day.
    • upvote/like-based systems: people don’t upvote your content (increasing its visibility) because you’re right, they do it because you say it confidently.
    • on the Internet, nobody knows that you’re a dog: concern trolling made easy.

    Now look at what @startle@toast.ooo said: “Dunno man, seems like it might be the fascists.”. IMO that user is being spot on, those five things make social media specially easy to manipulate for fascists². And they’re mostly the ones creating this dichotomisation of society³, because that’s how they’re able to congregate the nutjobs into a political discourse. Suddenly the village idiot doesn’t simply say “they’re hiding aliens from us” (stupid, but morally OK), the discourse becomes “the Jews are hiding aliens from us” (stupid and Antisemitic).

    1. By “assumers” I mean individuals who are quick to draw conclusions based on little to no reasoning, evidence, or thought. This plague exists since the dawn of time, it’s just that decontextualisation gives them more room to assume shit out of nowhere.
    2. Fascists often babble about “virtue signalling”, without realising that themselves are prone to signal adherence to their stupid beliefs. They don’t want to be in the receiving end of their own witch hunts.
    3. By “society” I mean at the very least Western Europe plus the Americas; probably more. It is not exclusive to USA.




  • Not even parrots - the birds are actually smart.

    I’m not a lawyer but I can see a good way for lawyers to use ChatGPT: tell it to list laws that are potentially related to the case, then manually check those laws to see if they apply. This would work nicely in countries with Roman law; and perhaps in countries with tribal law too (the article is from USA), as long as the model is fed with older cases for precedent.

    And… really, that’s the best use for those bots IMO - asking it to sort, filter and search information from messy and large systems. Letting it write things for you, like those two lawyers did, is worse than laziness: it stinks stupidity.

    It’s also immoral. The lawyer is a human being, thus someone who can be held responsible for one’s actions; ChatGPT is not and, as such, it should not be in charge of decisions that affect human lives.