• 0 Posts
  • 24 Comments
Joined 1 year ago
cake
Cake day: June 22nd, 2023

help-circle








  • Thank you for writing the explanation! I still think that this doesn’t need a blockchain. Instances could broadcast user creation, so each instance could validate user age on its own (or ask other trusted instances when they first “saw” that user).

    Fundamentally, blockchain solves the problem that there is no central source of trust, but in the Fediverse people necesarily trust the instance that they sign up, so a blockchain can’t add much in my opinion.


  • I see. I’m not convinced that proving the account creation date makes much of a difference here. Obviously the instance records when you sign up, so you would only need this to protect against malicious instances. But if a spammer is manipulating their instance to allow them to spam more, you have a much bigger problem than reliably knowing their account creation date.








  • This article is full of errors!

    At its core, an LLM is a big (“large”) list of phrases and sentences

    Definitely not! An LLM is the combination of an architecture and its model parameters. It’s just a bunch of numbers, no list of sentences, no database. (Seems like the author confused the word “LLM” with the dataset of the LLM???)

    an LLM is a storage space (“database”) containing as many sample documents as possible

    Nope. This applies to the dataset, not the model. I guess you can argue that memorization happens sometimes, so it might have some features of a database. But it isn’t one.

    Additional data (like the topic, mood, tone, source, or any number of other ways to categorize the documents) can be provided

    LLMs are trained in an unsupervised fashion. Just sequences of tokens, no labels.

    Typically, an LLM will cover a single context, e.g. only social media

    I’m not aware of any LLM that does this. What’s the “context” of GPT-4?

    software developers have gone to great lengths to collect an unfathomable number of sample texts and meticulously categorize those samples in as many ways as possible

    The closest real thing is the RLHF process that is used to fine tune an existing LLM for a specific application (like ChatGPT). The dataset for the LLM is not annotated or categorized in any way.

    a GPT uses the words and proximity data stored in LLMs

    This is confusing. “GPT” is the architecture of the LLM.

    it’s still only able to combine words in ways that it has seen before from its LLM

    This isn’t accurate, depending on the temperature setting, an LLM can output literally any word at any time with a non-zero probability. It can absolutely produce things it hasn’t seen. Also the phrasing “from its LLM” suggests that the author misunderstood what an LLM is.

    Also I think it’s too simple to just assert that LLMs are not intelligent. It mostly depends on your definition of intelligence and there are lots of philosophical discussions to be had (see also the AI effect).




  • Yet often it was his own stubborn and uncompromising nature that defined his life – his choices paint a picture of a man who was unable to heed the words of others. This undendinly antagonistic nature cost him friends, honours and ultimately put him into the dark role of colonialist.

    He was “stubborn and uncompromising”, which makes him “antagonistic”, therefore a colonialist and racist. That’s a pretty low bar. I don’t think it makes sense to define racism in a way that makes all 19th century naturalist racist.