Doing the Lord’s work in the Devil’s basement

  • 0 Posts
  • 57 Comments
Joined 1 year ago
cake
Cake day: May 8th, 2024

help-circle





  • They do math, just in a very weird (and obviously not super reliable) way. There is a recent paper by anthropic that explains it, I can track it down if you’d be interested.

    Broadly speaking, the weights in a model will form sorts of “circuits” which can perform certain tasks. On something hard like factoring numbers the performance is probably abysmal but I’d guess the model is still trying to approximate the task somehow.





  • I don’t know any AI artists (as in someone who prompts a model and then calls the result a work of art), although most traditional artists i know have come to incorporate AI one way or another in their process.

    You don’t really hear about it because it’s all intermediate material used during the production phase. For example, as a hobbyist writer, one thing i struggle with is writing action scenes cause i don’t have visual memory and i tend to forget a lot about continuity and “spatial realism” (“this guy starts in this corner of the room so there’s no way he could grab that object at that point”, shit like that). With AI I can generate some kind of “story board” of my scene, which helps me write it much better. It’s just laid out visually in front of me and i catch a lot more details.

    Sometimes when i’m toying with an idea i’ll also have a model generate a few variations on it, with different points of view, writing style, focus etc… Even if the writing is mediocre, it gives me a really good idea of how each version could pan out, and whether an angle works or not. I’ll then select the angle that works best and rewrite it entirely from scratch.

    There’s nothing innovative about it, people have been using assistants to avoid tedious work forever. It’s just that before AI you had to, you know, be rich and able to actually pay for the labor.



  • When you read that stuff on reddit there’s a parameter you need to keep in mind : these people are not really discussing Lemmy. They’re rationalizing and justifying why they are not on Lemmy. Totally different conversation.

    Nobody wants to come out and say “I know mainstream platforms are shit and destroying the fabric of reality but I can’t bring myself to be on a platform except it is the Hip Place to Be”. So they’ll invent stuff that paints them in a good light.

    You’ll still see people claiming that Mastodon is unusable because you have to select an instance - even though you don’t have to, you can just type Mastodon on Google, click the first link, and create an account in 2 clicks. It’s been ages. But the people still using Twitter need the excuse because otherwise what does it make them?


  • Zos_Kia@lemmynsfw.comtoMicroblog Memes@lemmy.worlddeepseek
    link
    fedilink
    English
    arrow-up
    4
    ·
    3 months ago

    If you take into account the optimizations described in the paper, then the cost they announce is in line with the rest of the world’s research into sparse models.

    Of course, the training cost is not the whole picture, which the DS paper readily acknowledges. Before arriving at 1 successful model you have to train and throw away n unsuccessful attempts. Of course that’s also true of any other LLM provider, the training cost is used to compare technical trade-offs that alter training efficiency, not business models.






  • They have no ability to actually reason

    I’m curious about this kind of statement. “Reasoning” is not a clearly defined scientific term, in that it has a myriad different meanings depending on context.

    For example, there has been science showing that LLMs cannot use “formal reasoning”, which is a branch of mathematics dedicated to proving theorems. However, the majority of humans can’t use formal reasoning. This would make humans “unable to actually reason” and therefore not Generally Intelligent.

    At the other end of the spectrum, if you take a more casual definition of reasoning, for example Aristotle’s discursive reasoning, then that’s an ability LLMs definitely have. They can produce sequential movements of thought, where one proposition leads logically to another, such as answering the classic : “if humans are mortal, and Socrates is a human, is Socrates mortal ?”. They demonstrate the ability to do it beyond their training data, meaning they do encode in their weights a “world model” which they use to solve new problems absent from their training data.

    Whether or not this is categorically the same as human reasoning is immaterial in this discussion. The distinct quality of human thought is a metaphysical concept which cannot be proved or disproved using the scientific method.