• 2 Posts
  • 118 Comments
Joined 2 months ago
cake
Cake day: May 11th, 2024

help-circle






  • Hackworth@lemmy.worldtoLemmy Shitpost@lemmy.worldMildred
    link
    fedilink
    English
    arrow-up
    10
    arrow-down
    1
    ·
    19 hours ago

    Mildreds out here, tryin ta take ‘Mildred’ back, and y’all throwin shade like people ain’t have to live with dumbass fuckin names like Mildred. Leave Mildreds alone! They’ve got enough to deal with. Wait, they chose Mildred, like as an adult? In the present? Pfft.



  • Hackworth@lemmy.worldtopics@lemmy.worldArctic Circle (OC)
    link
    fedilink
    English
    arrow-up
    4
    arrow-down
    1
    ·
    1 day ago

    Claude, re-write my post.

    <Claude> Sure thing boss!

    Generative AI fascinates me, and I use it frequently for both work and leisure. This photo, however, is the first to make me feel a twinge of sadness, as it reminds me that I now have to question the authenticity of everything I see.

    Then I recall that even traditional photographs often undergo extensive post-processing, which somewhat alleviates my concern. To distract myself, I create an AI-generated video of the river flowing, momentarily forgetting my worries.

    However, I can’t help but notice that many of these AI generations are of poor quality. Generative AI is simultaneously wonderful and troubling - a double-edged sword of technology.

    Despite these mixed feelings, I must admit this is a beautiful photo.


  • Hackworth@lemmy.worldtopics@lemmy.worldArctic Circle (OC)
    link
    fedilink
    English
    arrow-up
    5
    arrow-down
    2
    ·
    edit-2
    1 day ago

    I love generative A.I. I use it all the time, professionally and recreationally. This photo is the first to make me sad that I have to question the veracity of everything now. Then I remember the amount of post-processing a photo like this usually goes through and feel less sad. Then I make an AI video of the river flowing and forget what I was worried about. But all these generations are kinda trash. Such wonderful, terrible things. Beautiful photo, though.











  • Sure, thanks for your interest. It’s an incomplete picture, but we can think of LLMs as an abstraction of all the meaningful connections within a dataset to a higher dimensional space - one that can be explored. That alone is an insane accomplishment that is changing some of the pillars of data analysis and knowledge work. But that’s just the contribution of the “Attention is All You Need” paper. Many implementations of modern generative AI combine LLM inference in agentic networks, with GANs, and with rules-based processing. Extracting connections is just one part of one part of a modern AI implementation.

    The emergent properties of GPT4 are enough to point toward this exponential curve continuing. Theory of mind (and therefore deception) as well as relational spatial awareness (usually illustrated with stacking problems) developed solely from increasing the parameter count describing the neural network. These were unexpected capabilities. As a result, there is an almost literal arms race on the hardware side to see what other emergent properties exist at higher model sizes. With some poetic license, we’re rending function from form so quickly and effectively that it’s seen by some as freeing and others as a sacrilege.

    Some of the most interesting work on why these capabilities emerge and how we might gain some insight (and control) from exploring the mechanisms is being done by Anthropic and by users at Hugging Face. They discovered that when specific neurons in Claude’s net are stimulated, everything it responds with will in some way become about the Golden Gate Bridge, for instance. This sort of probing is perhaps a better route to progress than blindly chasing more size (despite its recent success). But only time will tell. Certainly, Google and MS have had a lot of unforced errors fumbling over themselves to stay in what they think is the race.