• Max-P@lemmy.max-p.me
    link
    fedilink
    arrow-up
    1
    ·
    10 months ago

    Probably an AI model that fits in that size. It might not be our best models, but it probably would be a lot more useful to aliens than whatever we’d decide to fit on 4GB.

    They’d get mostly all the inner workings of our languages and how we do conversations and generally be able to answer basic questions about humanity.

    • DogMuffins@discuss.tchncs.de
      link
      fedilink
      English
      arrow-up
      2
      ·
      10 months ago

      What ? “an AI model” is not a compression algorithm. Why give the aliens an AI trained with some wikipedia articles when you could just give them wikipedia.

      • Max-P@lemmy.max-p.me
        link
        fedilink
        arrow-up
        1
        arrow-down
        1
        ·
        edit-2
        10 months ago

        Because an LLM is more than just data: it’s like a big network of how syllables and words go together based on some context. And that’s useful because language is how we communicate, how we connect ideas together, it’s how we share stories. It’s not just Wikipedia articles, it’s a database of relationships between words and concepts. It approximates how we think as humans.

        Yes, AI is hella overhyped. Everyone wants to AI everything. But really for this particular situation, I think the model data would actually be the best precompiled database of knowledge we can possibly provide to learn about humans for the size.

        No it’s not magic compression, but 4GB worth of parameters is still a lot. GPT4All has models just under 4GB. They’re not particularly impressive compared to OpenAI’s offerings, but I think you can extract a lot more practical information to do first contact out of a basic model than 4GB worth of Wikipedia. It’s extremely lossy compression, it’s never gonna spit out articles vebatim, it will hallucinate a ton of stuff.

        If we had more space I’d send all the major AIs we have like Dall-E, LLaMa and GPT 4. Imagine you’re an alien, you’re presented with a keyboard and a monitor, and know nothing about us. You can use Dall-E to try random letters and words and see if the output makes sense. Maybe you find out what a cat, dog, bat, frog, apple looks like. You can then input those words in ChatGPT, and get context as to when those are used. What’s “a horse”? What’s “riding”? Put those into Dall-E, now you know what a “cat riding a horse” looks like. It can generate as many as you want, any combination. Eventually you can figure out how to ask ChatGPT if cats typically ride horses, cars, bycles, what do cats do.

        Now imagine you’re a very advanced alien species that can easily process the model’s parameters. You’ve just downloaded the basics of humanity. They can map their language to our model’s parameters, and basically speak to us in our language, and translate our answers to theirs, and basically have a basic conversation.

        • DogMuffins@discuss.tchncs.de
          link
          fedilink
          English
          arrow-up
          2
          ·
          10 months ago

          Sorry chief you haven’t really explained why an AI model would be the best format.

          It’s less dense than Wikipedia text. End of.