• Vespair@lemm.ee
    link
    fedilink
    arrow-up
    1
    ·
    32 minutes ago

    Fun time to remind everyone that pigeons aren’t native to the Americas, and were brought over originally by the French primarily as a food source (although also as messengers; carrier pigeons).

    “Squab” on a menu is pigeon.

    • Sjmarf@sh.itjust.worksOP
      link
      fedilink
      arrow-up
      1
      ·
      2 hours ago

      Google may not have enabled them in your region. Here in the UK they just appeared for me one day, a few months after I initially saw screenshots of them online. I didn’t do anything to enable them.

    • daniskarma@lemmy.dbzer0.com
      link
      fedilink
      arrow-up
      8
      arrow-down
      24
      ·
      edit-2
      7 hours ago

      Do the search yourself you’ll see that you get a better answer.

      Question: How to clean a rescued pidgeon

      ChatGPT: Cleaning a rescued pigeon should be done carefully to avoid stressing or harming the bird. Here’s a step-by-step guide to help you:

      1. Assess the Pigeon’s Condition Check for injuries: If the pigeon is injured, weak, or sick, it’s best to seek veterinary help before attempting to clean it. Observe behavior: If the pigeon is acting stressed or aggressive, you may need to wait until it calms down.
      2. Prepare the Cleaning Area Warm water: Fill a basin or sink with lukewarm water, not hot or cold. Mild soap: Use a gentle soap or baby shampoo without strong fragrances or chemicals. Towels: Have clean towels ready for drying the pigeon afterward.
      3. Handling the Pigeon Safely Wear gloves: This can protect both you and the bird. Secure hold: Hold the pigeon gently but firmly. Support its body with one hand and keep its wings close to its body to prevent flapping.
      4. Cleaning the Pigeon Partial cleaning: Focus on specific dirty areas rather than submerging the whole bird if possible. Use a damp cloth: For lightly soiled areas, use a damp cloth or sponge to wipe the feathers gently. Bath for heavier dirt: If a full wash is needed, place the pigeon in the water up to its chest. Avoid getting its head and neck wet unless absolutely necessary.
      5. Rinsing Rinse thoroughly: If you used soap, make sure to rinse all of it out of the feathers using clean, lukewarm water. Avoid the face: Be careful not to get water in the pigeon’s eyes, nose, or ears.
      6. Drying the Pigeon Pat dry: Gently pat the pigeon with a towel to remove excess water. Avoid rubbing as it could damage the feathers. Warm place: Keep the pigeon in a warm, draft-free area until it is completely dry. You can use a hairdryer on the lowest setting if the pigeon tolerates it, but keep it at a safe distance to avoid overheating.
      7. Post-Cleaning Care Observe behavior: Watch the pigeon for signs of stress or discomfort after the bath. Provide food and water: Ensure the pigeon has access to fresh water and a safe, comfortable place to rest. Additional Tips Do not use strong disinfectants or chemicals that can harm the bird’s sensitive skin and feathers. Limit baths to when absolutely necessary, as frequent cleaning can strip the natural oils from the pigeon’s feathers. If the pigeon seems injured or unwell, contact a wildlife rescue center or avian veterinarian for guidance.
      • bitjunkie@lemmy.world
        link
        fedilink
        arrow-up
        2
        ·
        3 hours ago

        Unsurprising that GPT is better given that OpenAI has been working on it and training it for ages and the MS partnership made every other big boy feel the need to rush some garbage model to market

        • daniskarma@lemmy.dbzer0.com
          link
          fedilink
          arrow-up
          8
          arrow-down
          6
          ·
          edit-2
          6 hours ago

          Could you be a little more constructive and point me at the points that are wrong and useless?

          Thank you.

          • ben_dover@lemmy.world
            link
            fedilink
            arrow-up
            23
            arrow-down
            1
            ·
            5 hours ago

            i think the point is that the answer is not reliable. it might be completely correct or borderline wrong, or something in between, and there’s no way to tell without verifying everything it says - and then one could look it up oneself in the first place already.

            • daniskarma@lemmy.dbzer0.com
              link
              fedilink
              arrow-up
              1
              arrow-down
              4
              ·
              edit-2
              2 hours ago

              Same as most human iterations then?

              At least I fact check everything I read. Like I did with this post and the church of the anti-AI got angry they got fact checked.

            • person420@lemmynsfw.com
              link
              fedilink
              arrow-up
              2
              arrow-down
              12
              ·
              5 hours ago

              If you’re using AI verbatim without looking up answers and verifying results, then that’s on you.

              When you Google something, do you take the first result and just assume it’s fact? You shouldn’t for AI either.

              • Grazed@lemmy.world
                link
                fedilink
                arrow-up
                7
                ·
                3 hours ago

                The main problem I see is that Google just shouldn’t include AI results. And they definitely shouldn’t put their unreliable LLM front and center on the results page. When you google something, you want accurate information, which the LLM might have, but only if that data was readily available to begin with. So the stuff it can help with is stuff the search would put first already.

                For anything requiring critical thought or research, the LLM will often hallucinate or misrepresent. The danger is that people do not always apply critical thinking. Defaulting to showing an LLM response is extremely dangerous, and it’s basically pointless.

                • person420@lemmynsfw.com
                  link
                  fedilink
                  arrow-up
                  1
                  ·
                  48 minutes ago

                  I don’t know. I find it to be a helpful tool. There’s definitely times it’s wrong (very very wrong sometimes) and there’s sometimes it’s right. It’s up to the user to figure that out.

                  Maybe I’m old and cynical, but I don’t take anything I read on the Internet, especially something automatically generated, at face value. It’s just another tool I could use to help get to the answer I’m looking for.

                • gamermanh@lemmy.dbzer0.com
                  link
                  fedilink
                  English
                  arrow-up
                  1
                  arrow-down
                  1
                  ·
                  3 hours ago

                  Honestly? It’s a great place to start, especially with every search engine being worse than anything pre-2018

                  I used to have to post my error codes to a forum if googling them didn’t immediately get my anywhere and pray someone would reply something actually useful some day

                  Now I can ask ChatGPT to point me at something and go from there. If it assumes wrongly about anything I can correct it rather simply. Its really good at turning documentation written by somebody who hasn’t spoken to another human in 15 years into something my stupid ass can better understand, too

                  AI is a powerful as shit tool, people who slag it as not having any utility are about as ignorant as the people saying it’s the second coming of Jesus himself

                  I’m personally planning to host a local model to avoid supporting commercial shit, but that’s a project for down the line rn

          • MrMcGasion@lemmy.world
            link
            fedilink
            arrow-up
            1
            ·
            4 hours ago

            While I’m not the person you replied to and don’t know what their argument would be, I’ll take a shot at giving my own answer. In many cases when people post examples of AI giving unhelpful or bad information, there’s often someone who runs off to their favorite LLM to see if it gives a better result, and it usually does, so it gets treated like user error for using the wrong LLM or not wording the prompt properly. When in other examples that person’s favorite LLM which gave the correct answer this time, is the bad example hallucinating or mixing unrelated concepts, and other people are in the comments promoting other LLMs that gave them a good reply this time. None of the LLMs are actually trustworthy consistently enough to be trusted alone, and you won’t really know what answer is trustworthy unless you ask several LLMs and then go research the topic on your own anyway to figure out which answer is the most correct. It’s a valid point that ChatGPT got the answer more right than Gemini this time, but it’s somewhat useless to know that because other times ChatGPT is the one hallucinating wildly, and Gemini has the right answer, but since they’ve all been wrong before who do you trust.

            LLMs are like asking an arrogant person who thinks they know everything, who rather than admitting what they don’t know, will pull an answer out of their butt, and while it might be a logical answer, it isn’t based in reality, and may still be wildly wrong. If you already mostly know the answer, maybe asking the arrogant person works, because you already know enough to know if they are speaking from their actual knowledge or making up an answer, but if you don’t already have knowledge on a topic, you won’t know whether the arrogant person is giving useful information or not.

            • LifeInMultipleChoice@lemmy.dbzer0.com
              link
              fedilink
              English
              arrow-up
              1
              arrow-down
              2
              ·
              3 hours ago

              It’s a tool, if you misuse a tool you can get hurt, if you use it right it can make your task easier. Ultimately it will likely cause problems in the future but the answer is easy, stop using Google and move to another search engine. You won’t get the responses from Googles ai then. Enough people do so, Google will change it maybe.

  • ikidd@lemmy.world
    link
    fedilink
    English
    arrow-up
    28
    ·
    8 hours ago

    If you asked me to clean a bird, this is exactly what I would think of. Though usually it involves removing lead shot, not labels.

    • Qwazpoi@lemmy.world
      link
      fedilink
      arrow-up
      15
      ·
      8 hours ago

      I think it took note of it being about rescue pigeons and assumed they would have a tag on them and then gave cooking directions for them

  • RagnarokOnline@programming.dev
    link
    fedilink
    arrow-up
    129
    arrow-down
    2
    ·
    12 hours ago

    lol, definitely missed some important context.

    I guess it thought OOP meant “clean” as in how do you dress the bird before you cook it. (As in: “clean a fish” means to filet a fish and prep it for cooking.)

    • jaybone@lemmy.world
      link
      fedilink
      arrow-up
      4
      ·
      8 hours ago

      But first it said they are usually clean. So that can’t be the context. If there was a context. But there is no context because AI is fucking stupid and all these c-suite assholes pushing it like their last bowel movement will be eating crow off of their golden parakeet about two years from now when all this nonsense finally goes away and the new shiny thing is flashing around.

      • jj4211@lemmy.world
        link
        fedilink
        arrow-up
        3
        ·
        edit-2
        6 hours ago

        There are signs of three distinct interpretations in the result:

        • On topic, the concept of cleaning a wild bird you are trying to save
        • Preparing a store bought Turkey (removing a label)
        • Preparing a wild bird that is caught

        It’s actually a pretty good illustration of how AI assembles “information shaped text” and how smooth it can look and yet how dumb it can be about it. Unfortunately advocates will just say “I can’t get this specific thing wrong when I ask it or another LLM, so there’s no problem”, even as it gets other stuff wrong. It’s weird as you better be able to second guess the result, meaning you can never be confident in an answer you didn’t already know, but when that’s the case, it’s not that great for factual stuff.

        For “doesn’t matter” content, it may do fine (generated alternatives to stock photography, silly meme pictures, random prattle from background NPCs in a game), but for “stuff that matters”, Generative AI is frequently more of a headache than a help.

  • PyroNeurosis@lemmy.blahaj.zone
    link
    fedilink
    English
    arrow-up
    49
    arrow-down
    1
    ·
    12 hours ago

    I like the idea that these are scientist-monitored birds that you remove the tag from before dressing and eating.

    • rockstarmode@lemmy.world
      link
      fedilink
      arrow-up
      4
      ·
      edit-2
      7 hours ago

      This actually happens very frequently in the US. When hunters harvest a bird they report their kills in compliance with hunting regulations. If any of your birds have leg or neck bands you report that information as well. The bands have a tracking number on them, and scientists use them to monitor populations and migration patterns. It’s literally part of their plan.

      You get to keep the bands as well (I only have experience with banded geese and ducks). They’re a neat memento.

    • Gork@lemm.ee
      link
      fedilink
      arrow-up
      8
      ·
      8 hours ago

      It would make an interesting footnote in the scientific article that had this pigeon in the experiment.

      “1 This pigeon was removed from the experiment as it was eaten by a local Florida Man.”

  • Lost_My_Mind@lemmy.world
    link
    fedilink
    arrow-up
    38
    ·
    12 hours ago

    I’m going to build a bunch of cyborgs, who follow orders exclusively via googleAI.

    I figure once I release about 4 billion of them into the world, either google stops doing evil shit, or they do REALLY evil shit. We shall see what happens…

    • mindaika@lemmy.dbzer0.com
      link
      fedilink
      English
      arrow-up
      1
      ·
      8 hours ago

      It is kind of an Interesting idea: what does “statistical average morality” look like when it’s got a 3m tall power frame and a handheld howitzer?

  • 🇰 🌀 🇱 🇦 🇳 🇦 🇰 ℹ️@yiffit.net
    link
    fedilink
    English
    arrow-up
    6
    arrow-down
    1
    ·
    edit-2
    8 hours ago

    Be more specific with the question.

    “How do I wash a pigeon”

    Or put every single word in quotations so it actually searches the whole string and doesn’t truncate the query to just 2 or 3 keywords because this is clearly ignoring the “rescued” part.

  • Kecessa@sh.itjust.works
    link
    fedilink
    arrow-up
    21
    arrow-down
    6
    ·
    edit-2
    11 hours ago

    Pigeon = edible bird

    Cleaning a bird > preparing a bird after killing it (hunting term)

    AI figured the “rescued” part was either a mistake or that the person wanted to eat a bird they rescued

    If you make a research for “how to clean a dirty bird” you give it better context and it comes up with a better reply

    • DannyBoy@sh.itjust.works
      link
      fedilink
      English
      arrow-up
      25
      arrow-down
      3
      ·
      edit-2
      10 hours ago

      The context is clear to a human. If an LLM is giving advice to everybody who asks a question in Google, it needs to do a much better job at giving responses.

        • HighlyRegardedArtist@lemmy.world
          link
          fedilink
          arrow-up
          3
          ·
          7 hours ago

          I have to disagree with that. To quote the comment I replied to:

          AI figured the “rescued” part was either a mistake or that the person wanted to eat a bird they rescued

          Where’s the “turn of phrase” in this, lol? It could hardly read any more clearly that they assume this “AI” can “figure” stuff out, which is simply false for LLMs. I’m not trying to attack anyone here, but spreading misinformation is not ok.

          • Kecessa@sh.itjust.works
            link
            fedilink
            arrow-up
            1
            ·
            edit-2
            3 hours ago

            I’ll be the first one to explain to people that AI as we know it is just pattern recognition, so yeah, it was a turn of phrase, thanks for your concern.

            • HighlyRegardedArtist@lemmy.world
              link
              fedilink
              arrow-up
              1
              ·
              3 hours ago

              Ok, great to know. Nuance doesn’t cross internet well, so your intention wasn’t clear, given all the uninformed hype & grifters around AI. Being somewhat blunt helps getting the intended point across better. ;)

          • Rivalarrival@lemmy.today
            link
            fedilink
            English
            arrow-up
            1
            ·
            edit-2
            17 minutes ago

            My point wasn’t that LLMs are capable of reasoning. My point was that the human capacity for reasoning is grossly overrated.

            The core of human reasoning is simple pattern matching: regurgitating what we have previously observed. That’s what LLMs do well.

            LLMs are basically at the toddler stage of development, but with an extraordinary vocabulary.

    • FlorianSimon@sh.itjust.works
      link
      fedilink
      arrow-up
      13
      arrow-down
      2
      ·
      edit-2
      10 hours ago

      I like how you’re making excuses for something that it is very clear in context. I thought AI was great at picking up context?

      • lunarul@lemmy.world
        link
        fedilink
        arrow-up
        4
        ·
        9 hours ago

        I thought AI was great at picking up context?

        I don’t know why you thought that. LLMs split your question into separate words and assigns scores to those words, then looks up answers relevant to those words. It has no idea of how those words are relevant to each other. That’s why LLMs couldn’t answer how many "r"s are in “strawberry”. They assigned the word “strawberry” a lower relevancy score in that question. The word “rescue” is probably treated the same way here.

      • iAmTheTot@sh.itjust.works
        link
        fedilink
        arrow-up
        6
        ·
        9 hours ago

        I don’t think they are really “making excuses”, just explaining how the search came up with those steps, which what the OP is so confused about.

    • huginn@feddit.it
      link
      fedilink
      arrow-up
      4
      arrow-down
      1
      ·
      10 hours ago

      Let me take the tag off my bird then snap it’s wings back together

      • bluewing@lemm.ee
        link
        fedilink
        arrow-up
        2
        ·
        8 hours ago

        Bought in a grocery store - see squab - they are usually clean and prepped for cooking. So while the de-boning instructions were not good, the AI wasn’t technically wrong.

        But while a human can make the same mistake and many here just assume the question was about how to wash a rescued pigeon - maybe that’s not the original intent - what human can do that AI cannot is to ask for clarification to the original question and intent of the question. We do this kind of thing every day.

        At the very best, AI can only supply multiple different answers if a poorly worded question is asked or it misunderstands something in the original question, (they seem to be very bad at even that or simply can’t do it at all). And we would need to be able to choose the correct answer from several provided.