Skip to content
0
  • Home
  • Recent
  • Tags
  • Popular
  • World
  • Users
  • Groups
  • Home
  • Recent
  • Tags
  • Popular
  • World
  • Users
  • Groups
Skins
  • Light
  • Brite
  • Cerulean
  • Cosmo
  • Flatly
  • Journal
  • Litera
  • Lumen
  • Lux
  • Materia
  • Minty
  • Morph
  • Pulse
  • Sandstone
  • Simplex
  • Sketchy
  • Spacelab
  • United
  • Yeti
  • Zephyr
  • Dark
  • Cyborg
  • Darkly
  • Quartz
  • Slate
  • Solar
  • Superhero
  • Vapor

  • Default (Sketchy)
  • No Skin
Collapse

Wandering Adventure Party

  1. Home
  2. Uncategorized
  3. A thought that popped into my head when I woke up at 4 am and couldn’t get back to sleep…

A thought that popped into my head when I woke up at 4 am and couldn’t get back to sleep…

Scheduled Pinned Locked Moved Uncategorized
llmsclaudechatgpt
70 Posts 37 Posters 122 Views
  • Oldest to Newest
  • Newest to Oldest
  • Most Votes
Reply
  • Reply as topic
Log in to reply
This topic has been deleted. Only users with topic management privileges can see it.
  • Lars Marowsky-Brée 😷L Lars Marowsky-Brée 😷

    @etchedpixels @jzb I'm inclined to slightly disagree and think this is denialism (understandably, given how bad their ethics are; it'd be much easier if indeed they had no useful function).

    The problem is that, despite all the scenarios where they're inappropriate and wrong, they do.

    And we're unwilling (as a society) to fully consider their risks and costs, because "there's no glory in prevention".

    That's the challenge we need to overcome.

    DanielD This user is from outside of this forum
    DanielD This user is from outside of this forum
    Daniel
    wrote on last edited by
    #53

    @larsmb @etchedpixels @jzb in a $work context I've found llms quite good at automating what would otherwise be "find the plausible stack overflow answer and copy-paste it, changing the names" or "write a shit load of boilerplate" or "explain the awful mess that this module is and work out what it was supposed to be for" or even "do a refactor in less time than it would take me to figure out the LSP support in this language and do it myself".

    All things that should not be useful if we'd collectively made better choices, but given where we are now have value in context

    The Penguin of EvilE 1 Reply Last reply
    0
    • DanielD Daniel

      @larsmb @etchedpixels @jzb in a $work context I've found llms quite good at automating what would otherwise be "find the plausible stack overflow answer and copy-paste it, changing the names" or "write a shit load of boilerplate" or "explain the awful mess that this module is and work out what it was supposed to be for" or even "do a refactor in less time than it would take me to figure out the LSP support in this language and do it myself".

      All things that should not be useful if we'd collectively made better choices, but given where we are now have value in context

      The Penguin of EvilE This user is from outside of this forum
      The Penguin of EvilE This user is from outside of this forum
      The Penguin of Evil
      wrote on last edited by
      #54

      @dan @larsmb @jzb That's not really an LLM problem though - that's a very targetted problem being solved using an LLM as a large hammer, and a hammer that makes mistakes where formal methods and formal method dervied tools do not in general do

      As to "find the plausible stack overflow answer and copy-paste it, changing the names", part of my job at Intel was catching people doing this and dropping them in the shit. Automated versus wilful human copyright violation 😎

      DanielD 1 Reply Last reply
      0
      • Gabriel PettierT Gabriel Pettier

        @jzb i think the problem is more that workers have much greater work ethics than generally acknowledged, and if a tool allow them to work faster, they'll do more work, not not reclaim more time.

        but more than une explanation can be true at the same time.

        Wolf480plW This user is from outside of this forum
        Wolf480plW This user is from outside of this forum
        Wolf480pl
        wrote on last edited by
        #55

        @tshirtman @jzb
        it's also that if people who send work your way learn that you get it done quickly and reliably, they'll send work your way more often

        Gabriel PettierT 1 Reply Last reply
        0
        • Wolf480plW Wolf480pl

          @tshirtman @jzb
          it's also that if people who send work your way learn that you get it done quickly and reliably, they'll send work your way more often

          Gabriel PettierT This user is from outside of this forum
          Gabriel PettierT This user is from outside of this forum
          Gabriel Pettier
          wrote on last edited by
          #56

          @wolf480pl @jzb yeah, but they do mostly rely on workers honesty to learn that.

          And there is always more work to do, in my experience as a dev.

          Wolf480plW 1 Reply Last reply
          0
          • The Penguin of EvilE The Penguin of Evil

            @dan @larsmb @jzb That's not really an LLM problem though - that's a very targetted problem being solved using an LLM as a large hammer, and a hammer that makes mistakes where formal methods and formal method dervied tools do not in general do

            As to "find the plausible stack overflow answer and copy-paste it, changing the names", part of my job at Intel was catching people doing this and dropping them in the shit. Automated versus wilful human copyright violation 😎

            DanielD This user is from outside of this forum
            DanielD This user is from outside of this forum
            Daniel
            wrote on last edited by
            #57

            @etchedpixels @larsmb @jzb like I said, things that should not be useful but are.

            DanielD 1 Reply Last reply
            0
            • Gabriel PettierT Gabriel Pettier

              @wolf480pl @jzb yeah, but they do mostly rely on workers honesty to learn that.

              And there is always more work to do, in my experience as a dev.

              Wolf480plW This user is from outside of this forum
              Wolf480plW This user is from outside of this forum
              Wolf480pl
              wrote on last edited by
              #58

              @tshirtman @jzb They rely on shit getting done to learn that.

              If you finish a task early, that'll unblock your coworker who's been waiting for you to finish it, so they'll know you did it.

              If you start a task late, sure, you spend less time doing things, but do you get to relax for the first half of day, knowing that you have a backlog of things to do?

              1 Reply Last reply
              0
              • DanielD Daniel

                @etchedpixels @larsmb @jzb like I said, things that should not be useful but are.

                DanielD This user is from outside of this forum
                DanielD This user is from outside of this forum
                Daniel
                wrote on last edited by
                #59

                @etchedpixels @larsmb @jzb as an industry we've spent this many decades failing.to "sharpen the saw", is it surprising we're now all gung ho about the enchanted broadswords we've just been gifted? They're so much better at opening bottles than the old way!

                1 Reply Last reply
                0
                • Lars Marowsky-Brée 😷L Lars Marowsky-Brée 😷

                  @jzb Is is an inherent limitation of how LLMs currently exist and are implemented.
                  They do strive to minimize it through scale, but it's also a reason why they do get "creative" in their answers.
                  Like with any stochastic algorithm, they perform best if you can (cheaply) validate the result. e.g., does a program pass the tests still?

                  This is much harder for complex questions about the real world.

                  @em_and_future_cats

                  Florian Berger (privat)F This user is from outside of this forum
                  Florian Berger (privat)F This user is from outside of this forum
                  Florian Berger (privat)
                  wrote on last edited by
                  #60

                  @larsmb

                  Side note: I'd call them anything but 'creative'.

                  If anything, the behavior is better described as 'evasive', since the model effectively keeps talking, without any substantial data backing up what's being conveyed.

                  Or, as Hicks, Humphries and Slater put it: They're bullshitting.

                  https://doi.org/10.1007/s10676-024-09775-5

                  N 1 Reply Last reply
                  0
                  • Em & future cats 🇺🇦🐈🏳️‍🌈E Em & future cats 🇺🇦🐈🏳️‍🌈

                    @larsmb @jzb
                    I think if it’s a “closed system” where you feed it information and tell it to only use the information it has and say when it “comes up empty” it should be okay. And to speed up the process of citations that does seem useful. (He would also double check on the accuracy of it , like say if it says something is on page 34 it should be there otherwise it’s not valid)

                    mathewM This user is from outside of this forum
                    mathewM This user is from outside of this forum
                    mathew
                    wrote on last edited by
                    #61

                    @em_and_future_cats @larsmb @jzb Honestly, if I could use a local LLM and have it quickly answer something and show me where it found the answer, that would be great and a valid use of the technology. I haven’t found a good application that can do it, though.

                    Em & future cats 🇺🇦🐈🏳️‍🌈E 1 Reply Last reply
                    0
                    • Joe BrockmeierJ Joe Brockmeier

                      A thought that popped into my head when I woke up at 4 am and couldn’t get back to sleep…

                      Imagine that AI/LLM tools were being marketed to workers as a way to do the same work more quickly and work fewer hours without telling their employers.

                      “Use ChatGPT to write your TPS reports, go home at lunchtime. Spend more time with your kids!” “Use Claude to write your code, turn 60-hour weeks into four-day weekends!” “Collect two paychecks by using AI! You can hold two jobs without the boss knowing the difference!”

                      Imagine if AI/LLM tools were not shareholder catnip, but a grassroots movement of tooling that workers were sharing with each other to work less. Same quality of output, but instead of being pushed top-down, being adopted to empower people to work less and “cheat” employers.

                      Imagine if unions were arguing for the right of workers to use LLMs as labor saving devices, instead of trying to protect members from their damage.

                      CEOs would be screaming bloody murder. There’d be an overnight industry in AI-detection tools and immediate bans on AI in the workplace. Instead of Microsoft CoPilot 365, Satya would be out promoting Microsoft SlopGuard - add ons that detect LLM tools running on Windows and prevent AI scrapers from harvesting your company’s valuable content for training.

                      The media would be running horror stories about the terrible trend of workers getting the same pay for working less, and the awful quality of LLM output. Maybe they’d still call them “hallucinations,” but it’d be in the terrified tone of 80s anti-drug PSAs.

                      What I’m trying to say in my sleep-deprived state is that you shouldn’t ignore the intent and ill effects of these tools. If they were good for you, shareholders would hate them.

                      You should understand that they’re anti-worker and anti-human. TPTB would be fighting them tooth and nail if their benefits were reversed. It doesn’t matter how good they get, or how interesting they are: the ultimate purpose of the industry behind them is to create less demand for labor and aggregate more wealth in fewer hands.

                      Unless you happen to be in a very very small club of ultra-wealthy tech bros, they’re not for you, they’re against you. #AI #LLMs #claude #chatgpt

                      ☃️SnögubbenJ This user is from outside of this forum
                      ☃️SnögubbenJ This user is from outside of this forum
                      ☃️Snögubben
                      wrote on last edited by
                      #62

                      @jzb You make an excellent point, and also proving the fact that many of these tools simply do not work.

                      As for my own profession, the idea of replacing software engineers with energy hungry slop code machines is simply a way to cut down on staff during hard times, but making it look good to the stock market.

                      1 Reply Last reply
                      0
                      • mathewM mathew

                        @em_and_future_cats @larsmb @jzb Honestly, if I could use a local LLM and have it quickly answer something and show me where it found the answer, that would be great and a valid use of the technology. I haven’t found a good application that can do it, though.

                        Em & future cats 🇺🇦🐈🏳️‍🌈E This user is from outside of this forum
                        Em & future cats 🇺🇦🐈🏳️‍🌈E This user is from outside of this forum
                        Em & future cats 🇺🇦🐈🏳️‍🌈
                        wrote on last edited by
                        #63

                        @mathew @larsmb @jzb
                        My husband is using notebook llm (yes it’s google uhgggg 🙄) and he “says” that he can set it to “local” and only data he puts into it. I don’t really know how it works because I haven’t used it and I don’t plan to. So I can’t say for certain if this is actually true or not. 😕

                        Lars Marowsky-Brée 😷L 1 Reply Last reply
                        0
                        • Em & future cats 🇺🇦🐈🏳️‍🌈E Em & future cats 🇺🇦🐈🏳️‍🌈

                          @mathew @larsmb @jzb
                          My husband is using notebook llm (yes it’s google uhgggg 🙄) and he “says” that he can set it to “local” and only data he puts into it. I don’t really know how it works because I haven’t used it and I don’t plan to. So I can’t say for certain if this is actually true or not. 😕

                          Lars Marowsky-Brée 😷L This user is from outside of this forum
                          Lars Marowsky-Brée 😷L This user is from outside of this forum
                          Lars Marowsky-Brée 😷
                          wrote on last edited by
                          #64

                          @em_and_future_cats @mathew @jzb It is not. No LLM can ignore the data in the training set. And NotebookLM is definitely not a local instance.

                          1 Reply Last reply
                          0
                          • Florian Berger (privat)F Florian Berger (privat)

                            @larsmb

                            Side note: I'd call them anything but 'creative'.

                            If anything, the behavior is better described as 'evasive', since the model effectively keeps talking, without any substantial data backing up what's being conveyed.

                            Or, as Hicks, Humphries and Slater put it: They're bullshitting.

                            https://doi.org/10.1007/s10676-024-09775-5

                            N This user is from outside of this forum
                            N This user is from outside of this forum
                            Nicolás Alvarez
                            wrote on last edited by
                            #65

                            @flberger @larsmb is that the correct DOI link?

                            Florian Berger (privat)F 1 Reply Last reply
                            0
                            • Joe BrockmeierJ Joe Brockmeier

                              A thought that popped into my head when I woke up at 4 am and couldn’t get back to sleep…

                              Imagine that AI/LLM tools were being marketed to workers as a way to do the same work more quickly and work fewer hours without telling their employers.

                              “Use ChatGPT to write your TPS reports, go home at lunchtime. Spend more time with your kids!” “Use Claude to write your code, turn 60-hour weeks into four-day weekends!” “Collect two paychecks by using AI! You can hold two jobs without the boss knowing the difference!”

                              Imagine if AI/LLM tools were not shareholder catnip, but a grassroots movement of tooling that workers were sharing with each other to work less. Same quality of output, but instead of being pushed top-down, being adopted to empower people to work less and “cheat” employers.

                              Imagine if unions were arguing for the right of workers to use LLMs as labor saving devices, instead of trying to protect members from their damage.

                              CEOs would be screaming bloody murder. There’d be an overnight industry in AI-detection tools and immediate bans on AI in the workplace. Instead of Microsoft CoPilot 365, Satya would be out promoting Microsoft SlopGuard - add ons that detect LLM tools running on Windows and prevent AI scrapers from harvesting your company’s valuable content for training.

                              The media would be running horror stories about the terrible trend of workers getting the same pay for working less, and the awful quality of LLM output. Maybe they’d still call them “hallucinations,” but it’d be in the terrified tone of 80s anti-drug PSAs.

                              What I’m trying to say in my sleep-deprived state is that you shouldn’t ignore the intent and ill effects of these tools. If they were good for you, shareholders would hate them.

                              You should understand that they’re anti-worker and anti-human. TPTB would be fighting them tooth and nail if their benefits were reversed. It doesn’t matter how good they get, or how interesting they are: the ultimate purpose of the industry behind them is to create less demand for labor and aggregate more wealth in fewer hands.

                              Unless you happen to be in a very very small club of ultra-wealthy tech bros, they’re not for you, they’re against you. #AI #LLMs #claude #chatgpt

                              MamlenaM This user is from outside of this forum
                              MamlenaM This user is from outside of this forum
                              Mamlena
                              wrote on last edited by
                              #66

                              @jzb
                              I'm going to leave an alternative idea. As a Marketer, the value I see in AI is: "it tells people what to think" it's the ideal media type for propaganda. If you can control what ChatGPT replies which I found out is very easy, you control their brains. Just teach them to use it for everything instead of thinking or searching the data. Imo currently, the models don't work so well to save labour time but work well enough to answer short random questions so people can use it as Search bar.

                              1 Reply Last reply
                              0
                              • Joe BrockmeierJ This user is from outside of this forum
                                Joe BrockmeierJ This user is from outside of this forum
                                Joe Brockmeier
                                wrote on last edited by
                                #67

                                @riverpunk oooh. Apparently I'm a centaur. Cool. @pluralistic

                                1 Reply Last reply
                                0
                                • Joe BrockmeierJ Joe Brockmeier

                                  A thought that popped into my head when I woke up at 4 am and couldn’t get back to sleep…

                                  Imagine that AI/LLM tools were being marketed to workers as a way to do the same work more quickly and work fewer hours without telling their employers.

                                  “Use ChatGPT to write your TPS reports, go home at lunchtime. Spend more time with your kids!” “Use Claude to write your code, turn 60-hour weeks into four-day weekends!” “Collect two paychecks by using AI! You can hold two jobs without the boss knowing the difference!”

                                  Imagine if AI/LLM tools were not shareholder catnip, but a grassroots movement of tooling that workers were sharing with each other to work less. Same quality of output, but instead of being pushed top-down, being adopted to empower people to work less and “cheat” employers.

                                  Imagine if unions were arguing for the right of workers to use LLMs as labor saving devices, instead of trying to protect members from their damage.

                                  CEOs would be screaming bloody murder. There’d be an overnight industry in AI-detection tools and immediate bans on AI in the workplace. Instead of Microsoft CoPilot 365, Satya would be out promoting Microsoft SlopGuard - add ons that detect LLM tools running on Windows and prevent AI scrapers from harvesting your company’s valuable content for training.

                                  The media would be running horror stories about the terrible trend of workers getting the same pay for working less, and the awful quality of LLM output. Maybe they’d still call them “hallucinations,” but it’d be in the terrified tone of 80s anti-drug PSAs.

                                  What I’m trying to say in my sleep-deprived state is that you shouldn’t ignore the intent and ill effects of these tools. If they were good for you, shareholders would hate them.

                                  You should understand that they’re anti-worker and anti-human. TPTB would be fighting them tooth and nail if their benefits were reversed. It doesn’t matter how good they get, or how interesting they are: the ultimate purpose of the industry behind them is to create less demand for labor and aggregate more wealth in fewer hands.

                                  Unless you happen to be in a very very small club of ultra-wealthy tech bros, they’re not for you, they’re against you. #AI #LLMs #claude #chatgpt

                                  Their friend, Svavar,S This user is from outside of this forum
                                  Their friend, Svavar,S This user is from outside of this forum
                                  Their friend, Svavar,
                                  wrote on last edited by
                                  #68

                                  @jzb

                                  I've just shown my partner who is a therapist how to use a LLM to write some of the bullshit reports the insurance companies make her fill in.

                                  1 Reply Last reply
                                  0
                                  • Joe BrockmeierJ Joe Brockmeier

                                    @larsmb @em_and_future_cats Well, as designed, they are -- I'm not sure that's a built-in limitation of LLMs or not. To be fair, I am not an expert on the tech.

                                    As something of an aside...

                                    It would be really interesting if you could pair the natural language instruction input with predictable output.

                                    That is, for example -- if I could query, say, all the data in Wikipedia but get only accurate output. Or if you had something like Ansible with natural-language playbook creation.

                                    "Hey, Ansible -- I want a playbook that will install all of the packages I have currently installed and retain my dotfiles" (or something) and be guaranteed accurate output... that would be amazing.

                                    Except that I also worry about losing skills to do those things. I worry about the loss of incidental knowledge when researching if a computer can return *only* what you ask for and sacrifice accidental discovery.

                                    (I also still think search engines were something of a mistake and miss Internet directories. Yeah, I'm fun at parties....)

                                    Eve La FéeE This user is from outside of this forum
                                    Eve La FéeE This user is from outside of this forum
                                    Eve La Fée
                                    wrote on last edited by
                                    #69

                                    @jzb

                                    > It would be really interesting if you could pair the natural language instruction input with predictable output.

                                    It exists, it's mostly accurate and can learn from its mistakes. It's a bit expensive though. Its called a human assistant.

                                    @larsmb @em_and_future_cats

                                    1 Reply Last reply
                                    0
                                    • N Nicolás Alvarez

                                      @flberger @larsmb is that the correct DOI link?

                                      Florian Berger (privat)F This user is from outside of this forum
                                      Florian Berger (privat)F This user is from outside of this forum
                                      Florian Berger (privat)
                                      wrote on last edited by
                                      #70

                                      @nicolas17

                                      Huh, no. Something was missing. The correct one is https://doi.org/10.1007/s10676-024-09775-5 . Fixed it in the original post. Thanks!

                                      1 Reply Last reply
                                      0
                                      • Pteryx the Puzzle SecretaryP Pteryx the Puzzle Secretary shared this topic on

                                      Reply
                                      • Reply as topic
                                      Log in to reply
                                      • Oldest to Newest
                                      • Newest to Oldest
                                      • Most Votes


                                      • Login

                                      • Login or register to search.
                                      Powered by NodeBB Contributors
                                      • First post
                                        Last post