7 comments

  • pedalpete 13 days ago
    I've found where LLMs can be useful in this context is around free-associations. Because they don't really "know" about things, they regularly grasp at straws or misconstrue intended meaning. This, along with the volume of language (let's not call it knowledge) result in the LLMs occasionally bringing in a new element which can be useful.
    • gotts 13 days ago
      Can you list some examples where free-associations from LLM were useful to you?
      • pedalpete 13 days ago
        A lot of where I've benefited is in some marketing language. Rarely, or almost never has ChatGPT come up with something and I've thought "that's exactly what we wanted", but through iterations, it's taken me down paths I might not have found myself.

        Unfortunately, ChatGPT doesn't have a good search interface, so I can't search through older chats, but I know when I was looking at re-naming our company, it didn't come up with our new name, but it lead me down a path which did lead to our name.

        I was trying to understand a patent, and we were looking at the algorithm which was being used. ChatGPT misunderstood how the algorithm worked, but pointed to it's knowledge of a similar algorithm which worked differently, but was better suited to our purposes.

        Calling this "free-association" may be taking some liberty. Many people would consider these errors, or hallucinations, but in some ways, they do look very similar to what many would call free-association IMO.

      • PeterStuer 13 days ago
        Long, long time ago (1999, before LLM's) I made a virtual museum exhibit creator for education. The collection explorer created a connected graph where the nodes were the works of art and the edges were based on commonalities from their textual descriptions. It used very rudimentary language technology so it 'suffered' from things like homographs. Rather than being seen as a problem, the users liked the serendipity it brought for ideation.

        I assume free but not random association could be a comparable support for ideation in research.

      • bongodongobob 13 days ago
        Assume free-associations = hallucinations. Assume hallucinations are exactly what makes LLMs useful and your question can be rephrased as "Can you list some examples where LLMs were useful to you?"
        • firewolf34 13 days ago
          Is not the purpose of a model to interpolate between two points? This is the underlying basis of "hallucinations" (when that works out /not/ in our favour) or "prediction" (when it does). So it's a matter of semantics and a bit of overuse of the term "hallucination". But the model would be useless as nothing more than a search engine if it were to just regurgitate it's training data verbatim.
        • ec109685 13 days ago
          Hallucinations are lies. So not the same thing.
          • Teleoflexuous 13 days ago
            For LLM to lie it would need to know the truth. That's an incredible level of anthropomorphization.
          • malux85 13 days ago
            Hallucinations are not always lies, they are more like a transformation in the abstraction space.
            • throwup238 13 days ago
              That is some weapons grade spin :-)
          • littlestymaar 13 days ago
            All lies aren't useless, some can be insightful even when blatantly wrong in themselves (for instance: taken literally every scientific model is a lie). I can definitely see how an LLM hallucinating can helps fostering creativity (the same way psychedelics can), even if all they say is bullshit.
          • bongodongobob 13 days ago
            I'm using hallucination to mean "not exactly the thing", not outright lying. So maybe the "truth" is "My socks are wet." A hallucination could be "My socks are damp."
          • HeatrayEnjoyer 13 days ago
            Lies require intent. I can ask a model to lie and it will provide info it knows is inaccurate, and can provide the true statement if requested.

            Hallucinations are inaccuracies it doesn't realize are inaccurate.

    • robwwilliams 13 days ago
      This approach is already useful in functional genomics. A common type of question requires analysis of hundreds of potentially functional sequence variants.

      Hybrid LLM+ approaches are beginning to improve efficiency of ranking candidates and even proposing tests and soon I hope—higher order non-linear interactions among DNA variants.

    • deegles 13 days ago
      I like thinking of LLMs as "word calculators." Which I think really encapsulates how they aren't "intelligent" as the marketing would have you believe but also show how important the inputs are.
  • KhoomeiK 13 days ago
    A group of PhD students at Stanford recently wanted to take AI/ML research ideas generated by LLMs like this and have teams of engineers execute on them at a hackathon. We were getting things prepared at AGI House SF to host the hackathon with them when we learned that the study did not pass ethical review.

    I think automating science is an important research direction nonetheless.

    • srcreigh 13 days ago
      That’s pretty wild. What was the reason behind failing ethics review?
      • robbomacrae 13 days ago
        I'm generally a proponent of AI and LLM but to me the decision was the right one. You are tasking people with implementing an idea generated by an algorithmic model with (I'm guessing) zero oversight that might have very little training that teaches it the importance of coming up with ideas worth implementing. Some may be more useful than others so it won't be fair from an accomplishment or motivation point of view.

        Imagine you've already invested time going to this event and want to win the prize/credit but to do so you have to implement a plugin that makes webpages grayscale because of a random idea generator. Maybe some people would find that interesting but others would see it as wasting their time.

        • jimmySixDOF 13 days ago
          Individual ideas can be subject to Ethical Review Board approvals and that should go for a hackathon project same as any study proposed in Academia or drug trial etc -- but to apply some wavey handed lum sum out of bounds lable just based on source seems like arbitrary opinionated overreach.
        • golol 13 days ago
          As long as all participants are well-informed then there is absolutely no ethical issue...
          • rsfern 13 days ago
            How do you make sure the participants are well informed? What if an idea suggested by a model turns out to be dangerous to implement, but nobody at the hackathon has quite the relevant experience to notice?
            • naasking 13 days ago
              Such as?
              • HeatrayEnjoyer 13 days ago
                rsfern is asking exactly that
                • naasking 13 days ago
                  No, I'm asking for an example of an idea that an LLM might produce that is too dangerous to implement but nobody at the hackathon has the relevant experience to notice. You can shut down any endeavour by imagining boogeymen that aren't actually real.
                  • rsfern 12 days ago
                    I don’t think anything needs to be shut down necessarily, I’m just suggesting reasons why a reasonable ethics board might be hesitant to green light such a hackathon if it’s not clear the organizers have done their due diligence on safety

                    I might be biased in terms of the safety profile, my background is materials and chemistry, and there are loads of ways you can get into trouble if you don’t really have experience in the materials and synthesis routes you’re working with

                    One example I’ve heard of from my field (alloy design) is an ML model that suggested an composition high in Magnesium - perfectly reasonable if you’re interested in lightweight strong alloys, but the synthesis method was arc melting, which is a high risk for starting a metal fire if you aren’t careful because Mg has a low vapor pressure

                    If you’re doing organic chemistry it’s maybe even worse because there can be all kinds of side products or runaway exothermic reactions, and if you’re doing novel chemistry it might take deep experience in the field to know of those things are likely

                    All these concerns are manageable, but I think an ethics review panel would want to at least see that there is a reasonable safety review process in place before letting students try out random experiments in topic areas in which the models likely haven’t been fine tuned with safety in mind.

        • taneq 13 days ago
          Surely the ideas themselves are what should be examined for ethical suitability, rather than the meta-idea of “ask an LLM for ideas”?
      • CJefferson 13 days ago
        One obvious problem is, what if the ideas were obviously unethical?

        I would personally let this pass ethics if someone read all the generated ideas, and took personal responsibility for them passing the basic ethics rules, or got them through the ethics committee if required, exactly the same as they would their own ideas.

    • brigadier132 13 days ago
      I don't think LLMs are the right approach for this. Coordinated science would basically be a search problem where we verify different facts using experiments and use what we learn to determine what experiment to do next.
      • visarga 13 days ago
        When you can run experiments quickly it becomes feasible to use ML and evolutionary methods to do novel discoveries, like AlphaTensor's better matrix multiplication than Strassen, and AlphaZero's move 37, upturning centuries of game strategy.

        The paper "Evolution through Large Models" shows the way. Just use LLMs as genetic mutation operators. Evolutionary methods are great at search, LLMs are great at intuition but get stuck on their own, they combine well. https://arxiv.org/abs/2206.08896

        The interplay between LLMs and Evolutionary Algorithms, despite differing in objectives and methodologies, share a common pursuit of applicability in complex problems. Meanwhile, EA can provide an optimization framework for LLM's further enhancement under black box settings, empowering LLM with flexible global search capacities.

        Since chatGPT was first released hundreds of millions of people have been using it for assistance, and the model outputs influenced their actions, maybe even supported scientists to make new discoveries. The LLM text is filtered through people and ends up as real world consequences and discoveries that are reported in text, and get in the next training set closing the loop.

        Trillions of AI tokens per month do this slow feedback game. AI speeds up the circulation of useful information and ideas in human society, and AI feedback gets filtered by the contact with people and the real world.

  • barathr 13 days ago
  • UncleOxidant 13 days ago
    The ideas aren't the hard part.
    • tokai 13 days ago
      This. Any researcher should, over a lunch, be able to generate more idea than can be tackled in a life time.
      • falcor84 13 days ago
        The fact that a human expert can also do it doesn't mean the AI isn't valuable. Even if you just consider the monetary aspect, those few API calls would definitely be cheaper than buying the researcher lunch. But the big benefit is being able to generate those ideas immediately and autonomously every time there's new data.
        • passwordoops 13 days ago
          I think what they are saying is that idea generation is not a pain point and not really worth solving. Taking ideas and making them happen... that's the hard part where an artificial agent could come in much more handy
      • kordlessagain 13 days ago
        The number of the ideas has nothing to do with the quality of the ideas. Some ideas a gold, many aren’t.
    • fpgamlirfanboy 13 days ago
      Tell that to PhD advisor that took credit for all my work because they were his ideas (at least so he claimed).
      • passwordoops 13 days ago
        Unfortunately the good ones who do not steal credit are few and far between. Current incentives select for this behaviour. Not just in academia, but about everywhere.

        Go to any meeting and state the obvious fact that "any idiot can have an idea. Making it happen is the tough part" then watch how the decision makers react

  • SubiculumCode 13 days ago
    In some fields of research, the amount of literature out there is stupendous, and with little hope of a human reading, much less understanding the whole literature. Its becoming a major problem in some fields, and I think, in some ways, approaches that can combine knowledge algorithmically are needed, perhaps llms.
    • wizzwizz4 13 days ago
      Traditionally, that's what meta-analyses and published reviews of the literature have been for.
  • deegles 13 days ago
    It would be fun to pair this with an automated lab that could run experiments and feed the results into generating the next set of ideas.
  • not-chatgpt 13 days ago
    Cool idea. Never gonna work. LLMs are still generative models that spits out training data, incapable of highly abstract creative tasks like research.

    I still remember all the GPT-2 based startup idea generators that spits out pseudo-feasible startups.

    • bigyikes 13 days ago
      Ignoring the “spits out training data” bit which is at best misleading, it’s interesting that you use the word “abstract” here.

      I recently followed Karpathy’s GPT-from-scratch tutorial and was fascinated with how clearly you could see the models improving.

      With no training, the model spits out uniformly random text. With a bit of training, the model starts generating gibberish. With further training, the model starts recognizing simple character patterns, like putting a consonant after a vowel. Then it learns syllables, and then words, and then sentences. With enough training (and data and parameters, of course) you eventually yield a model like GPT-4 that can write better code than many programmers.

      It’s not always that clear cut, but you can clearly observe it moving up the chain of abstraction as the training loss decreases.

      What happens when you go even bigger than GPT-4? We have every reason to believe that the models will be able to think more abstractly.

      Your “never gonna work” comment flies in the face of exponential curve we find ourselves on.

      • ethanwillis 13 days ago
        If we keep extrapolating eventually GPT will be omniscient. I really can't think of any reason why that wouldn't be the case, given the exponential curve we find ourselves on.
        • esafak 13 days ago
          How do you know you're not on a logistic curve?

          Don't you think costs and the availability of training data might impose some constraints?

          • dragonwriter 13 days ago
            With real world phenomena that have resource constraints anywhere, a good rule of thumb is: if it looks like an exponential curve, walks like an exponential curve, and quacks like an exponential curve, it’s definitely a logistic curve
          • HeatrayEnjoyer 13 days ago
            The entire universe is training data.
            • esafak 13 days ago
              It is, but we -- humans, and computers -- are limited in our ability to learn from it. We both learn more easily from structured data, like textbooks.
        • fire_lake 13 days ago
          This has the form of a religious belief.
          • mistermann 13 days ago
            And also non-religious belief...paradoxical!
        • inference-lord 13 days ago
          I think they're being factitious?
          • ethanwillis 13 days ago
            I am. And I think it says a lot about the state of things that many people think I'm being completely serious.
    • ramraj07 13 days ago
      I have asked chat GPT to generate hypotheses on my PhD topic that I know every single piece of existing literature about and it actually threw out some very interesting ideas that do not exist out there yet (this was before they lobotomized it).
      • ta988 13 days ago
        Did you try with the API directly? I've had great results with my own prompts, much less so with the chatgpt one.
      • voxl 13 days ago
        > (this was before they lobotomized it)

        Of course, of course. Because god forbid anyone be able to reproduce your suggestion. Funnily enough I tried the same and have the exact opposite experience.

    • growthwtf 13 days ago
      I think that ship has sailed, if you believe the paper (which I do).

      LLMs are already super-human at some highly abstract creative tasks, including research.

      There are numerous examples of LLMs solving problems that couldn't be found in the training data. They can also be improved by using reasoning methods like truth tables or causal language. See Orca from Microsoft for example.

    • CuriouslyC 13 days ago
      they don't just spit out training data, they generalize from training data. They can look at an existing situation and suggest lines of experimentation or analysis that might lead to interesting results based on similar contexts in other sciences or previous research. They're undertrained on bleeding edge science so they're going to falter there but they can apply methodology just fine.
    • llm_trw 13 days ago
      They just need to be better at it than humans, which is a rather low bar when you go beyond two unrelated fields.
    • krageon 13 days ago
      When you're this confident and making blanket statements that are this unilateral, that should tell you you need to take a step back and question yourself.