Hacker Newsnew | past | comments | ask | show | jobs | submitlogin
How is AI impacting science? (michaelnotebook.com)
94 points by occamschainsaw on Dec 30, 2023 | hide | past | favorite | 114 comments


I know the article focused on the deeper/code-ier aspects of AI. However, I think that it's mundane aspects are going to be huge.

I once had a co-researcher that asked for help with his Matlab code. We were in a smell-lab together working on mice. His code was about controlling some scent valves in response to a mouse putting it's nose in a hole and breaking a laser beam. Importantly, this code was running his thesis, essentially, and he had no coding experience prior to this experiment.

I said sure, I'd love to help, but you owe me a six-pack. So we sat down one day and started going through his Matlab code. After the 18th nested 'if' statement, I had to up the payment to a case of beer.

LLMs would have helped my co-researcher out a lot. He might have actually gotten the code working at least.

Most HNers don't appreciate, I think, how difficult coding is most people. AIs are already helping bridge that gap.

Another more mundane area is in communication. Research papers are notorious for being obtuse and jargon filled, to the point where many fields are nearly speaking another language entirely. AIs can help with that as well. Not only in the writing side, but also in the reading side. Can can put a paper into them and ask for summaries, you can put multiple papers in and ask for review papers. You must be very careful, of course, but the speed increase is just amazing.


You have to understand the code the AI emits, to be able to check if it's done what you wanted. That's sometimes more challenging than writing it yourself, as it may use features you're not familiar with.

Which is not to say it can't be useful, but don't expect a cure all.


Reading code is easier than starting from scratch. Not sure how that is for a researcher though, just my SWE perspective


They're different kinds of easy...

Writing fresh code is easy because you just express your ideas. What's hard is hitting all the contingencies appropriately; it can take years of 'burn in' to get right.

Reading code is easy - especially if you don't know the language well - because someone has already assembled all the correct syntax, etc. What's hard is that the code may be convoluted from years of monkey-patches to handle all those inconvenient contingencies, to the point of obscuring the core ideas. eg, making you search through ten nested indirections to figure out how something actually executes...

On the writing side, LLM's make it much easier to have the idiomatics of the language easily on hand. (eg, I never ever remember the exact way to use 'super' in python... but ML assistance is happy to fill in the parens as soon as I start them.)


Not really, reading existing code is hard because you have to be on guard that your bias isn’t skewing your judgement about what’s going on.

What I’m getting from these comments is (as expected), most people are happy to take the easy assumption filled route rather than actually applying rigour.


My co-researcher was thrown in with the sharks on that project. He'd never even taken calculus, let alone any form of coding class. And this was Matlab, by the by, a language that not only holds your hand, but chews your food for you too. He never really groked matrices either.

We were a neuro lab, so coding was seen as something similar to a hammer: Bang away with it all you want, but at the end of the day, it's going back in a dark drawer where it belongs with the other tools.


What they're describing is dealing with a person who already doesn't understand what their code does.


This is where AI agents will shine as you give it the test cases it needs to meet and it will write the code to that specification AND verify those through writing unit tests, executing the code then verifying the results.

English as a programming language makes this all work.


You have to know how to write useful test cases to know it works, etc

Some of the comments here actually freak me out.


I'd think you'd be positively appalled by the state of code in academia then.

It's kludge after hot patch after voodoo incantation. I once worked on a computer that under no circumstances could it be put on the web. It was running Windows 98 in ~2017 and was only used for a single program to control an instrument. I know for a fact that it's still online and running. It was slow as all get out, but no one had any kind of time, let alone gumption, let alone skill, to try to bring it up a millennium.


Most academic code is written fairly naively (it's usually the code that's written by people who have some idea and try to be "smart" that's a problem), and so can be cleaned up and made maintainable pretty easily. If AI causes more "smart" code to be written, that's going to make things worse.

On old hardware/software, that's normal, and will continue to be normal, simply because the value is in the instrument, and money spent trying to get the software/drivers to work on more modern systems would be better spent on buying a new and better instrument. The only way this would change is if we force all hardware vendors to fully open-source (and ideally upstream) their code (look at the whole IoT ecosystem).


This is a good point, using an LLM means that writing even more smart code is now unlocked.

One still needs to find that balance between smart, functional / practical and maintainable. That’s the art of coding.


People will still write a vague or incorrect description, and any AI will presumably have some error rate. You can't escape verification.


I think the gap is quickly narrowing on the need to fully comprehend most computer programs for most use cases, especially if bugs and changes can all be managed through natural language by the primary stakeholders.


  I think the gap is quickly narrowing on the need to fully comprehend most 
  computer programs for most use cases, ...
What makes you assume that current "AI" "comprehends"?

I wish it could, but it appears not to.

As an example, ask it to code up something from a "less popular domain" (e.g., Clifford algebra).

You might notice that it appears to not have enough data to make good guesses. Some "AI" hype folks appear to move the goalpost then.

However, maybe the stock market and some companies benefit from all this sensationalist stuff.

I wish I had a HAL 9000 beside me, but I am sorry Dave, I cannot let you do that.


YMMV but Copilot (Bing w/ GPT4 toggle) does alright: https://gist.github.com/SmilyOrg/c0f60a41e9fcbbffdbf5454a500...

Obviously it could be completely wrong, but it certainly _seems_ to comprehend geometric algebra more than I do.


Off topic: is there a version of your site that doesn't require flash? It looks like you have some really cool simulations there.


Name one thing that HAL 9000 did that current LLMs can't do (for better or worse).

We are basically at the point where we could build the fembots from Ex Machina or David, the little boy from A.I., if we were just a little better at robotics. We'll see the teddy bear Supertoy from the latter movie under peoples' Christmas trees inside five years.


> Name one thing that HAL 9000 did that current LLMs can't do (for better or worse).

The ability to reason or understand seems to be not there in current LLMs.

See my other comment:

https://news.ycombinator.com/item?id=38819544

Simply put: "fake it 'till you make it" = probabilistic guesses given a large data set = seems to fool us people believing that an LLM "thinks" and "reasons".


I once tasked an LLM (e.g., ChatGPT 3.5)

There you go. Try it with 4. Still doesn't work? Let's revisit when 5 comes out. You will lose the argument eventually, it's only a question of when.

with deriving the geometric product (Clifford algebra) based on a set of definitions and axioms (e.g., the distributive property). Unfortunately, it failed, making numerous errors along the way.

Also maybe try asking it something that ordinary humans can be expected to accomplish, rather than complaining that nobody has invented Ramanujan-as-a-Service yet.

Also: Further, "AI" beating masters of chess also seems to be a product of "beat a human with probabilistic guesses given a large data set".

That was true for chess. Lee Se-Dol's defeat by AlphaGo was different. That wasn't supposed to happen. That was a combination of probabilistic Monte Carlo methods and... something else entirely.

That's really when the train of "AI can do X but not Y" left the tracks, IMO.


  There you go. Try it with 4. Still doesn't work? Let's revisit when 5 comes out. 
You appear to assume that current "AI" is able to "understand" and "think". What makes you so sure?

So, it "thinks" and "understands", it does not make (probabilistic) guesses?

Definition: guess

  to give an answer to a particular question when you do not have all the facts and so 
  cannot be certain if you are correct
https://dictionary.cambridge.org/dictionary/english/guess

To your other point:

  Also maybe try asking it something that ordinary humans can be expected to
  accomplish, rather than complaining that nobody has invented
  Ramanujan-as-a-Service yet.
Hmm... is it too much to ask such an "AI" for formal (i.e., well-defined) stuff? I would argue that mathematics or any formal language is perhaps more accurate than human language.

  You will lose the argument eventually, it's only a question of when.
Oh, I hope so, but I am skeptical, as I am still not convinced of an LLM being an "AI".


> You appear to assume that current "AI" is able to "understand" and "think". What makes you so sure?

"To understand" and "to think" are two very different things. Understand means more or less to encode and compress effectively from a perspective of such a system - there's quite a bit of evidence that they do that.

As for "thinking" that is impossible for LLMs as thinking is an action - and LLMs aren't agents that can plan and take action.

Actually AlphaGO and AlphaZero were agents capable of thinking - just in a extremly simplistic world which is the game of Go, Shogi or Chess. But they had a world model (which was fully known as it was for simple games) and a way to plan the action they will take by evaluating what impact will they have upon the world and how beneficial it will be for them.

Just that extending that system/agent to the real world is very hard.


You appear to assume that current "AI" is able to "understand" and "think". What makes you so sure?

A flippant way to answer the question might be to ask, "What makes you so sure you're not arguing with an AI right now?"

A less-flippant answer is simply that we would need to agree on what understanding and thinking mean before we could debate whether AI is doing it now or can potentially do it in the future. I doubt we'll come to agreeable terms there, but I'd start by suggesting a simplistic hierarchy of cognition:

0) Memory: the ability to store information. Trivial.

1) Knowledge: the ability to retrieve that information associatively.

By that I'd imply a requirement that knowledge must be retrievable by indexing at a conceptual level. Knowledge isn't "I have four bananas and the third is ripe"; that's memory. Knowledge is "Bananas are edible fruits." It's more or less unquestionable that this is satisfied by embedding memorized items in a high-dimensional vector space where conceptual similarity corresponds to distance measured by some norm or another.

2) Understanding: the ability to cast new knowledge in terms of existing knowledge. This is where the increasingly-popular compression-as-intelligence metaphors start to come into play. Anytime you use an embedding scheme to compress data, you are exhibiting understanding. When Stable Diffusion encodes information about entire works of art with one or two bits of actual information, which it can later use to synthesize new works, that's "understanding" IMO.

3) Reasoning: I'm fine with the definitions of inductive reasoning (applying specific understanding to the solution of general problems) and deductive reasoning (deriving specific outputs from general understanding.) AI models perform induction whenever they recognize and encode patterns, and they perform deduction anytime they solve a specific problem using those patterns. At this point the burden of proof falls on anyone who claims these processes aren't happening.

So that leaves "thinking", which is a topic in philosophy rather than neuroscience or compsci, and forms of reasoning that go beyond the simplistic inductive and deductive models. Hypothesis formation is interesting, for instance. That's one area where someone using a GPT 3.x model might walk away with their negative preconceptions regarding AI being fully validated, as you did, while someone using GPT 4 might walk away with a furrowed brow and a dazed look. Maybe, once the previous bases are covered, Occam's Razor is all you need to perform abductive reasoning. If not, what else?

These are all interesting points of debate that go much deeper than "LOL, it can't even construct proofs using Clifford algebra." If you're satisfied with that, then again, there's not much room for the conversation to progress.


Is this a joke ? You think an LLM can be in charge of a spaceship on an interstellar mission ?


You wouldn't use an LLM for mission management, but you could certainly use one as the human interface to the actual control system/autopilot. Which (in any event) doesn't require any sort of deep-learning implementation at all.


How would they know if the code produced by ChatGPT had a side effect that effected their research? Isn't it risky to not fully understand the behavior of your experiment?


I get what you're saying, but the reality is that most researchers aren't any more code literate than the average person. Which is to say, they know just enough to make things worse. Hopefully those errors are removed before publication, but in my experience, they are not. Especially when it comes to coding.

Also, a larger point is that this was fundamental research. A very hard thing. You're supposed to be getting it wrong most of the time. Then when you think you've got it right, you try to publish. Then you hope the reviewers actually read anything you wrote and catch all the errors [0]. But a lot of the time, errors sneak through and really only get discovered after a few years/decades. That's fine in the end, as science is a 'strongest-link' system where we really only care about the big and correct discoveries. 'Weak-link' systems are the opposite, they are governed by the small little falsehoods and errors.

[0] Very unlikely, reviews almost always only review to see if the paper is of a high enough quality for the journal to include. I want to be clear on this point: Reviewers do not re-do experiments and typically do not go back through your math, let alone the code you used. Now, each field is different and has different standards, and those vary wildly. But generally, reviewers aren't there to make sure the science is right.


It sounds like that researchers should just learn to properly code.


Yeah, we were in a neuro lab. So your patch-clamping, pipette drawing, and tetrode dropping skills were a bit more prized.

My co-researcher had never taken calculus, let alone any coding. Trying to get him up to speed with all the various skills he'd missed out on would have taken another 4 years of full time 'remedial' education. Of course, there was no time for that, let alone budget.

I totally agree with you though. And that's why I really think that AI would have helped out a lot. It's not perfect in any way, but your velocity through all the pitfalls of learning is a lot faster with the AI helping out.


If the set theorists knew how to farm, then we wouldn't need the NSF to give them money.


Given that it is relevant to their research, they should consider learning how to code.

Otherwise you'll be at the mercy of whatever autogenerated code that you happen to be using, since you literally don't have the skill to check if it's actually working.

The other alternative is to work with a programmer.


this is new technologies, and there are no scientists have salary higher than programmers


I want to echo this point.

Our 'salary' was ~$24k/year. So, just enough to stay alive and for a beer on Fridays.

If my co-researcher had the coding abilities he needed, he would have left.

And, apologies if I wasn't clear enough here, we were grad students.


Breaking down code in cells and asking chatGPT for help is quite easy and mostly foolproof, if you have enough will to learn things.


Indeed, build up the complexity from smaller, verifiable code components.


I think in this case that problem even exists in their own code that they don't fully understand.


> LLMs would have helped my co-researcher out a lot.

This describes me quite well. During the last few days I was working on getting the matrix for an integral operator, using Chebyshev polynomials. Numpy has regular and Chebyshev polynomials, but the documentation is a bit of a mess. I asked ChatGPT for help. It didn't work that well. But it got me started. It took a bit, because Numpy has actually 2 APIs for all their polynomials, and the ChatGPT code was mixing them. Once I figured that out, I decided to stick to a single API, and I finished the code snipped in about half an hour.

Without ChatGPT it would have been much more difficult. How many people out there are aware of the Chebyshev polynomials in numpy? Certainly nobody in my circle. ChatGPT was aware, but it was also hallucinating quite a bit. Still, it helped me get the job done.


Anecdotal, but I’ve talked to some physicists working in the field of quantum computing, and some of them think that it’s possible that advancements in AI will provide somewhat efficient solutions to some computational problems (namely in the NP class), and the solutions will be “good enough” for actual businesses (e.g. in logistics) and researchers (e.g. in chemistry), to a degree that it might negatively affect future funding for quantum computing research. And the pace of advancement in AI will continue to accelerate, while the pace of advancements in quantum computing is notoriously slow.


Yep, I'm very interested in "good enough" optimization techniques using ML, to massively speed up optimization problems that require a lot of computation that doesn't parallelize easily. I'm not an expert in it, but I work adjacent to it, and it seems like a promising direction to me.


They’ve always gotten lots of funding in quantum for techniques with big promises that haven’t delivered yet. I’m not even saying they’re misleading us so much as it has had basically no payoff compared to HPC, FPGA’s, and even analog computing.

The bigger concern of companies like that is if someone bankrolls dirt-cheap FPGA’s or Adapteva-style cores with open architectures targeted by a toolchain like Synflow’s or Cray’s Chapel. From there, domain-specific applications (esp optimized kernels) targeting those tools. Then, MPP-style hardware as cheap as commodity servers to scale it up and out. I’m talking engineering the thing by combining proven strategies, not doing new research.

Even $100 million put into such systems would deliver more value in more areas than $1 billion put into quantum computing. If not, we’d at least get huge speed-ups with the parallel architectures for a while and then the QC folks eventually deliver something better in some areas. I’d be happy both ways so long as one is in my building or it’s in the cloud for $2.30/hr. :)


Quantum doesn’t help for NP problems.

In the case of codebreaking, where QC may have an advantage, AI is unlikely to provide an alternative because modern AI is all based on finding probabilistic patterns and cryptography is explicitly designed to be resistant to that kind of attack.


^ Just pointing out, the parent didn't say NP-hard or -complete, so they weren't wrong (but could be misunderstood). But saying Quantum doesn't help for NP isn't right either.


Fair enough!


> Quantum doesn’t help for NP problems.

Notice that the GP is about the other way around. Solving NP-hard problems will absolutely make quantum computers useless.

And in the case of physics simulation (the actually valuable use for a quantum computer), there's no guarantee that you can't find a probabilistic solution.


Can you please elaborate?


The class of this problems where quantum computing gives an advantage is called BQP and it includes things like factoring but it is not known to include NP complete problems like the traveling salesman.


Thank you!


I have not programmed a quantum computer before. My current state of ignorance, and so im relying only on intuition here, is that it is massively parralel in the same way that the surface of water in your cup is massively parralel.

To program a serial algorithm in a quantum computer would be to miss the point. You could encode the interactions between clumps of qbits to function as nodes in a wave simulation.

There would be few enough nodes in the wave that it would be not equivalent to a true wave simulation. It would be a discretization, but now it functions as an analog asic like proxy for the real thing. If the groups of qbits are smaller than the true nodes in a wave surface your computation would be faster than the real thing, but lower precision.

In a classical computer this would have to be done with buffers, or a very narrow set of parralel deterministic programs, otherwise impossible. (examples: a subset of cellular automata rules, gravity sort)

Is any of that right or am i completely off base.


Quantum is not parallel.

Quantum is quantum. We did a few quantum gate simulations in my college back in the day.

The gist is that entangling bits (say b0, b1, b2, b3 are all entangled), has special properties. You can almost perform retroactive computing. you can send these values to quantum gates to define properties, like b0 + b1 = b2 * b3.

Later, you unentangled the values and they will 'snap' into some true state. Like b0,b1,b2,b3 == 1111, but the values 0001 or 0010 are also possible.

There is nothing 'parallel' about this. It's just a quantum operator using the properties of entanglement / disentanglement for the purposes of computation.

-------

I'd say that quantum is closer to retroactive computing than parallel. Operations seemingly go backwards in time thanks to entanglement.

That doesn't mean quantum is useful in all algorithms, but it seemingly has applications in cryptography and 3SAT / NP Complete and... Lol simulations of quantum effects.

--------

Ultimately, Quantum is only useful if you get enough bits at a time entangled. For now, it's faster for standard computers to brute-force all 2^32 bit combinations rather than using lasers+diamonds to entangle 32-bits and perform the operations.

Quantum needs more entanglement and at lower costs to become competitive. But maybe it happens. At very least, quantum computers do exist but who knows how long it will be before they are competitive.


One of the challenges of quantum computing is getting the useful results back into "classical" mode, where you can use them. Theoretically that's one of the hurdles of getting better theoretical performance than a classical computer. What you describe is more akin to how a GPU works, but in this case, you can't "just get" the result from the GPU.


Well that sounds like good news, are there any practical applications for quantum computing other than breaking commonly used encryption that would push the modern world even further towards a complete surveillance state?


The #1 proposed use of quantum computers is... simulating quantum effects.

Modern quantum simulators running on supercomputers is an exponential (O(k^n)) kind of operation. But a quantum computer can simulate any quantum effect in just one step, O(1) time. Because ya know... a quantum computer is just a computer that has isolated quantum effects and allows a programmer to control them easily.


Isn't O(1) an oversimplification? In comparison wouldn't electrical computers then also be able to simulate all electrical effects in O(1) Time?


> In comparison wouldn't electrical computers then also be able to simulate all electrical effects in O(1) Time?

Yes and OpAmps / analog electrical circuits were way faster than digital computers for decades because of this.

When you know that the current of a diode is related to the exponent of the diode's voltage, you can do silly things like calculate logarithms using OpAmps and diodes in a mere nanosecond or so.

Most 1980s synths used OpAmps as the basis of the math / calculations for signal processing, because digital computers just weren't fast enough to compete yet. Those circuits still work today.

Alas, digital computers are too cheap, accurate, and fast these days so nearly everything is digital now.

----------

I guess your point though is that maybe a quantum computer (or OpAmp / electrical computer) can only physically simulate the effects that the hardware contains.

Ex: an electrical computer cannot simulate BJTs unless it has a BJT somewhere. So it would still come down to the computers physical load out.

Similarly, a Quantum simulation on a quantum simulation computer would only have some subset of primitives implemented.


I'm not in the field but in my understanding it's going to make it easier to do parallel competitions and also will allow us to find the inputs that can lead to a particular output in some types of equations. Factoring numbers is only one of the "practical" equations.

Materials science and chemistry (e.g. predicting how drugs may behave in our bodies) will benefit, or other computationally heavy "try out many permutations and see which ones fall though" tasks may become a bit easier. Climate modelling, supply chain logistics, financial modelling, and of course AI.


I think AI and quantum computing are really the same thing, the difference being the same as girltalk in contrast with boytalk.

what I'm saying is that there's no quantum computing; but I actually mean to say I simply do not understand whatever 'quantum' means in the context of computing given my own opinions of what computing really is (I have a philosophical opinion).

in my view (as distorted and twisted as it is), what quantum computing devices really do is sensing or measuring. computing is classical in nature and no amount of hand waving will change this.

whatever happens to logic in a qubit? if both A and B are anything between 0 and 1, what's the negation of such a thing?


>whatever happens to logic in a qubit? if both A and B are anything between 0 and 1, what's the negation of such a thing?

The problem is the pop sci explanation where qbits are everything between 0 and 1 at the same time is super wrong and misleading.

Your qbits are in states that have "amplitudes". You do things that change the amplitudes, and that happens to do computation. From the amplitudes of your qbits, you can figure out what the probabilities of observing different outcomes is.

It's not that the qbit is anything between 0 and 1. It's in a precise state, that results in percentage probabilities for different outcomes.

The negation of true 80% of the time is false 80% of the time.


To be fair, in some quantum computers, it does work. For example if you count a beam of photons following the same path with (almost) all possible polarization combinations as a qubit, that is indeed closer to "everything between", however it's got limited utility indeed when compared to the broader view.


Apologies in advance—I don't know how to say this in a loving way, but I also don’t mean to pass judgment—couldn’t this just be a skill issue? Why not look into how quantum computing works and resolve that once and for all?

https://quantum.country/ is a pretty approachable starting point if you’re curious. The math is no more intimidating than any other skilled discipline.


you can just say it, it's not worse than getting downvoted for having off colored opinions

"aren't you just too stupid to understand?"

maybe I am... maybe I just think differently with another brand of depth

what I wonder now, given another comment about amplitudes, is what happens to combinatorics? (and alphabets, and languages as sequences of strings from finite alphabets) but it's just simpler to dismiss me as foolish idiot.


People responded with civility which, in retrospect, you may not have deserved.


Im just as ignorant as you about it. The quantum part to me doesnt seem to matter. I don't see why its any fundamentally different from making a computer out of any other substrate. Light, sand, chemicals, at the end of the day you compute a function by following a set of steps which may or may not have constraints of seriality.

I think von neumann cant handle purely parralel algorithms efficiently. Even gpus. I think thats whats where the "all qbits interacting simultaniously for n steps after being put in initial conditions" is about. Its purely parralel.

It could be the case that information cant be extracted before it is computed, and that for some algorithms the peak efficiency serial version is equal to the best efficiency purely parralel version. It could also be the case thats not true and that classes of quantum algorithms could be better.

At the end if the day the way i see it substrate is just an engineered means of applied ops per time.


The quantum part does matter in the sense that certain kinds of "maths" are not directly accessible in a classical computer. You can sort of simulate Shor's algorithm on a classical computer but you cannot get the same low complexity. Of course, whether the quantum computer uses substrate A or B doesn't matter (other than practically).

The only way I ever found an access to understanding quantum computing is by doing the math, as other pop-sci explanations don't really reflect what is happening (at least for me).


Regarding the negation, you can consider it in a few different ways, depending on your use case.

If your qubit is "all values", the negation is the lack of any value, e.g. "a measurement that results to 0".

However, in most cases, you will have a segment within the possibilities. E.g. one qubit can hold "vertically polarised light at frequencies between X and Y at phase P", then the negation could be one of "horizontally polarised light at that frequency and phase", or, more common, same polarization and frequency but the phase is 180 degrees off P. That way, if you add them together, you get the cancellation. Another negation is "all other possible combinations".

However I must note that when you call a qubit "all possible combinations" that's not a typical qubit value.

Think of it like this, "binary" is "0 or 1", but a "binary value" is only one of either 0 or 1. A qubit value is, depending on your interpretation, photons of some polarisation, frequency and phase, or electrons of a particular spin, or some other combination of those. Some of those can have multiple combinations, for example a qubit value could be the combination of a bunch of photons at different values. Or, an electron with an "unknown spin". Or an electron with an unknown spin that's the complement of another electron at another unknown spin, but when you measure either you will reveal the other, and so on. So, only some qubit values will have "all possible combinations" i.e. "when observed, it could literally be anything".

The internal of "what was the value before you observe it" is, well, controversial, and to most people, almost irrelevant, and even, "not good to consider".


I can tell you one impact -- there are huge amounts of ChatGPT helped, or entirely written, junk papers appearing at conferences, which is causing serious problems getting reviewers. Some large events have introduced a pre-review phase to filter out junk, but it's getting harder, as LLMs are good at producing plausable looking nonsense.

Honestly, I'm not sure of the long term solution here. Conferences and journals may have to introduce a system where you need an existing member of the community to "vouch" for your work to be allowed to submit (they don't have to say it should be published, just worth reviewing).


LLM-written content has a certain gestalt to it that’s pretty easy to sniff out once you’ve seen it a few times, just look at Quora and LinkedIn these days.

While I think it’s probably fair for people with English as a second language to use LLMs as a writing assistant (for example), there definitely needs to be some kind of author disclosure statement at the very least. This could be similar to how more papers are now including contribution statements when there are many coauthors.


That gestalt can easily be smoothed out. You simply need to prompt LLM to do so in your query. Tell the LLM to depart from its default style. Say make your response more brief, well written and add a witty conclusion at the end

Then you say rewrite the final paragraph to be slightly shorter and incorporate a certain example. Yada yada.

Likely you can even do this "write your response in a style that is not typical for an LLM"


It’s easier to illustrate with prose: ask it to write in the style of Terry Pratchett or Hunter S Thompson and the style changes drastically. You can even ask it to rewrite your scientific paper in the style of Shakespeare: https://chat.openai.com/share/498804da-6d59-4a0c-91d1-da4cea...


Are there any molecular biologists here? I'm curious: what happens to society if/when we figure out protein folding? Hypothetically, how does the world change if we had a 100% accurate way to model quaternary structure from primary structure?


My brother worked in protein folding (although his statement about specifically AlphaFold was years after he left that field) but I showed him the AlphaFold results from like 5 years ago and his reaction was "oh ... Yeah they solved protein folding"

So at least according to him we've lived in that world for the last 5 years.

As a person working with / tangentially to people in the same field I would say that it's made things faster and more scalable, but protein structure isn't the be all end all of things. Researchers use AlphaFold a lot for filtering potential candidates, but that is only one step in a lot of steps. A SNP mutation -> protein structure change -> functional change is already difficult without then considering that the vast majority of mutations that create function change in humans are not in exons, so something like AlphaFold (in the form I'm familiar with) would be useless for those as well.

Eventually though an AI system that can go mutation->function change is entirely possible, although it is much much further in the future. In that case though I think you'll be quite close to a future where combined with things like CRISPR therapeutic treatment for all heritable disease would be possible.


AlphaFold did not solve protein folding. You can tell this is the case because it doesn’t correctly model structures with missense variants.

AlphaFold is a huge advance and people are extending it to try to tackle this (eg AlphaMissense) but to say it solved protein folding is hyperbolic.


>"AlphaFold can do in TEN MINUTES what a PhD dissertation required, fifteen years ago."

Quoting a PhD immunologist from Texas, who uses AlphaFold daily.


Agreed - and this is not in tension with my comment whatsoever.


Is that not the same as saying maths is solved thanks to spreadsheets/excel? All of science and technology is to some extent automation and increased speed of what used to be slow and laborious.


My sincerest response is that we are now post-moore's-law[1]; instead of "transistor density doubling every 18 months," I have (over the past year of Perplexity.AI) more of the belief that "global intelligence will double every 18 months."

Just my ¢¢ half-sanity.

[1] Isn't a single hydrogen diatom just 1nm wide? How much smaller can transistor gates be safely assembled? Didn't we physically stop getting physically smaller around 14nm?


I heard a lot of different opinion about AlphaFolding, how it's not revolutionary or not revolutionary, or something like that.


It is revolutionary in what it does. There's no question about it. The problem is that a lot of people think it does something different or more practical, like creating new drugs. That is one of the ultimate goals, but it's way down the road.


Counting down to dekhn appearance in 3…


Uh ok. Now I feel obliged to say that AlphaFold is the obvious outcome you'd get from having a company dedicated to winning games deciding to win at CASP.

To me, "solving" structure prediction (explicitly acknowledging there are areas where AF doesn't make accurate predictions), is a clear and satisfying win, although it still doesn't answer some of the fundamental physics questions around folding.

I am glad to see the existence proof but want to see more outcomes; in particular, I'd like to see a lab-in-the-loop that actually produces something of high value (higher than the cost of building the lab-in-the-loop).


A lot of things become simpler, but hard problems remain. In the cell, a structure of a protein is the result of its sequence and all the other environmental conditions (cofactors, ligands, binding partners, lipids,…), and furthermore, the structure is a dynamical construct, so it can change over time. So the main advance for society is when we can accurately predict how to jam or unblock a machine in a cell without influencing all the other machines in a cell. The essential pieces of this puzzle are the possible shapes of the pieces that make up the machines and AF2 gives you those. But the puzzle is huge. And very important.


Even if we had a perfect library of useful proteins and their utilities, the composition of these protiens, and the emergent behaviour within "machines" made by them, would be complex enough to be handfuls of classes of new engineering altogethor.

I think its like saying "we know how to make bricks and iron and bolts now, shouldnt it be easy to make a full scale functioning replica of the greater tokyo metropolitan area?"

Even if you came up with a spec for a complex tissue or just a fluid that functions as a standalone chemical factory, you would need to fabricate it. However many specific protiens from your library, in specific ratios, mixed and maybe even... positioned.

In short protein folding is just the first fundamental step towards this print any biological design out of proteins world that I guessed you are alluding to.


It would not have a huge practical impact in the short term. There are many steps from figuring out protein folding to figuring out new, effective and safe drugs. Here's a good Nature article discussing just that: https://archive.is/4QNKy


We could do computer-only search for enzymes that catalyze desired reactions in enzyme-friendly reaction environments. Assuming that the way that makes the above 100% accurate also yields us the tooling to simulate the candidates in operation accurate enough to let the search/optimizer learn from the simulation feedback.


The static shape of a protein doesn't automatically give you a prediction of its functional properties. There's a hell of a lot more biophysics going on that we have no predictive models for that are needed to understand catalysis, allostery, assembly, etc etc etc. We don't even have good comprehensive data for any of that (compared to sequences or structure) to model with.

Fold prediction is an incredibly useful tool for scientists and genetic engineers to help design new proteins, but it doesn't magically solve molecular or cell biology. Designing new functions and mechanisms is still going to involve a huge amount of labor and brute-force experimentation.


okay so the prevailing sentiment seems to be: a) We kind of already have b) it's a big deal but not that big of a deal because knowing the precise protein structure does not become useful until you know how that structure interacts

So my takeaway is that we just need sufficiently advanced quantum computers that can do molecule-level simulation on a large cell, THEN we'll be there.


there was a talk about this a few days ago https://media.ccc.de/v/37c3-12061-alphafold_how_machine_lear...


Although the article focused primarily on AlphaFold, many other ML approaches are making impactful contributions in the general scientific field. One example is the diffusion model and its use of stochastic differential equations (SDEs).

Microsoft has an initiative called AI4Sciencie (https://www.microsoft.com/en-us/research/lab/microsoft-resea...) which published a fair amount of SDE/diffusion-based method to solve scientific problems


It's from May, pity, it would have been really interesting if the author had discussed DeepMind's recent FunSearch paper as well: https://www.nature.com/articles/s41586-023-06924-6

But it goes to show just how fast we are currently progressing.


Someone needs to host an event where those people, others doing genetic programming, and those doing executable synthesis are all in the same place. Let them bounce ideas off each other. All their brainstorming is recorded to be published into the public domain.

We might see some interesting combos.

EDIT: Maybe use a sample from the Humies as benchmarks for the techniques, including new LLM’s. Let people try every approach in parallel mixing the best of each.

https://human-competitive.org/awards


Anyone here who knows a good project that attempts to solve large sparse linear systems using ML, preferably in Python?


Why would ML be any good at that?


I’m not a scientist but I enjoy reading the literary genre they write (the headlines). Most of the time, the articles themselves are hard or impossible to understand.

ChatGPT can lend a helping hand


> how to predict the 3-dimensional structure of a protein from the sequence of amino acids making up that protein

Keep in mind that proteins and deducing the structure of them (which in turn would help deduce their function) may not be a good thing. Proteins can also be poisons, prions, etc.

But beyond that, we should not have such easy to access knowledge. It's just too much power for our current greedy, short-term thinking.


Chipping flints into sharpened edges may not be a good thing. Sharpened flints can also be stone axes, spear heads, etc.

But beyond that, we should not have such easy to access cutting technology. It’s just too much power for our current greedy, short-term thinking.


Glib analogy, but there's got to be a limit to everything. Beyond resorting to superficial analogies, we should give serious thought into what level of technology is simply too much.

Moreover, I actually agree with you in some ways, even though you meant to be sarcastic. Modern cutting technology, which certainly should include deep-well oil drilling, IS too much.

In fact, all technology has a limit beyond which we should not go. We just have to apply deep thinking to figure out what it is, or at least do our best to try.

You


You’re more likely to get cut with a dull knife. It can cause it to slip.

Technology is fine. It’s this artificial cookie-cutter society we’re forcing on everyone. We have a mental health and education problem.

The tech isn’t the issue, it’s the people who feel the need to use it for negative purposes.

Limiting tech is just a bandaid.


People will always use technology for negative purposes as long as it can help them get ahead in the short term and they aren't held accountable for the long-term consequences.

Technology is not fine. Not all of it is bad, but to say all of it is fine makes no sense. Yes, we have a mental and health education problem, but it's reasonable to have rules for all societies. Limiting technology that makes getting ahead too easy makes sense because even in the most utopian and good societies, if gaining something is TOO EASY, then it WILL be taken.

It's much too simplistic to frame technology as fine. If all technology is fine, perhaps every corner store should sell automatic weapons? Can you imagine a society, ANY society, in which this makes sense?


Technology is never a problem. Automatic weapons can also be used in defense. Someone with a knife wouldn’t be able to confront a mass killer with a rifle.

You are talking about mental problem. Do you know what have contributed to mental problem ever since the existence of human being? Isolation and loneliness, which technologies such as telephone, email and yes, social network can help with.

Name any technology that should be banned, and I can give you a reason why it should be not.


Nowhere did OP suggest limiting tech, they suggested we are not ready to handle the powers we are gaining responsibily.

Your point seems in agreement with this to a degree.

More importantly, you claim it to be a mental health/education issue; yet you can't show me a period of human history without wars and power struggles.


Cars can be used to ram people. But most people simply dont want to do that.

What is different about cars and new tech is that new tech comes with the ability to track and monitor all users. So historically we didnt mind if one in ten million people wanted to ram people with their truck on purpose. But now the power to monitor everyone all the time to prevent these "one in ten million" crimes exists.

Its up to you to decide if the trade off is worth it.


I want to become a von neumann probe, fill the universe with my offspring and go to war.

Until tech can do that it hasnt gone far enough.


I was being sarcastic, I don’t think I was necessarily being glib.

I suppose the deeper point I was trying to make was that “this time is different/special” is shallow thinking, as is believing that the real issue is with our “current greedy short term thinking”.

The real issue is the human condition, which remains largely as it has throughout our history. We need to choose a way to live knowing we will die, and then live that way.


This time is different....highest CO2 level in 800,000 years, highest human population ever. Seems like a bit of uniqueness there.


How do you feel about antibiotics? How do you feel about medical advances that reduced infant mortality from more than 20% in the 19th century to less than 1%? How about the elimination of smallpox or how fertilizer and other agricultural technologies have eliminated most famines. Now how about possible cures for cancer or alzheimer’s? You need a much more nuanced answer than “we should not have such easy access to knowledge.”


> How do you feel about antibiotics? How do you feel about medical advances that reduced infant mortality from more than 20% in the 19th century to less than 1%? How about the elimination of smallpox or how fertilizer and other agricultural technologies have eliminated most famines. Now how about possible cures for cancer or alzheimer’s?

Like everything else, it's only good up to a point!

- I believe in basic medical care, but medical technology that permits immortality is bad.

- Likewise, is reducing mortality always good? Now we have extreme overpopulation. In short, reducing OUR mortality has also INCREASED the mortality of non-human animals, which I believe have just as much right to live as us.

- Fertilizer and agriculture? Most industrial fertilizer poisons the land and agriculture has eliminated millions of acres of wild forest and destroyed a lot of ecosystem. Agriculture is also responsible for the huge population growth that is causing global warming.

- Cure for cancer? Maybe good, maybe bad IF that cure results in significantly extended lifetimes due to how it works.

I do have MANY more nuanced answers. That is why I believe in looking at each of your points AND others individually. My answer was specifically directed at AI, which I believe to be universally bad.

Most of the technologies you listed have significant negative consequences, and it's not clear to me that those are outweighed by the positives, especially if our modern society results in the destruction of the phytoplankton of the ocean producing half our oxygen.

I suspect in the long term, it would be better to have more mortality, smaller populations, and less technology than our current world. It's not a popular view because people are emotional about such things, but it's hard to imagine how a world with 8+billion and mass eco-destruction to be better IN THE LONG TERM.


Yes, strangely, people get emotional about being asked to let their kids die of scarlet fever and other bacterial infections. I’m a birder too, but some new luddism isn’t going to save the environment. If we stop where we’re at, we’re stuck burning fossil fuels, stuck with slash-and-burning the Amazon. If you want save to animals, you need better messaging than “maybe cancer isn’t bad.” Look around the world, places with great longevity have lower population growth, not higher. Places with better economies and more technology are also greener. When you say things like “AI is universally bad” you’re making an ethical judgement about a means rather than an end and ruling out many possible good ends. You’ve read Peter Singer, now go read Hans Rosling.


My message isn't that "cancer isn't bad". It's that medical research into cancer may lead to other outcomes which are definitely not good.

Yes, I am making an ethical judgment about the means of AI, because after some analysis I do believe that the negative effects will be far worse than the positive. True, there are good applications of it. I never denied that.

Of course, my line of reasoning is far more vast than what I wrote here. I'm just arguing a specific point here, not presenting a thesis. I do think there is more to saving animals than arguing against human growth. Of course, people need incentives.


I wholeheartedly disagree with your assertions on mortality. Global population growth is slowing and is set to peak in the 2060s and will decline afterwards. We're more at risk of not having enough hands to take care of an aging population. An educated populace has fewer children and many countries are already feeling the strain of being below the replacement rate.


You says you have nuanced view yet you make universal statement don't you see a contradiction?

Take your stand on AI, which is more diverse than the LLMs and diffusion models that are currently making the news. Would it be bad to replace peoples by computer vision and robotics in a waste sorting center, enabling more recycling and less microplastics in the environment due to a better sorting of the waste stream ?

I see that as a good thing enabled by AI.


AI for me is an exception to my nuance, simply because it's too powerful and has too many horrible consequences. Yes, your application could be considered good. But what happens if there are people at the waste recycling plant who can't get other jobs? It's not a nice job but we have to first figure out what those people can do.


How do you feel about extremes?

Antibiotics have saved billions of peoples lives from an early death. But use of antibiotics in farm animals is putting us at risk again via resistant superbugs that are sending us scrambling for new drugs to combat them.

Or, to take the analogy to 11, should each individual on the planet have access to a button that if pushed would cause the planet to explode? I seems reasonable the answer would be no, but now how do you set up a system of rules limiting who has access to the technology to make that button?


1. These are not even comparable dangers. Reduction in infant mortalty by 98% versus what.. some hormone issues sometimes in some people??

2. The laws of physics have so far prevented us from making technology like a tiny portable button that blows up a planet. Id assume a universe with weak energy constraints like that would have yielded much larger scale organisms, or perhaps none at all.


Uh you don't comprehend 1 then.

1 is you get a scratch then die because your body rots and you can do nothing about it. May want to spend some time learning about antibiotics resistance.


Funnily enough, I wrote a sci-fi story a long time ago where there was a machine like the replicator in Star Trek that could make anything, but it was easy to turn it into a gigaton bomb.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: