Does Neuroscience Challenge the Existence of Free Will?

  • Thread starter Thread starter Ken G
  • Start date Start date
  • Tags Tags
    Free will Neural
Click For Summary
The discussion centers on the implications of Benjamin Libet's research, which suggests that decisions occur in the brain before conscious awareness, raising questions about free will and determinism. Participants explore whether this indicates a conflict between determinism and free will, proposing that neurological processes may be deterministic while free will could exist in a non-physical realm. The conversation critiques the reductionist view that equates physical processes with determinism, arguing instead for a more nuanced understanding that includes complexity and chaos theory. The idea that conscious and unconscious processes are distinct is emphasized, with a call for a deeper exploration of how these processes interact in decision-making. The limitations of current neuroscience in fully understanding consciousness and free will are acknowledged, suggesting that a systems approach may be more effective than reductionist models. Overall, the debate highlights the complexity of free will, consciousness, and the deterministic nature of physical processes, advocating for a more integrated perspective that considers both neurological and philosophical dimensions.
  • #181
Q_Goest said:
I don’t see how physical states preclude change. Physical states exist in both space and time. In http://en.wikipedia.org/wiki/Phase_space#Thermodynamics_and_statistical_mechanics" for example, if a system consists of N particles, then a point in the 6N-dimensional phase space describes the dynamical state of every particle in that system, as each particle is associated with three position variables and three momentum variables. In this sense, a point in phase space is said to be a microstate* of the system. Physical states require both dimensional and temporal information to describe them, so I don’t know why one would claim that physical states don’t have a “temporal extent”. I don’t know what that means.

So you would disagree with the Wiki definition of states in classical physics as " a complete description of a system in terms of parameters such as positions and momentums at a particular moment in time"?
http://en.wikipedia.org/wiki/State_(physics )

All you are saying in pointing out that 6 dimensions can capture "all" the dynamics of a particle is that this is the way dynamics can be modeled in reductionist terms. You can freeze the global aspects (the ones that would be expressed in time) and describe a system in terms of local measurements.

Yet we know from QM that the position and momentum cannot be pinned down with this arbitrary precision - this seems a very strong ontological truth, no? And we know from chaos modelling that a failure to be able to determine initial conditions means that we cannot actually construct the future of a collective system from a 6N description. And we know from thermodynamics that we cannot predict the global attractor that will emerge in such a 6N phase space even if we did have exactly measured initial conditions - the shape of the attractor can only emerge as a product of a simulation. Etc, etc.

So we know for many reasons that a state based description of reality is a reduced and partial model good for only a limited domain of modelling. To then use it as the unexamined basis of philosophical argument is a huge mistake. Even if as you argue, it is a "mainstream" mistake.

You probably still don't understand why states are the local/spatial description and exclude global temporal development. But "at a particular moment in time" seems a pretty clear statement to me. It is the synchronic rather than diachronic view. Surely you are familiar with the difference?

I’d like to honestly understand why there are people that feel the nonlinear approach (dynamics approach, systems approach, etc…) such as Alwyn Scott, Even Thompson, many others, is so appealing, but these are not mainstream ideas. Would you not agree? The mainstream ideas surrounding how consciousness emerges regards computationalism which doesn’t seem to fit with this other approach. From where I sit, weak emergence and separability of classical systems are well founded, mainstream ideas that are far from being overturned. They are used daily by neuroscientists that take neurons out and put them in Petri dishes and subject them to controlled experiments as if they were still in vivo. Then they compare this reductionist experiment with the brain and with computer models which are clearly only weakly emergent. So what is it that is really being promoted by this systems approach?

Perhaps your area of expertise is computer science and so yes, this would not be the mainstream view in your world. But I am not sure that you can speak for neuroscience here. In fact I know you can't.

I have repeatedly challenged you on actual neuroscience modelling of the brain, trying to direct your attention to its mainstream thinking - the effects on selective attention on neural receptive fields has been one of the hottest areas of research for the last 20 years. But you keep ducking that challenge and keep trying to find isolated neuron studies that look comfortably reductionist to you.

You just don't get the irony. Within neuroscience, that was the big revolution of the past 20 years. To study the brain, and even neurons and synapses, in an ecologically valid way. Even the NCC hunt of consciousness studies and the brain imaging "revolution" was based on this.

People said we have been studying the brain by isolating the components. And it has not really told us what we want to know. We stuck electrodes into the brains of cats and rats. But they were anaethetised, not even conscious. And it was single electrodes, not electrode arrays. But now (around 20 years ago) we have better equipment. We can record from awake animals doing actual cognitive tasks and sample activity from an array of regions. Even better, we can stick humans in a scanner and record the systems level interactions.

Yet you say the mainstream for neuroscience is people checking the electrical reponses of disected neurons in petri dishes, or IBM simulations (gee, you don't think IBM is just about self-promotion of its supercomputers here?).

I used to write for Lancet Neurology, so I think I have a better idea of what is mainstream in neuroscience.

Again, remember that my claim here is not that reductionism (the computer science view of life) is wrong. Just that it is the subset of the systems view you arrive at when you freeze out the issue of global constraints. It is the adibiatic view. Where the larger Ps model also has to be able to deal with the non-adibiatic story - where global constraints actually develop, evolve, change, in time.
 
Last edited by a moderator:
Physics news on Phys.org
  • #182
Q_Goest said:
Oh... and by the way. I (think I) agree with both of you regarding the fundamental issue that the knowledge paradox seems to flounder on. The problem starts out by using as an axiom that phenomenal states are not physically describable. They are not physical states. Once you define qualia that way, you may as well go whole hog and admit that the causal closure of the physical is false. These two axioms are at odds which is why there is a paradox.

Or instead, you could recognise that you had made a wrong move in assuming P and M to be ontologically separate (rather than epistemically separable - big difference).

The axiom that P and M are separate can be false, and yet the axiom that P is close is true.

And you again seem to be missing the point that axioms are epistemological assertions of modelling convenience rather than statements of ontological truth. They are "truths" that seem reasonably in the grounds of generalised experience rather than truths that are known to be true due to some magical kind of direct revelation.
 
  • #183
Q_Goest said:
Perhaps you could start a thread that highlighted those views and provide references so I can dig deeper.
You're asking apeiron to provide references to support his claims... dude you like to live with risk! :biggrin:

Q_Goest said:
The point being that laymen terms are not applicable to what is the ‘efficient’ cause.
I don't think I get your point here. What is the difference you see between layman causality and efficient causality?

Q_Goest said:
One can say the gun caused the spaceship but as weak emergence would have it, the ability to emit a spaceship is dependant on the ability of individual cells in the game of life to change state from white to black and back again which is a function of the rules to the game
Or one can say that the gun is an algorithm, which indeed it is (and a simple one: it's just a periodic oscillator), thus the behavior does not need to tied with the particular rule of CGL: any system mathematically equivalent to these guns is in deep the same system. So if we afford to say that the gun cause the spaceship, then the causality is in fact the identity to a Turing machine. That said, a non trivial consequence is that free will is a set of algorithm, define as those who can behave as something we will recognize has having free will. (...) It's late I'm becoming unclear I guess. See you. :smile:
 
Last edited:
  • #184
  • #185
Hi apeiron. I found that last post to be totally understandable. Not like the previous post at all. I really wonder how you manage to switch the flowery talk on and off like that. No offense intended.

apeiron said:
Yet we know from QM that the position and momentum cannot be pinned down with this arbitrary precision - this seems a very strong ontological truth, no?
But pinning down particles isn't important to classical mechanics. Sure, the real world isn't classical, but that's not the point. The conventional view is that quantum mechanics isn't a factor in how the brain works because there are sufficient statistical aggregates of particles that individual particles don't matter. They're simply averaged together. Is this "systems view" dependent on individual particles? Clearly Alwyn Scott for example, makes the point as do many others that classical mechanics has this 'more than the sum' feature already intrinsic to it and I’d think Scott’s views probably mirror your ideas fairly closely.

And we know from chaos modelling that a failure to be able to determine initial conditions means that we cannot actually construct the future of a collective system from a 6N description.
But is that because of not having the initial conditions of the individual particles? Or not having the initial conditions of the classically defined states? Density, internal energy, entropy, etc... of the system is obviously independent of specific individual particle states, but not of the aggregate. So not knowing individual particle states will lead to indeterminate future states, and yes, the classical model is a model. But one has to show that there is a "meaningful difference" between having initial particle conditions and having initial classical conditions. I see no meaningful difference. Sure the classical approach isn't exact, but it is exact to the degree you have the initial conditions of the classical states and that's what's important unless the quantum mechanical states are being roped into causing different classical states by downward causation which isn't possible.

And we know from thermodynamics that we cannot predict the global attractor that will emerge in such a 6N phase space even if we did have exactly measured initial conditions - the shape of the attractor can only emerge as a product of a simulation. Etc, etc.
Can you provide an example of a global attractor? One that regards classical mechanics and can't be predicted by weak emergence? This is fundamentally where we disagree. I would say Benard cells are a perfect example of a weakly emergent structure, and I’d contest that’s a mainstream idea not just my own. Regarding mainstream, perhaps that feels like a knock so instead I’ll say “cutting edge” or something. Anyway, as Davies points out.
Thus we are told that in Benard instability, … the molecules organize themselves into an elaborate and orderly pattern of flow, which may extend over macroscopic dimensions, even though individual molecules merely push and pull on their near neighbors. This carries the hint that there is a sort of choreographer, an emergent demond, marshalling the molecules into a coherent, cooperative dance, the better to fulfil the global project of convective flow. Naturally this is absurd. The onset of convection certainly represents novel emergent behavior, but the normal inter-molecular forces are not in competition with, or over-ridden by, novel global forces. The global system ‘harnesses’ the local forces, but at no stage is there a need for an extra type of force to act on an individual molecule to make it comply with a ‘convective master plan’.
Also from Davies
Strong emergence cannot succeed in systems that are causally closed at the microscopic level, because there is no room for additional principals to operate that are not already implicit in the lower-level rules.
However Davies does allow for some kind of emergence at the border between classical and quantum mechanics, which is where separability breaks down also.

You probably still don't understand why states are the local/spatial description and exclude global temporal development. But "at a particular moment in time" seems a pretty clear statement to me. It is the synchronic rather than diachronic view. Surely you are familiar with the difference?
No, I'm not. Feel free... Regardless, why should it be controversial to suggest that there exists a physical reality at a particular moment in time? If the argument drops into quantum mechanics, there’s no point in arguing. At that point, we have to suggest that neurons interact due to some quantum mechanical interaction, which isn’t worth arguing about.

Perhaps your area of expertise is computer science and so yes, this would not be the mainstream view in your world. But I am not sure that you can speak for neuroscience here. In fact I know you can't.
I guess I disagree.

I have repeatedly challenged you on actual neuroscience modelling of the brain, trying to direct your attention to its mainstream thinking - the effects on selective attention on neural receptive fields has been one of the hottest areas of research for the last 20 years. But you keep ducking that challenge and keep trying to find isolated neuron studies that look comfortably reductionist to you.
I think you misunderstand. My daughter has selective attention. I have no doubt they influence neural receptive fields! lol But that’s like saying the spaceship is caused by the gun that Lievo keeps talking about. Unless you can clearly define selective attention and neural receptive fields, it won’t help.

You just don't get the irony. Within neuroscience, that was the big revolution of the past 20 years. To study the brain, and even neurons and synapses, in an ecologically valid way. Even the NCC hunt of consciousness studies and the brain imaging "revolution" was based on this.

People said we have been studying the brain by isolating the components. And it has not really told us what we want to know. We stuck electrodes into the brains of cats and rats. But they were anaethetised, not even conscious. And it was single electrodes, not electrode arrays. But now (around 20 years ago) we have better equipment. We can record from awake animals doing actual cognitive tasks and sample activity from an array of regions. Even better, we can stick humans in a scanner and record the systems level interactions.

Yet you say the mainstream for neuroscience is people checking the electrical reponses of disected neurons in petri dishes, or IBM simulations (gee, you don't think IBM is just about self-promotion of its supercomputers here?).
I don’t disagree with anything you said here except that I’m sure IBM isn’t somehow influencing neuroscience with profits from their computers. Sure, we’ve made progress in understanding how brains work, just as you say. That’s all the kind of work that’s necessary for the reductionist approach. I suspect you intend to mean that all this experimentation is unique somehow to a systems approach, but I don’t see why.
 
  • #186
apeiron said:
I supplied Q Goest with the references he requested long ago. Perhaps he did not read them? That often happens here doesn't it :wink:?

https://www.physicsforums.com/showpost.php?p=2501587&postcount=7
Not sure why you brought up that link.
You quoted Bedau who only accepts weak emergence as I've pointed out.
You quoted Emmeche who rejects strong downward causation and his medium downward causation is frighteningly like weak downward causation. It either has to drop into one or the other category, which isn't clear.
You quoted yourself on Physicsforums
You quoted Google
And this one: http://www.calresco.org/
and a handful of other web sites. I don't want to read web sites though except perhaps the Stanford Dictionary of Philosophy or maybe Wikipedia.

I guess we should just disagree and leave it at that.
 
Last edited by a moderator:
  • #187
Lievo said:
I don't think I get your point here. Why do you agree to attribute gun with freedom but not valves,
I'm not attributing the gun with freedom (free will). I'm saying it has just as much as the valve, which is none.

Lievo said:
and what is the difference you see between layman causality and efficient causality?
Layman causality is not efficient causality.
 
  • #188
apeiron said:
I supplied Q Goest with the references he requested long ago. Perhaps he did not read them? That often happens here doesn't it :wink:?
Perhaps he had a look and the references were not supporting your claims unless a lot of creativity was involved. That sometime happens, doesn't it :wink:?
 
Last edited:
  • #189
Q_Goest said:
Layman causality is not efficient causality.
That what you said and I perfectly understood that it was what you said. My question again: please explain why you think there is a difference.
 
Last edited:
  • #190
Q_Goest said:
I’d like to honestly understand why there are people that feel the nonlinear approach (dynamics approach, systems approach, etc…) such as Alwyn Scott, Even Thompson, many others, is so appealing, but these are not mainstream ideas. Would you not agree? The mainstream ideas surrounding how consciousness emerges regards computationalism which doesn’t seem to fit with this other approach. From where I sit, weak emergence and separability of classical systems are well founded, mainstream ideas that are far from being overturned.

This seems like a contradiction to me, nonlinear approach, systems approach. i.e. complex systems approach isn't at odds with computationalism (and is indeed part of it). Any time you read about spiking networks, you're reading about physiologically derived neuron models that are inherently nonlinear and give rise to chaos when coupled together with biologically derived coupling terms (excitatory, inhibitory, diffusive). Since there are several operating in a network, you have a complex system. Yes, this is mainstream (but in a nascent manner).

Evidence that it's mainstream:

Journals

Physics Review E now includes biological systems:
http://pre.aps.org/
neuroscience in that journal:
http://www.ncbi.nlm.nih.gov/pmc/articles/PMC2713719/

AIP: Chaos: Journal of Interdisciplinary Science (you should actually read this:)
http://chaos.aip.org/about/about_the_journal

A neuroscience example from Chaos:
http://chaos.aip.org/resource/1/chaoeh/v18/i2/p023102_s1?isAuthorized=no

Neuroscience:
http://www.ncbi.nlm.nih.gov/pubmed/10362290

Chemical Physics and Physical Chemistry:
http://onlinelibrary.wiley.com/doi/10.1002/cphc.200500499/fullPublished Authors:

Wulfram Gerstner (head of Computational Neuroscience department at the Mind-Brain Institute in Laussane, Switzerland)
http://icwww.epfl.ch/~gerstner//BUCH.html

Ermentrout:
http://www.pitt.edu/~phase/

Eugene Izhikevich's:
http://www.braincorporation.com/

Izhikevich wrote the textbook, "Dynamical Systems in the Neurosciences"
http://www.izhikevich.org/publications/dsn.pdf

There's Tsumoto:
http://pegasus.medsci.tokushima-u.ac.jp/~tsumoto/achieve/index-e.htmlIt's the complex global behavior that is unpredictable (which is why we have to run simulations, then interpret them geometrically: i.e. pretty pictures that we interpret from our personal experiences and what we know experimentally about the neurons). There's no "deterministic solution" to resulting complex behavior. Yet, it still can be quantified with average lifetimes, finding the different qualitative regimes through bifurcation analysis, etc, etc. You can go back and look at a bunch of numbers that the code spits out, but they're not very meaningful in the standard mathematical analysis. We need Poincare's geometric approach.
 
Last edited by a moderator:
  • #191
Q_Goest said:
Hi apeiron. I found that last post to be totally understandable. Not like the previous post at all. I really wonder how you manage to switch the flowery talk on and off like that. No offense intended.

Do you have examples of this flowery language? Systems science being an interdisciplinary subject, there is a problem that there are a lot of different jargons - ways of saying the same things coming from different fields.

But pinning down particles isn't important to classical mechanics. Sure, the real world isn't classical, but that's not the point. The conventional view is that quantum mechanics isn't a factor in how the brain works because there are sufficient statistical aggregates of particles that individual particles don't matter. They're simply averaged together. Is this "systems view" dependent on individual particles? Clearly Alwyn Scott for example, makes the point as do many others that classical mechanics has this 'more than the sum' feature already intrinsic to it and I’d think Scott’s views probably mirror your ideas fairly closely.

QM shows there is a problem even at the fundamental physical level. And the systems argument is that the same thing is going on over all scales of analysis.

Let me remind you what the systems approach actually is.

It says systems are formed hierarchically from two complementary kinds of causality - bottom-up construction and top-down constraint. And they are "a system" in that they are mutual or synergistic in their causality. Each is making the other. And so the develop (emerge) together in holistic fashion.

You are arguing from the reductionist viewpoint where there is a definite atomistic grain to reality. You have a bunch of stuff that already exists (it is not emergent) and it constructs some kind of global order (the forms emerge from the materials).

The reductionist story cannot of course address the reasons why the local atomistic grain exists. And while a "global state" might emerge, this is not the same as the global constraints of a system.

The temperature or pressure of an ideal gas is a macroscopic measurement, not a constraint. The constraints of the ideal gas would be the walls of the container, the external bath that keeps the system at a constant equilibrium, etc. All the general conditions that allow the system to have "a condition".

The systems approach instead says all that exists locally are degrees of freedom. Well, exist is too strong a word for random fluctuations - an unlimited number of degrees of freedom. So there is nothing definite at the scale of the local grain at the beginning.

But then unconstrained fluctuations lead to randomly occurring interactions and at some point constraints start to form as a result. Constraints have the effect of limiting local degrees of freedom. There is suddenly the beginnings of less freedom in the fluctuations and so a more definite direction to their action. The global constraints in turn become more definite, feeding back again on the local freedoms and the whole system undergoes a phase transition to a more ordered state.

This is a model familiar from many fields - Peircean semiotics, spin glasses, generative neural nets, Haken's synegetics, second order cybernetics, Hofstadter's strange loops, Salthe's hierarchy theory, Ulanowicz's ecological ascendancy, etc.

The key point is that there is now no definite local atomistic grain. It is part of what must emerge. It is like solitons or standing waves. The "particles" composing the system are locally emergent features that exist because of global constraints on local freedoms.

So QM says there are no locally definite particles until there is an act of observation - a decoherence of the wavefunction by an observing world.

And the same at the neuroscience level. A neuron's receptive field is constrained first by a developmental history (learning over time) and then even by attentional effects (top-down shaping over the course of 100 ms or so).

So reductionism gives you the frozen, static, simple description of a system. The local grain just exists (whereas in the systems view it is shaped by constraint). And the global constraints are unchanging (whereas in the systems view, they are part of what has to self-organise).

Consider the parallels with genetic algorithms. A pool of code (representing many local degrees of freedom) is forced to self-organised by imposing some general global constraints. A program is evolved. (This is not exactly what I am talking about, just an illustration that might be familiar to you.)

As to Scott, it is a long time since I spoke to him or read his book, so I can't actually remember how close he is to my current views. Though from dim memory, I think he was a little more simplistic - like John Holland and others pushing the Santa Fe view of complexity that was vogue in the US at that time. (Gell-Mann and Kauffman were deeper IMO)

But is that because of not having the initial conditions of the individual particles? Or not having the initial conditions of the classically defined states? Density, internal energy, entropy, etc... of the system is obviously independent of specific individual particle states, but not of the aggregate. So not knowing individual particle states will lead to indeterminate future states, and yes, the classical model is a model. But one has to show that there is a "meaningful difference" between having initial particle conditions and having initial classical conditions. I see no meaningful difference. Sure the classical approach isn't exact, but it is exact to the degree you have the initial conditions of the classical states and that's what's important unless the quantum mechanical states are being roped into causing different classical states by downward causation which isn't possible.

If everything is nicely "linear" - static and unchanging at the atomistic level due to static and unchanging global constraints - then coarse-graining can be good enough for modelling.

But the discussion was about systems that are complex and developing - such as brains. Where the global constraints are, precisely, non-holonomic. Where the local grain (for example, neural receptive fields) are dynamically responsive.

Taking an ideal gas as again a standard classical physics model of a system, note that we can impose a temperature on the system, but we cannot determine the individual kinetic freedoms of the particles. Just constrain them to a gaussian distribution.

So this is coarse graining in action. The individual motions are unknown and indeed unknowable (Maxwell's Demon). But they are constrained to a single statistical scale - that of the system's now atomistic microstates. And we can calculate on that basis. We have frozen out the sources of uncertainty (either local or global) so far as our modelling is concerned.

Can you provide an example of a global attractor? One that regards classical mechanics and can't be predicted by weak emergence? This is fundamentally where we disagree. I would say Benard cells are a perfect example of a weakly emergent structure, and I’d contest that’s a mainstream idea not just my own. Regarding mainstream, perhaps that feels like a knock so instead I’ll say “cutting edge” or something. Anyway, as Davies points out.

I gave you the Collier reference on Benard cells. Can you instead supply me with a reference where the global organisation was predicted purely from a local model of the molecules thrown together?

...I'll answer the rest of your points later...
 
  • #192
Lievo said:
Perhaps he had a look and the references were not supporting your claims unless a lot of creativity was involved. That sometime happens, doesn't it :wink:?

Sometimes people say that is what happened - yet strangely cannot then support their opinion in public. At least Q Goest argues his corner. If you were more confident of your views, perhaps you would too?
 
  • #193
Q_Goest said:
Also from Davies

However Davies does allow for some kind of emergence at the border between classical and quantum mechanics, which is where separability breaks down also.

Davies correctly says that global constraints are not some extra force. Force is a localised, atomised, action - efficient causality. Constraints are just constraints. They might sound "forceful" because they act downwards to constrain the local degrees of freedom (as I say, shape them to have some distinct identity). But they are a complementary form of causality. A constraining or limiting action, not a constructive or additive one.

He is also right in saying that you cannot have constraints emerging if the analysis only recognises the micro-scale. If you close off causality at the microscale in your modelling, there is indeed no room for anything else.

No, I'm not. Feel free... Regardless, why should it be controversial to suggest that there exists a physical reality at a particular moment in time? If the argument drops into quantum mechanics, there’s no point in arguing. At that point, we have to suggest that neurons interact due to some quantum mechanical interaction, which isn’t worth arguing about.

The synchronic~diachronic dichotomy is used frequently in the emergence literature - as in the Bedau paper you appear to prefer.

http://people.reed.edu/~mab/papers/principia.pdf

This is not about QM issues but the general modelling of structures and processes - systems modelling.

I think you misunderstand. My daughter has selective attention. I have no doubt they influence neural receptive fields! lol But that’s like saying the spaceship is caused by the gun that Lievo keeps talking about. Unless you can clearly define selective attention and neural receptive fields, it won’t help.

What is undefined about the concepts of selective attention and neural receptive fields in the literature?

And Lievo's guns are precisely not an example of anything I have been talking about. What could be a more reductionist view of reality than CA? He may see "spaceships" and "guns" looking at the rigid operations of a finite state automaton. But that "meaning" is completely absent from the CA itself. It emerges nowhere within it.

Bedau uses these guns to argue for weak emergence. And I agree, so far as emergence goes, it is as weak as can be imagined. I am talking about something else here.

I don’t disagree with anything you said here except that I’m sure IBM isn’t somehow influencing neuroscience with profits from their computers.

I was saying IBM dreams up these kinds of stunts to sell more supercomputers to universities.

Perhaps you haven't been keeping track of the controversies? IBM getting savaged by its own original Blue Brain scientist.

Why did IBM let Mohda make such a deceptive claim to the public?
I don't know. Perhaps this is a publicity stunt to promote their supercompter. The supercomputer industry is suffering from the financial crisis and they probably are desperate to boost their sales. It is so disappointing to see this truly great company allow the deception of the public on such a grand scale.

http://nextbigfuture.com/2009/11/henry-markram-calls-ibm-cat-scale-brain.html
 
  • #194
from apeiron's link (criticizing IBM):

In real life, each segment of the branches of a neuron contains dozens of ion channels that powerfully controls the information processing in a neuron. They have none of that. Neurons contain 10's of thousands of proteins that form a network with 10's of millions of interactions. These interactions are incredibly complex and will require solving millions of differential equations.

This is exactly right, it's not an easy problem. My Textbook:

"From Molecules to Networks: and introduction to cellular and molecular neuroscience"

It is no exaggeration to say that the tasks of understanding how intrinsic activity, synaptic potentials, and active potentials spread through and are integrated within the complex geometry of the dendritic trees to produce the input-output operations of the neurons is one of the main frontiers of neuroscience.

But we do have NEURON and GENESIS to help us with this, using the compartmental models developed by Rall and Shepherd. (Sheperd co-authored this chapters of the textbook).

I think what Q_Goest doesn't recognize is that passive currents (compartment models) only represent intrinsic currents, not the more interesting active currents (i.e. all the different channel dynamics. the feedback circuit between membrane potential and channel activation, the action potential.

Or the whole molecular/genetic behavior part that we've all continued to sweep aside. Neural activity also stimulates changes in genetic expression, so you have to talk about mRNA and transcription factors which modulate the geometry of the dendrites and the strength of both chemical and electrical synapses (by increasing and decreasing channel sites) so there's even more complicated feedback happening at the molecular level.

Yes, we use compartmental models. They're not nearly the whole story.
 
  • #195
Pythagorean said:
this is mainstream (but in a nascent manner).

Evidence that it's mainstream:

apeiron said:
the effects on selective attention on neural receptive fields has been one of the hottest areas of research for the last 20 years.

As a neuroscientist, I wish you both stop what is either poorly supported or wrong view about what is mainstream in neuroscience. Saying that this or this work is compatible with an interpretation is not an evidence that the interpretation is influencial (in other words I agree about the nascent manner but a nascent mainstream is simply self-contradictory). Saying that receptive field has been one of the hottest areas of research in the last 20 is simply wrong. Single units technics in animals have been the gold standard from maybe 1935 to about 1990. Since that what happens is an impressive rise of new technics devoted to recording in humans, mostly fMRI, and these technics can't record receptive field. You may not believe me, in which case simply look at how many papers you can retrieve with either brain + MRI or brain + receptive field in the last 20 y.

Q_Goest said:
However Davies does allow for some kind of emergence at the border between classical and quantum mechanics, which is where separability breaks down also.
Can you explain why separability should breaks down at this border, despite QM is perfectly computable?

Q_Goest said:
that’s like saying the spaceship is caused by the gun that Lievo keeps talking about. Unless you can clearly define selective attention and neural receptive fields, it won’t help.
These two concepts are perfectly and operationnaly defined. What's the problem?

apeiron said:
And Lievo's guns are precisely not an example of anything I have been talking about. What could be a more reductionist view of reality than CA? He may see "spaceships" and "guns" looking at the rigid operations of a finite state automaton. But that "meaning" is completely absent from the CA itself. It emerges nowhere within it.

Bedau uses these guns to argue for weak emergence. And I agree, so far as emergence goes, it is as weak as can be imagined. I am talking about something else here.
I'm sure you well aware that CGl is universal for computation. Meaning that any computable system you may think of can be implemented on it. So are you saying strong emergence is not computationnal? If not, on which basis would you decide that one CGL show or don't show strong emergence?
 
Last edited:
  • #196
We seem to be getting deeply into the neuroscience at this point, which is a perfectly appropriate place to go to study the neural correlates of mental qualia like free will. I would just like to point out at this point, as we have been talking about Pr, Ps, and M states, that we are free to adopt a physicalist perspective, and even choose to assume that physical states actually exist, and further assume that they form a closed system (whether it be Ps or Pr that are the most useful approaches, or whether either approach can be more or less useful in a given context). All the same, every single one of those is an assumption involved in a modeling approach, not a single one is an axiom (because they are not self-evident), and not a single one has convincing evidence to favor it. They are just choices made by the scientist to make progress.

In my view, which may be a minority but is logically bulletproof, there is no reason to imagine that we know that physical states either exist, or are closed, or even that it makes any sense to imagine that either of those are true beyond the usual idealizations we make to get somewhere. What is demonstrably true is that everything we mean by a physical state arises from perception/analysis of a class of experiences, all done by our brains. We can notice that our perceptions are correlated with the concepts we build up around the idea of a physical state, and we gain predictive power by building up those concepts, and not one single thing can we say beyond that. This is just something to bear in mind as we dive into the physicalist perspective, either at a reduced or systems level-- it is not at all obvious that this approach will ever be anything but a study of the neural correlates of mental states, i.e., the mental states may always be something different.
 
  • #197
Lievo said:
As a neuroscientist, I wish you both stop what is either poorly supported or wrong view about what is mainstream in neuroscience. Saying that this or this work is compatible with an interpretation is not an evidence that the interpretation is influencial (in other words I agree about the nascent manner but a nascent mainstream is simply self-contradictory).

Nascent mainstream is not self-contradictory. Nonlinear science is nascent to the mainstream, but it is mainstream (i.e. the work is published in well-known peer-reviewed journals). I don't know what you think my interpretation is; I responded to Q_Goest who (wrongly) put computational models and nonlinear science at odds.

You would agree, I hope, that Hodgkin's Huxley is a (~60 year old) mainstream model. It's a nonlinear model. Popular among computaitonal scientists is the Morris Lecar model (because it's 2D instead of the HH 4D, making phaseplane analysis and large networks much easier to handle).

Theoretical Neuroscience institutes and centers have popped up all over the world in the last 20 years, including exactly the kind of work I'm talking about (you have Redwood at Berkeley, CTN in New York, The Seung Lab at MIT, Computaitonal Neuroscience programs within neuroscience deparments)

We can argue about the semantics of "mainstream" but this a well-funded, productive area of research.

Here... from 2001, TEN years ago...

Neurodynamics: nonlinear dynamics and neurobiology
Current Opinion in Neurobiology
Volume 11, Issue 4, 1 August 2001, Pages 423-430
 
Last edited:
  • #198
Lievo said:
As a neuroscientist, I wish you both stop what is either poorly supported or wrong view about what is mainstream in neuroscience... You may not believe me, in which case simply look at how many papers you can retrieve with either brain + MRI or brain + receptive field in the last 20 y.

You lose credibility with every post.

Check the top 10 papers from Nature Neuroscience over the past 10 years.
http://www.stanford.edu/group/luolab/Pdfs/Luo_NatRevNeuro_10Y_anniv_2010.pdf

Attention effects, neural integration, homeostatic organisation and other forms of global top-down self-organisation feature prominently.

2001 Brainweb 2.0: the quest for synchrony
2002 Attention networks: past, present and future
2004 Homeostatic plasticity develops!
2006 Meeting of minds: the medial frontal cortex and social cognition

Your claim that scanning can't be used to research top-down effects is obvious nonsense.

http://www.jneurosci.org/content/28/40/10056.short
or
http://www.nature.com/neuro/journal/v3/n3/full/nn0300_284.html
or
http://www.indiana.edu/~lceiub/publications_files/Pessoa_Cog_Neurosci_III_2004.pdf

Or as wiki says...

In the 1990s, psychologists began using PET and later fMRI to image the brain in attentive tasks. Because of the highly expensive equipment that was generally only available in hospitals, psychologists sought for cooperation with neurologists. Pioneers of brain imaging studies of selective attention are psychologist Michael I. Posner (then already renowned for his seminal work on visual selective attention) and neurologist Marcus Raichle.[citation needed] Their results soon sparked interest from the entire neuroscience community in these psychological studies, which had until then focused on monkey brains. With the development of these technological innovations neuroscientists became interested in this type of research that combines sophisticated experimental paradigms from cognitive psychology with these new brain imaging techniques. Although the older technique of EEG had long been used to study the brain activity underlying selective attention by cognitive psychophysiologists, the ability of the newer techniques to actually measure precisely localized activity inside the brain generated renewed interest by a wider community of researchers. The results of these experiments have shown a broad agreement with the psychological, psychophysiological and the experiments performed on monkeys.
http://en.wikipedia.org/wiki/Attention

And are you suggesting electrode recording is somehow passe?

http://www.the-scientist.com/2009/10/1/57/1/
 
Last edited by a moderator:
  • #199
Lievo said:
So are you saying strong emergence is not computationnal?

Yes, by design Turing machines are isolated from global influences. Their internal states are simply informational, never meaningful. Are you unfamiliar with Searle's Chinese Room argument for example?

BTW, I don't accept the weak~strong dichotomy as it is being used here because it too hardwires in the very reductionist assumptions that are being challenged.

As I have said, the entire system is what emerges. So you have a development from the vaguely existing to the crisply existing. But if weak = vague, and strong = crisp, then perhaps you would be making a fair translation.
 
  • #200
Ken G said:
All the same, every single one of those is an assumption involved in a modeling approach, not a single one is an axiom (because they are not self-evident), and not a single one has convincing evidence to favor it. They are just choices made by the scientist to make progress.

I agree. An axiom is just an assumption formulated for the purposes of modelling. Calling it self-evident is just another way of saying I can't think of anything better at the moment.

Mathematicians of course have often believed they were accessing Platonic truth via pure reason. But scientists follow the pragmatist philosophy of CS Peirce.
 
  • #201
Yes, and if a convincing case can be argued for an equation along the lines of
M = Ps <--> Pr,
I think it would be even more interesting if the relation was something more akin to
P(M) = Ps <--> Pr,
where P means "the projection onto the physical." It would not be necessary for P to be invertible, so the physicalist claim that
M = P-1(Ps <--> Pr)
does not necessarily logically follow.

It is apparent that changes in the Ps <--> Pr interaction correlate with changes in M, and can be viewed as causal of changes in M because the detection of causality is one of the main properties of the P operation. However, if E signifies the evolution operator, involved in making some change, we still cannot say
E[M] = P-1(E[Ps <--> Pr]),
as that requires not only that P is invertible, but also that it commutes with E. Instead, what we can say is
E[P([M])] = E[Ps <--> Pr].
If we assert that EP = PE' as our definition of E', then we have
P(E'[M]) = E[Ps <--> Pr],
and this is the fundamental equation that systems-type neurologists study. But note we must wonder to what extent P is invertible, and to what extent P commutes with E. If neither holds, we have a particularly interesting situation.
 
  • #202
Ken G said:
Yes, and if a convincing case can be argued for an equation along the lines of M = Ps <--> Pr,

OK, you have lost me there. I'm not even sure if you are making a satirical argument. So you may have to explain it more fully :smile:.

A quick aside, to make a systems argument, one of the issues is having a systems notation. So I used set theory (as suggested by Salthe's hierarchy theory). But Peirce's sign of illation might be another route(http://www.math.uic.edu/~kauffman/TimeParadox.pdf).

Or because the = sign is always a statement about symmetry, and the systems view is about symmetry-breaking, then perhaps the notion of the reciprocal is the most accurate way to denote a process and its obverse - an invertible operation?

So Pr = 1/Ps could be another way of framing my argument. But then I was saying that Ps is the super-set, so not really invertible.

Anyway a logical notation that expresses the concepts is a live issue - and Louis Kauffman has highlighted the connections between Peirce, Nicod and Spencer-Brown. Another longer paper you might enjoy is http://www.math.uic.edu/~kauffman/Peirce.pdf.

Back to what you posted.

M = Ps <--> Pr - I translate this as the mind contains two constrasting views of causality, that are formed mutually as a symmetry-breaking of ignorance. But I think still the set theoretic view is more accurate.

My claim on P (models of physical causality) is that Ps = Pl + Pg. So systems causality is local construction plus global constraints.

Whereas Pr = Pl. So reductionist causality is based on just local construction.

However I then also claim that global constraints are still implied in Pr - they are just frozen and so can be left out of the modelling for simplicity's sake. Only the local construction has to be explicitly represented.

So Pr = Pl + not-Pg? Doesn't really work, does it?

But you raise an interesting issue just about the need for a formal notation that captures the ideas of systems causality. There is an abundance of notation to represent reductionist constructive arguments, but not really an equivalent for the systems view.
 
  • #203
Ken G said:
it is not at all obvious that this approach will ever be anything but a study of the neural correlates of mental states, i.e., the mental states may always be something different.
Sure it's logically sound. But could you think of a way to make a positive statement along this line?
 
  • #204
Pythagorean said:
Nascent mainstream is not self-contradictory. Nonlinear science is nascent to the mainstream, but it is mainstream (i.e. the work is published in well-known peer-reviewed journals). I don't know what you think my interpretation is; I responded to Q_Goest who (wrongly) put computational models and nonlinear science at odds.
I certainly agree your view is mainstream within the subfields that care about it. My point is that you're stretching the mainstream too far by applying it to neuroscience as a whole. The fact is most neuroscientists are not influenced by this view, because it does not have any impact in their day to day job. By the way, I guess what I don't like is that calling one's view mainstream is close to appeal to authority, which is hardly ever a good thing. When someone use that, usually that's because it doesn't have anything better to say. I know you can do better than that. :wink:
 
Last edited:
  • #205
Q Goest is the one that introduced mainstream as being scientifically meaningful.

I don't think most neuroscientists are influenced by the view, but I interpreted mainstream as pertaining to scientific peer review: the list of acceptable journals.
 
  • #206
In other words, these idea aren't beingrejected by the scientific community, and are passing peer review, even in the traditional neuroscience journals.

My traditional neuro advisor is happy to reach across the table and help us motivate our models biologically.
 
  • #207
Lievo said:
By the way, I guess what I don't like is that calling one's view mainstream is close to appeal to authority, which is hardly ever a good thing. When someone use that, usually that's because it doesn't have anything better to say. I know you can do better than that. :wink:

Yes, you have made your position clear on appeals to authority...

As a neuroscientist, I wish you both stop what is either poorly supported or wrong view about what is mainstream in neuroscience.

And why you continue to battle strawmen beats me.

It is only you who set this up as claims about being mainstream. You can check back to the post where I urged Q Goest to focus on the particular literature of neural receptive fields and top-down attentional effects if you like. You will see that it was hardly an appeal to authority but instead an appeal to consider the actual neuroscience.

https://www.physicsforums.com/showpost.php?p=3177690&postcount=36

Of course, you are now trying to say that the receptive field studies are not mainstream, or not new, or not influential, or something.

But all you have supplied as a source for that view is an appeal to authority. And a lame suggestion for a google search.

BTW, google scholar returns 142,000 hits for brain + fMRI and 164,000 for brain + receptive fields. What does that tell us?
 
  • #208
apeiron said:
You lose credibility with every post.
Your usual line when you got it wrong, it seems. :rolleyes:

apeiron said:
Check the top 10 papers from Nature Neuroscience over the past 10 years.
Small detail: this is Nature Reviews Neuroscience, which is not the same as Nature Neuroscience nor Nature.

Not so small detail: what you suggest is a bad methodology. Don't you know how to use pubmed?

I've check it anyway. Among these 10 papers, none but one discuss receptive fields, and only one reference is discussed. Among 114.

So again you said that "effects on selective attention on neural receptive fields has been one of the hottest areas of research for the last 20 years.", and again this is wrong. I can just wonder how you came to pretend otherwise despite it's wrong even using the data you emphasized.

apeiron said:
Your claim that scanning can't be used to research top-down effects is obvious nonsense.
Obvious strawman building, given I never made this claim. I'm better not: this my line of research. :biggrin:

apeiron said:
And are you suggesting electrode recording is somehow passe?
In a sense, yes. In a sense, no. The no part is that these kinds of data are still very interesting. The yes part is that data collection using electrode recording has been becoming too slow when compared to the 20-y-new neuroimaging technics in humans.
 
Last edited:
  • #209
Pythagorean said:
Q Goest is the one that introduced mainstream as being scientifically meaningful.

I don't think most neuroscientists are influenced by the view, but I interpreted mainstream as pertaining to scientific peer review: the list of acceptable journals.
Oh I just misunderstood you then. To me mainstream is "the common current thought of the majority". I've just missed the part where this definition was otherwise. My bad, sorry.
 
  • #210
apeiron said:
Yes, by design Turing machines are isolated from global influences.
If reality is computational, then all global influence are inside a TM. It may or not be the case, but if you assume the global environnement is not inside a Turing machine, there is no surprise you will conclude it's not computationnal.

apeiron said:
Their internal states are simply informational, never meaningful.
By definition?

apeiron said:
Are you unfamiliar with Searle's Chinese Room argument for example?
You actually give a sh.. about this argument?

apeiron said:
BTW, I don't accept the weak~strong dichotomy as it is being used here because it too hardwires in the very reductionist assumptions that are being challenged.

As I have said, the entire system is what emerges. So you have a development from the vaguely existing to the crisply existing. But if weak = vague, and strong = crisp, then perhaps you would be making a fair translation.
Is there any system that does not emerge, from these definitions?

apeiron said:
BTW, google scholar returns 142,000 hits for brain + fMRI and 164,000 for brain + receptive fields. What does that tell us?
*cough cough* I'm afraid it tells us you are not very familar with this engine. Brain + "receptive field" returns 88 700 hits. You see the trick?

You may notice that brain + MRI returns 1 130 000
 
Last edited:

Similar threads

  • · Replies 190 ·
7
Replies
190
Views
15K
  • · Replies 14 ·
Replies
14
Views
7K
  • · Replies 199 ·
7
Replies
199
Views
35K
Replies
14
Views
6K
Replies
11
Views
4K
  • · Replies 95 ·
4
Replies
95
Views
12K
  • · Replies 2 ·
Replies
2
Views
294
  • · Replies 60 ·
3
Replies
60
Views
9K
Replies
5
Views
3K
  • · Replies 27 ·
Replies
27
Views
7K