Freethought & Rationalism ArchiveThe archives are read only. |
02-25-2003, 07:49 AM | #11 |
Veteran Member
Join Date: May 2001
Location: US
Posts: 5,495
|
Qualities
I happen to like qualia, although I'm not sure whether I'm justified or not. Here's why:
1. Sense data arrives at the "surface of the mind", if you will. 2. There are certain measurable "raw" or "natural" qualities inherent in that data such as redness, hardness etc. I agree that in these examples there is no need for qualia - why substitute for the real thing? 3. As we understand our environment we ascribe some very esoteric qualities to certain aspects of it. For example, "The politics of the middle east are very complicated." The words "politics" and "complicated" are not natural sense data in the same way as the example in #2. 4. In making statements like "There are some politics involved" or "This is a complicated issue" it seems that we have invented compound non-material qualities in order to describe reality. Thus, while I don't admit qualia as being a mysterious set of additional qualities of reality I think they can be useful in a mind/brain duality debate. In my mind , qualia are highly abstract qualities of reality that we have learned to share and understand intersubjectively. Nevertheless, I believe their existence is dependent on the phenomenal operation of the brain and qualia can be used in developing an ontology for how we know things are "complicated" and "political" etc. Comments? Cheers, John |
02-25-2003, 09:18 AM | #12 |
Regular Member
Join Date: Feb 2003
Location: Amman, Jordan
Posts: 258
|
Mind/Brain Duality II/ Machines
If the conditions required for consciousness to emerge are pinned down in the not-so-far future. Do you think that we will be able to build machines that are conscious?
I think Edelman has made some progress in his Darwin VII machines... here is an abstract from one of his papers: Machine psychology: autonomous behavior, perceptual categorization and conditioning in a brain-based device. Krichmar JL, Edelman GM. W.M. Keck Laboratory, The Neurosciences Institute, 10640 John Jay Hopkins Drive, San Diego, CA 92121, USA. krichmar@nsi.edu In studying brain activity during the behavior of living animals, it is not possible simultaneously to analyze all levels of control from molecular events to motor responses. To provide insights into how levels of control interact, we have carried out synthetic neural modeling using a brain-based real-world device. We describe here the design and performance of such a device, designated Darwin VII, which is guided by computer-simulated analogues of cortical and subcortical structures. All levels of Darwin VII's neural architecture can be examined simultaneously as the device behaves in a real environment. Analysis of its neural activity during perceptual categorization and conditioned behavior suggests neural mechanisms for invariant object recognition, experience-dependent perceptual categorization, first-order and second-order conditioning, and the effects of different learning rates on responses to appetitive and aversive events. While individual Darwin VII exemplars developed similar categorical responses that depended on exploration of the environment and sensorimotor adaptation, each showed highly individual patterns of changes in synaptic strengths. By allowing exhaustive analysis and manipulation of neuroanatomy and large-scale neural dynamics, such brain-based devices provide valuable heuristics for understanding cortical interactions. These devices also provide the groundwork for the development of intelligent machines that follow neurobiological rather than computational principles in their construction. |
02-25-2003, 02:57 PM | #13 | |
Veteran Member
Join Date: Dec 2001
Location: Darwin
Posts: 1,466
|
Re: Re: Mind/Brain Duality
Quote:
IMO consciousness is an emergent property of genetic promoting! (oh dammit! I meant promting) like the basic scaffolding of the brain was genetically determined, and individualized consciousness emerged later. I post that in a hell of a rush. |
|
02-25-2003, 05:50 PM | #14 | |
Regular Member
Join Date: Feb 2003
Location: Denmark
Posts: 122
|
Re: Mind/Brain Duality II/ Machines
Quote:
|
|
02-26-2003, 01:40 AM | #15 | ||
Veteran Member
Join Date: Dec 2001
Location: Darwin
Posts: 1,466
|
Re: Re: Mind/Brain Duality II/ Machines
Quote:
I just found it, it was the psychologist Stuart Sutherland quoted here by Susan Greenfield in this interview Quote:
|
||
02-26-2003, 04:46 AM | #16 |
Regular Member
Join Date: Feb 2003
Location: Amman, Jordan
Posts: 258
|
If there is no metaphysical soul, and consciousness is explained by physical processes (regardless if the explanation came from cognitive sciences, molecular neurosicence, quantum mechanics, or all combined) then I would suppose that we can simulate those conditions on a highly sophisticated computer. As shown in Edelman's article I cited, the use of selectional rules in building machines gets them closer to perform "brain-like behaviour" like perceptual categorization. He did that, just by using a value system, which on a much more sophistacated level corresponds to what makes up our emotions. I see endless possibilities of simulations as our knowledge in the science of consciousness expands.
Marvin Minsky once wrote a beautiful article about that. "Will robots inherit the earth?" Regarding Frotiw's notion about how could we know that Data in startrek and thus a machine is conscious, I have a question. How could you know that anybody is conscious? or is feeling what you feel when they're hurt? Consciousness, by definition, is a private experience. |
02-26-2003, 07:57 AM | #17 | |
Junior Member
Join Date: Feb 2003
Location: Chicago
Posts: 95
|
Re: Re: Mind/Brain Duality II/ Machines
Quote:
To dunk into Wittgenstein, you can't verify the existence of anyone else's pain, let alone a machine's. They can look hurt, the x-ray could be gruesome, an EEG can indicate that neurons are firing off like mad in the "I'm in pain" part of the brain, they can tell you they're in pain, but it remains unverifiable that they are "feeling" pain. Couldn't that pain be a part of our "programming"? Is our feeling of pain really any different than your machine's programmed response to scream when kicked? There are people who don't feel pain. (in light of the above, I should say they claim not to feel pain, and they certainly carry on like they can't feel pain. I doubt we can verify the non-existence of pain in others, either.) It's rare, but it happens, and without the disincentive of headaches, sores and the blinding agony of broken bones, people who have this neurological anomaly often die young. They are perhaps victims of flawed programming. To sum up, I don't think that's a good test for consiousness, in that the measure is solely what we consider consiousness to be. If a machine as advanced as ST TNG's Data is ever built, to dismiss it as not concious because it doesn't feel pain the way you or I do, or it doesn't see blue the same way we do would be tantamount to calling someone illiterate because they can't speak your language. That's all I got for now. -Neil |
|
02-26-2003, 08:10 AM | #18 |
Regular Member
Join Date: Feb 2003
Location: Amman, Jordan
Posts: 258
|
Neil, thanks for the analogy! My point exactly
|
02-26-2003, 12:50 PM | #19 | |
Veteran Member
Join Date: May 2001
Location: US
Posts: 5,495
|
Re: Re: Re: Mind/Brain Duality II/ Machines
Quote:
Cheers, John |
|
02-26-2003, 12:52 PM | #20 |
Veteran Member
Join Date: Dec 2001
Location: Darwin
Posts: 1,466
|
I wonder if the conscious computer started to get bored would you observe it playing Solitare?
|
Thread Tools | Search this Thread |
|