Freethought & Rationalism ArchiveThe archives are read only. |
12-10-2001, 06:30 PM | #61 | |
Veteran Member
Join Date: Aug 2000
Location: OK
Posts: 1,806
|
Quote:
I have my own thoughts, but was curious what others thought on the issue. |
|
12-10-2001, 06:50 PM | #62 | |
Veteran Member
Join Date: Mar 2001
Posts: 2,322
|
Quote:
|
|
12-11-2001, 01:14 AM | #63 | |
Veteran Member
Join Date: Dec 2001
Location: Darwin
Posts: 1,466
|
The Problem is a Human Brain may only run at a mere 40hz and the fastest personal computer may run at around 2+gigahertz , But the big difference is the brain has billions these processes all running in parallel. We need this because a the most sophisticated robot will not have a clue how to react if it encounters a tiger of a cow, it may well end up bolting from the cow and and attempting to milk the tiger <img src="graemlins/boohoo.gif" border="0" alt="[Boo Hoo]" /> . But will still kill us at chess.
crocodile deathroll Quote:
|
|
12-11-2001, 03:10 AM | #64 | |||
Veteran Member
Join Date: Aug 2000
Location: Australia
Posts: 4,886
|
Quote:
Quote:
Quote:
[ December 11, 2001: Message edited by: excreationist ]</p> |
|||
12-11-2001, 03:40 AM | #65 |
Junior Member
Join Date: Dec 2001
Location: Germany
Posts: 30
|
Artificial HUMAN minds ?
How about attempting to make an artificial insect-mind first, then we will see about the 'higher' minds. Maybe we need some artificial evolution to construct a mind similar to ours... or we might as well just copy a real one to harddrive. - morte |
12-11-2001, 04:32 AM | #66 | ||
Veteran Member
Join Date: Aug 2000
Location: Australia
Posts: 4,886
|
Quote:
I found a link to the AI I was talking about. It autonomously learns, similar to a human, except that its only input is typed words, like children's stories. It hasn't actually seen the world though, but it learns about the relationships between different words so that it can use them somewhat competently. <a href="http://www.newscientist.com/hottopics/ai/gasbrain.jsp" target="_blank">gasnets and artificial insects</a> This uses artificially evolved creatures that act like insects (to some degree). Quote:
<a href="http://www.newscientist.com/hottopics/ai/clever.jsp" target="_blank">A "kitten" with 40 million artificial neurons</a> "Neural networks must be fine-tuned to perform particular tasks. But no human programmer could write the software needed to refine a network as complex as the CAM (Cellular Automata Machine) brain. Instead, this will be generated using an approach that simulates biological evolution. Through random mutations and breeding of the "genetic material" that describes the structure and connections of the network, the program will be evolved over many generations to get the optimum design. Robokoneko will not be built until this work has been completed on a computer simulation of the robot cat." <a href="http://www.newscientist.com/hottopics/ai/godofthenorns.jsp" target="_blank">Steve Grand Interview</a> He created the "Creatures" games that had norns. (They're a bit like the creatures in "Black & White") "COMPARED WITH ANIMALS, WHERE YOU WOULD PLACE THEM (the "norns")? They're like ants. They don't have thoughts of their own and they are trapped in a sensory loop, where the environment tells their senses what to receive, the brain generates behaviour, and, much like us, that behaviour changes the environment, and so on . . ." Now he's working on a neural net controlled robot called <a href="http://www.cyberlife-research.com/about/anatomy.htm" target="_blank">Lucy</a> which he is trying to give <a href="http://www.cyberlife-research.com/about/brainintro.htm" target="_blank">an imagination</a>. It will have to be taught how to do things, like an infant - it isn't programmed with advanced behaviours. I think we'd make huge neural networks that have human-like intelligence before we can copy a human brain. [ December 11, 2001: Message edited by: excreationist ]</p> |
||
12-11-2001, 04:44 AM | #67 | |
Veteran Member
Join Date: Aug 2000
Location: OK
Posts: 1,806
|
Quote:
Thus the applicable question for this thread would be: are we no more than bean sorting machines that are fated to do what we do because of the "beans" that come our way? I course I believe there are several answers to this puzzle, one of which is that the laws of nature are not prescriptive but descriptive. The laws of nature do not "force" anything to happen. They describe what does happen. |
|
12-11-2001, 04:57 AM | #68 |
Regular Member
Join Date: Mar 2001
Location: Winnipeg, Manitoba, Canada
Posts: 374
|
Laws of nature describe what happens as opposed to forcing things to happen? Huh? This makes no sense to me.
The answer you are looking for is that yes, we are no more than bean sorting machines in the sense that our thought process also uses real imput. I honestly don't see why this is a problem. I think that you see it as a problem because you have some unnecessary baggage attached to your concept of the word choice. The best solution would be to drop it.. i suggest reading through BD-from-KG's posts again.. devilnaut |
12-11-2001, 06:26 AM | #69 | |
Veteran Member
Join Date: Mar 2001
Posts: 2,322
|
Quote:
We've got to either use criteria (bean holes) to categorize or just not categorize. Unless you think consciousness is some kind of a force separate from ourselves, there has to be some kind of a neurological mechanism to feed the input into the hopper and, in fact, that is what imaging techniques are beginning to show us. |
|
12-11-2001, 07:21 AM | #70 |
Guest
Posts: n/a
|
That article on Hal is very interesting. I like the developmental approach to artificial learning systems, I think that has to be integrated into AI philosophy on a much broader basis. Still, it is possible, no, likely that we are centuries away from developing computers with the cognitive skills approaching that of toddlers.
“In contrast to children, who can take years to learn the basics of language, Hal can be trained in just a few days. This is due both to the intensity of the training and because the algorithms have no distracting inputs, says Hutchens.” In other words, this system is still highly specialized. That is necessary but it makes humans very susceptible to the so-called Eliza effect, our tendency to anthropomorphize cognitive skills that superficially resemble our own. I’m not trying to downplay the accomplishments of such research projects. Sophisticated studies of restricted domains is vital to the development of AI. However, it’s very easy to be mislead when descriptions like “understanding” and “the language skills of a five year old” are used. It makes a lot of people forget just how much conceptual ground has yet to be broken. |
Thread Tools | Search this Thread |
|