FRDB Archives

Freethought & Rationalism Archive

The archives are read only.


Go Back   FRDB Archives > Archives > IIDB ARCHIVE: 200X-2003, PD 2007 > IIDB Philosophical Forums (PRIOR TO JUN-2003)
Welcome, Peter Kirby.
You last visited: Today at 05:55 AM

 
 
Thread Tools Search this Thread
Old 04-17-2003, 06:38 PM   #11
Moderator - Science Discussions
 
Join Date: Feb 2001
Location: Providence, RI, USA
Posts: 9,908
Default

Bill Joy, the chief scientist at Sun Microsystems, had an article in Wired Magazine a little while ago making a similar proposal to limit research in potentially "dangerous" fields like nanotechnology, genetic engineering, and artificial intelligence. The Foresight Institute (a nanotech thinktank of sorts) has a bunch of links to articles criticizing this idea here:

http://www.foresight.org/hotnews/#anchor1465777

Some other good articles on this subject are Forward to the Future: Nanotechnology and Regulatory Policy by Glenn Reynolds, http://www.slofi.com/Don't_Count.htm by the authors of The Social Life of Information, and The Future Needs Us! by physicist Freeman Dyson (this article starts out as a review of Michael Crichton's Prey, so you have to scroll down a bit). The Center for Responsible Nanotechnology is another good site on nanotech regulation.
Jesse is offline  
Old 04-17-2003, 09:55 PM   #12
Veteran Member
 
Join Date: Feb 2002
Location: Singapore
Posts: 3,956
Default

Quote:
Originally posted by Late_Cretaceous

Reminds me of how a politician here in Canada commented that scientists are not qualified to make ethical decisions (supposedly only politicians are) - he was talking about cloning technology.
Yeah, why not, since from I had seen, most politicians don't even understand science, itself.
Answerer is offline  
Old 04-17-2003, 10:18 PM   #13
Veteran Member
 
Join Date: Mar 2003
Location: Memphis, TN
Posts: 6,004
Default

The US govt is already doing this. Severly limiting stem cell research for example. Stem cells hold a great hope for many types of new research. We need their noses out of what we do, not in it further. Let the politicians etc worry about the ethical issues as they come up, but do not stop basic research.
BioBeing is offline  
Old 04-18-2003, 03:23 AM   #14
Veteran Member
 
Join Date: Sep 2001
Location: Los Angeles Area
Posts: 1,372
Default

It never ceases to amaze me how many of us are preoccupied with doomsday scenarios. Along the same lines as Bill Joy's falling sky scenarios for nanotechnology is the so-called Technological Singularity that might arise if we were to pursue Artificial Intelligence. You can feel the general fear of these speculative futures in popular culture, in movies like The Matrix, and in legislative attempts to rein in the progress of science.
fando is offline  
Old 04-18-2003, 04:10 AM   #15
Veteran Member
 
Join Date: Dec 2001
Location: North of the South Pole
Posts: 5,177
Default

Quote:
Originally posted by Late_Cretaceous
Reminds me of how a politician here in Canada commented that scientists are not qualified to make ethical decisions (supposedly only politicians are) - he was talking about cloning technology.
:banghead:
Now the only requirement is ethical politicians...:banghead:
mongrel is offline  
Old 04-19-2003, 07:40 AM   #16
Veteran Member
 
Join Date: Mar 2003
Location: god's judge (pariah)
Posts: 1,281
Default

Quote:
Originally posted by mongrel
Now the only requirement is ethical politicians...:banghead:
They make those?
keyser_soze is offline  
Old 04-19-2003, 07:40 AM   #17
Veteran Member
 
Join Date: Aug 2000
Location: the 10th planet
Posts: 5,065
Default

not any more
Marduk is offline  
Old 04-19-2003, 07:52 AM   #18
Veteran Member
 
Join Date: Mar 2003
Location: god's judge (pariah)
Posts: 1,281
Default

Quote:
Originally posted by marduck
not any more
Had me going they did!
keyser_soze is offline  
Old 04-19-2003, 08:06 AM   #19
Veteran Member
 
Join Date: Mar 2003
Location: Calgary
Posts: 1,335
Default

A couple of weeks back in the Sydney Morning Herald, I recall reading a snippet where this guy (sorry for the vagueness) was suggesting that, at present, there is no need for scientific ethics. His rationale was that the philosophers are still debating the concept of ethics, so how can we apply limits on action that are based on a moving definition?
Extend that to having politicians with an inverted concept of ethics and no knowledge of science being the ones making the rules.... not promising.
Godot is offline  
Old 04-19-2003, 09:11 AM   #20
Moderator - Science Discussions
 
Join Date: Feb 2001
Location: Providence, RI, USA
Posts: 9,908
Default

Quote:
Originally posted by fando
It never ceases to amaze me how many of us are preoccupied with doomsday scenarios. Along the same lines as Bill Joy's falling sky scenarios for nanotechnology is the so-called Technological Singularity that might arise if we were to pursue Artificial Intelligence. You can feel the general fear of these speculative futures in popular culture, in movies like The Matrix, and in legislative attempts to rein in the progress of science.
Except the "technological singularity" isn't necessarily a doomsday scenario, it's often viewed as a positive thing by those who see it as likely. If Moore's law continues for another few decades I think there's a pretty good chance something like this will happen...see this article by Ray Kurzweil, for example:

The Law of Accelerating Returns
Jesse is offline  
 

Thread Tools Search this Thread
Search this Thread:

Advanced Search

Forum Jump


All times are GMT -8. The time now is 04:34 PM.

Top

This custom BB emulates vBulletin® Version 3.8.2
Copyright ©2000 - 2015, Jelsoft Enterprises Ltd.