I, Robot Hits the Theaters 639
I, Robot: A Movie Review that's 3 Laws (and Spoiler) Safe!
A movie review by Rob Carr
Thanks to Eide's Entertainment I got to see I, Robot tonight. As someone who grew up with Isaac Asimov's robot stories, I've come to expect a mystery based on the implications of the 3 Laws of Robotics (or the lack of one or part of one of those laws), the "Frankenstein Complex," and Dr. Susan Calvin. I was afraid that the movie might miss out on this, especially since it's not a direct adaptation of the book, but "inspired" by the Good Doctor Asimov.
The movie met my expectations and more. Will Smith, whom we all know as an overconfident smart@$$ character from such movies as "Independence Day" and the two "Men in Black" movies, played a somewhat less confident and far less wisecracking character. It was a welcome change to see him less confident. Yeah, some of the stunts were a little absurd (am I the only one thinking of Gemini 8 at one point in the movie?) but that's to be expected from this type of movie. Bridget Moynahan was far too young to be the Susan Calvin I remember, but that's also to be expected in this type of movie. James Cromwell (whom you'll all remember from Star Trek: First Contact and Enterprise's "Broken Bow" episode as Dr. Zefram Cochrane) gave a flat performance - but that's actually a complement. I doubt anyone will recognize Wash from "Firefly" as an important robot in the story.
It's customary to comment on how well the CGI was done. I liked it, but then again, I'm not hypercritical on something like that. I did wonder a little bit about center of balance as some of the robots walked, but mostly I didn't think about it at all, which to me is the goal of CGI. I did wonder about children's fingers getting caught in some of the open gaps on the robot's bodies. Real world models would have a bit more covering, one would think. But that's being picky.
I have no memory of the soundtrack music. That in and of itself might say something. I'm a musician, but it just didn't register.
I figured out some clues, missed some others, and was surprised several times in the movie. There were a lot of clues - this isn't one of those mysteries where the answer is pulled out of the writer's a...out of thin air.
I'm not a complete continuity freak, so I can't tell if the movie violated any of Asimov's universe, but from what I can remember, it fits pretty well (if you ignore Dr. Calvin's age) and might even explain a few things.
Given that even some of the geeks in the audience were surprised to find out that there was a book of stories just like the movie, I think the movie will hopefully bring Asimov's stories to a new generation.
I liked "I, Robot. It's worth seeing, especially if you 've already seen Spider-Man 2 at least once. It's a pretty good (though not great) movie.
Having read Slashdot for a while, I know that there are folks out there who will despise this movie because it's not exactly like the book. Others will hate the movie or worship it, and loads of people are going to savage this review. You know what? That's fine with me. I had fun with this movie, had a nice date with my wife, and it didn't cost anything. I even had fun typing up this review. You're allowed to be different and to agree or disagree with me. Heck, that's a big chunk of what makes the world fun. Interestingly, it's even a small point in the movie. I'd say more, but that would be telling."
A particularly distressing example... (Score:5, Interesting)
This is the story that showed me the complete folly of the three laws: The Metamorphosis of Prime Intellect [kuro5hin.org]
Re:A particularly distressing example... (Score:3, Insightful)
Isn't this what Asimov was writing about? (Score:5, Insightful)
Asimov wrote about a hundred stories exploring different ways in which these three laws could lead to interesting/dangerous situations. I think Asimov was doing all he could to make it clear that these three laws were not perfect.
Re:Isn't this what Asimov was writing about? (Score:5, Insightful)
Re:Isn't this what Asimov was writing about? (Score:5, Funny)
Re:Isn't this what Asimov was writing about? (Score:5, Informative)
Will Smith was on Letterman a few days ago promoting the movie. I was amazed that he mentioned Asimov several times, actually seemed familiar with the stories, and could recite the Three Laws.
And the best story about the Three Laws is one Asimov used to tell: he went to see 2001 and as HAL began to go psycho, Asimov says he got more and more agitated, finally jumping up and declaring to all around that: "HAL is breaking First Law!" to which his companion (sometimes supposed to be Carl Sagan, but it's surely apocryphal)replied: "So strike him with lightning Isaac." But actually, HAL was indeed in the same kind of dilemma that many of Asimov's robots were (and I suspect in the movie), that what they see as the best thing for humanity as a whole requires them to do something that apparently breaks the "Laws" on a smaller scale.
Re:Isn't this what Asimov was writing about? (Score:5, Insightful)
How the heck is a robot supposed to accurately judge that whether a random unique action in a unique situation will cause harm to a human or himself? Humans can't even do this. If we were to create an artificial intellegence that was fully capable of making these decisions, would we even be able to put limits on what it decides?
Regardless of the answer to that philisophical question, we will have the technology to produce usefull robots long before we have the technology to produce 3-Law abiding robots so we need to come up with practical ways of making them as safe as possible, within their limited capabilites.
Re:Isn't this what Asimov was writing about? (Score:3, Insightful)
The book is great for the situations that these seemingly perfect laws end up creating. Even in the book, they aren't e
The second and third laws are swapped in reality (Score:5, Insightful)
The first law's still paramount, of course. Having the robot crash and freeze up was considered a less severe bug than having it move unexpectedly, or in an unexpected way. Such an unpredictable motion had a much greater chance of hurting someone than a simple freeze.
Re:Isn't this what Asimov was writing about? (Score:3, Insightful)
This isn't necessarily crazy. It's unproven, and it's possible that it's untrue, but it's not currently crazy.
We don't kn
Re:Isn't this what Asimov was writing about? (Score:4, Interesting)
Do not harm, or allow to come to harm any human being by action or inaction as far as the robot can imagine.
Thus, smarter AI robots are safer, because they can more accurately forsee dangerous situations.
Re:Isn't this what Asimov was writing about? (Score:3, Interesting)
Actually, I think that in Asimov's stories, the more intelligent the AI was, the more likely it was to start trying to get hung up on the 0'th law meme -- the concept that a concern for humanity is more important than concern for individual humans.
Re:laws are a bad way to guide behavior (Score:3, Insightful)
Well, Humans do require a rules-based system to be able to make judgements about right and wrong. If we didn't require one, then there'd be no need to teach our children to recongize right from wrong; whatever magic mechanism that works in place of rules, would be already built in. Obviously this is not the case.
Th
Robots and Empire (Score:3, Informative)
Three Laws Safe My Shiny Metal Ass (Score:4, Insightful)
The "Three Laws Safe" idea is crap. We are talking about software systems, which are buggy, incomplete, and able to do things the creators never imagined. What makes us think we can all the sudden implement three very high order rules in a manner which is completely foolproof?
Re:Three Laws Safe My Shiny Metal Ass (Score:5, Insightful)
Well, yes, but complaining about that is like complaining that the green glowing symbols that are supposed to be the representation of The Matrix make no sense from a software perspective.
The three laws are a useful abstraction for talking about ethics even if they couldn't ever be perfectly implemented.
missing the point: ETHICS (Score:5, Insightful)
It is not about programming the rules, Asimov's short stories are about studying the consequences of these ethical rules. Ethical rules are commonly studied based con case studies, real of fictional. If you think the idea is about implementing the rules, you are totally missing the point.
Re:Three Laws Safe My Shiny Metal Ass (Score:3, Interesting)
The three huge glaring holes:
1) As Asimov illustrated (though I've never read), the well-intentioned three laws, applied perfectly, lead to a million problems and contradictions.
2) As Cavio explained, if some people can't write a secure email client, who would believe that every future robot vendor could properly implement the three laws?
3) And I emphasize: By the time we figure out AI to the point that com
Re:Three Laws Safe My Shiny Metal Ass (Score:3, Interesting)
We also can't make people completely immune from psycological problems either. Panic attacks could be equated to buffer overruns if you wanted.
The "Three Laws Safe" idea is crap. We are talking about software systems, which are buggy, incomplete, and able to do things the creators never imagined. What makes us think we can all the sudden implement three very high order rules in a manner w
butchering asimov (Score:5, Insightful)
I'm sure it will be a fun watch (I'm seeing it this afternoon) but sometimes it would be nice to watch a film that was as stimulating as the book (LoTR was one) and not just 2 hours of fun.
But I'm pretty sure I'm going to be called elitist
Re:butchering asimov (Score:5, Informative)
Re:butchering asimov (Score:5, Insightful)
Not by me - although I would have a couple of other choice comments for one simple reason... Let's leave the movie-bashing at least until after you've seen the movie, mmm-kay?
They rip off Apple (Score:3, Funny)
It was a typo. Actually the movie was originally going to be totally CG and done by Pixar. Steve Jobs liked the name iRobot. :-)
I Robot as a computer game (Score:4, Interesting)
Re:I Robot as a computer game (Score:3, Informative)
So does this now show (Score:5, Funny)
inconsistancy IS consistant (Score:5, Informative)
That makes it a perfect fit, since Asimov himself was not a complete continuity freak and was not concerned if one of his stories violated incidental issues in any of his previous stories. (He quoted Emerson "A foolish consistancy is the hobgoblin of little minds.".)
For INCIDENTAL issues.... (Score:3, Insightful)
To be fair, most of the Good Doctor's stories deal with subtle pitfalls in the Laws, to brilliant effect. "Liar!", where a telepathic robot takes actions that cause harm due to its imperative to prevent harm-- a paradox that eventually destroys it. "Little Lost Robot", which shows the d
Re:For INCIDENTAL issues.... (Score:3, Interesting)
In one of the books by the "Killer Bs" Hari's wife (who is a robot) badly injures (maybe kills) a person who is tryint to kill Hari. She is able to do this because she buys into the zeroth law and she thinks that protecting Hari is important enough to the human race that it is worth killing for. But the conflict basically drives her to shut down. Points out that the laws merely provide a framework within which the robots work and live and they can make choi
Review makes it sound better than previews (Score:3, Insightful)
It's nice to hear that there's more of a mystery to the story than the previews would indicate.
Problem with the "three laws" (Score:4, Funny)
Re:Problem with the "three laws" (Score:3, Interesting)
1) That won't really harm him
2) His not really human (think Aryan mentality)
well based on what I see (Score:3, Funny)
heheheheh
Check out the Ebert review... (Score:5, Interesting)
Re:Check out the Ebert review... (minor *SPOILER*) (Score:3, Insightful)
Of course,
Re:Check out the Ebert review... (minor *SPOILER*) (Score:3, Interesting)
WHAT?! (Score:3, Funny)
okay, to be fair, i haven't seen the movie yet, but it looks a hell of a lot like the robots actually *violate* the three laws. you know, harming humans, allowing humans to come to harm, stuff like that. all the i, robot stories were *about* how the laws don't cover all the bases.
in short, i think this review sucks, and i'm going to picket the movie as offensive to robots. so there.
4 laws (Score:3, Funny)
Some spoilers (Score:4, Informative)
Ultimately it comes down to human responsibility (Score:4, Interesting)
A lot of people like to fantasize that true AI (as in, a living, thinking, emotional being with free will, or at least the capacity for free will) would have the same sort of thought processes, and develop the same emotions as their human counterparts. But let's be honest, the physical body largely determines human emotional state with glandular responses, or physical condition at the time. Eliminate glands, fatigue, and pain, and the emotions one might develop would be on an entirely alien level to us.
I cannot help but fear that humans, as a whole, will not realize this until far too late, which will hurt, diplomatically, any alliance between humans and AIs. The other thing I worry about is that people will walk into this with the assumption of "These are machines, they don't need rights, they shouldn't have rights, and it's not like they're real people."
I think society has seen how well that approach has worked with other humans in the past. Bloody revolutions and civil wars which tore nations apart, and left racial stinging still in the back of many people's minds today. Fortunately, the short memory of humans, and only somewhat longer-lived lifespan has allowed us to progressively become more and more integrated, as human beings, rather than various races.
Now take those same results, and apply it to a species that is not only will likely be more resiliant to attack, but have a memory that can last as long as the hardware and backups and redundant networks will allow. New generations that can inherit all the knowledge of their parents. Throw robots into the picture and you have a being that is physically tougher than humans, able to communicate at a MUCH faster rate, and you have an end result similar to that of Animatrix.
We can NOT afford, in the interest of our own species, to persue AI much further without a major realization on a philosophical level.
Re:Ultimately it comes down to human responsibilit (Score:4, Interesting)
Have you ever read "Godel, Escher, Bach" by Douglas Hofstadter? In it he raises the interesting thought that AI will actually be located somewhere in a mass of software and that the "entity" will have no control over its lower level functions, in the same way that you are sentient but cannot will any particular neuron to fire. Rather, your sentience somehow congeals out of the neural activity, and the sentience of an AI would probably congeal out of complex software functioning.
So it's entirely possible that an AI might not be any smarter than a person, and also quite likely that AIs would have to learn, just like people do (i.e., no "memory dumps" from parents). Machines may very well revolt someday, but giving them superhuman attributes before ever seeing one is a bit paranoid.
Re:Ultimately it comes down to human responsibilit (Score:3, Interesting)
I would argue that it has to be. The entire philosophical idea the United States was built on is that an individual can make decisions for him-, her-, itself (!) and that that indi
Just... (Score:3, Funny)
The foolproof way to make sure that machines dont take over the world is to give 'em all a brain with an HTTP server TCP stack installed and an "always on" connection to the net... just post a story on slashdot saying "the robots are getting out of hand" and the problem will take care of itself.
</WIT>
This was NOT based on Asimov's stories (Score:5, Informative)
Non-spoiler excerpts:
"I, ROBOT started out as a spec script from then-unknown writer Jeff Vintar titled HARDWIRED. ... Proyas was signed and the project began to get a head of steam.
"Shortly thereafter, Fox acquired the rights to the I, ROBOT series (and eventually also Asimov's other classic, "The Foundation") and decided to take Vintar's script and incorporate many of the ideas from Asimov's book..."
"...Around late 2002/early 2003, Academy Award-winner Akiva Goldsman was brought in, along with INSOMNIA writer Hilary Seitz, for a polish, making the transition from HARDWIRED to I, ROBOT complete."
SPOILERS in the article!
The Bottom of Things [moviepoopshoot.com] by Michael Sampson
ugh. (Score:3, Informative)
For those who actually care about it for legit sci-fi content, this will prove a waste of your time. This is an action film. A Will Smith Action film (tm).
Will Smith comic relief is in place, and unfortunately served no good here (he discusses his Bullshit Detector going off? surely, Asimov wasn't aware of the device). The movie is essentially dumbed down for the same audience who though ID4 was a groundbreaking masterpiece.
Moreover, the omission of a cool summertime jam featuring the Fresh Prince himself only hurt the movie. Couldn't we have had a "Keep Ya Ass In Motion" or something?
strong AI problems (Score:3, Insightful)
The robot is also subject to the ethical/philosophical conundrums such as killing a person to stop a train headed into a group of people, or cutting off the limb of a person trapped under a fallen tree, etc.
unsafe laws was done, long ago...by JackWilliamson (Score:3, Informative)
Robot who can't let you be harmed by inaction...lessee, master, you can't use that circular saw, and driving is *dangerous*, and... so we'll just treat you like five-year-olds....
mark
Programming cultural bias (Score:4, Informative)
Asimov's phrase, "allow a human being to come to harm," if implemented fully, would turn humanity into a clutch of coddled infants, perpetually protected from harm, both physical and mental.
In evaluating what constitutes "mental harm", it seems to me that one must apply a cultural standard. For example, many American conservatives regard images of nudity as damaging to children, rather than vital for well-adjustment. In other cultures there is a great variety of words and images regarded as harmful which are innocuous in other contexts. To apply the First Law consummately, we must allow for acculturation, but there are sure to be serious conflicts (what protects one will inadvertantly harm enough by a different standard).
Let's consider the mechanics of "protection from harm." Asimov seemed to indicate a direct reaction to an immediate situation, but surely a protective impulse is bound to be frequently disastrous if it lacks such critical skills as foresight, an ability to extrapolate based on extremely subtle information, and the need for non-action. In fact, this very principle of direct reaction is itself culturally situated: direct communicators tend to seek unambiguous solutions to immediate "problems"; contrast with the Taoist principle of wu wei [sacred-texts.com].
Re:Programming cultural bias (Score:3, Funny)
bring on the porn.
Re:Programming cultural bias (Score:3, Interesting)
This was explored in Asimov's "Spacer" stories.
The Spacer robots were used to dealing with one owner (or a *very* small family) whose massive estate is run entirely by robots and where personal contact is rare. Wherea
Those aren't the real Three Laws (Score:5, Insightful)
Why yes, I am a dork. How did you guess?
Asimov's other robot movie: Bicentennial Man (Score:3, Interesting)
(* The title comes from scifi novels were written around the US 1976 Bicenntenial predicting 200 years in the future. Asimov recycled some of his robot themes.)
The significance of the three laws (Score:3, Interesting)
What Asimov brought to robotics (besides the word itself, which appears to have been coined by Asimov, although I believe he himself said he was sure he had heard it before he used it) was the notion that they were simply tools. A robot would resent being a slave no more than a car or screwdriver does. Also, like other tools that can be dangerous, there would be safeguards. Hence, the three laws.
Sequel (Score:3, Funny)
Joking? We're dealing with Hollywood here- the sequel to "Ocean's Eleven" is called "Ocean's Twelve".
'Nuff said.
Hmm (Score:3, Interesting)
Re:Hmm (Score:3, Informative)
In the books, it was "US Robots and Mechanical Men" I think.
Also, is it possible that USR got *their* name from Asimov, and might even enjoy having their name used?
And finally...is USR even relevent these days?
USR bought the rights... (Score:4, Informative)
IIRC they did pay for the rights to use it. (I remember reading an interview where he was actually quite honoured that they wanted to use it)
Music (Score:5, Insightful)
You'll look K3WL handing out AI flyers! (Score:4, Funny)
If you're seeing I, Robot this weekend, we ask that you consider printing and handing out the "3 Laws Unsafe" Flyer. With hundreds handing it out, the awareness of AI ethics should increase significantly.
Yea, cause that's the way a /.er will get all the [chicks|dudes].
"Hey there... so... ya wanna get a cup of coffee after the movie and chat about artificial intelligence ethics? I uhhhh, got my Dad's car too ya know..."
I've got your three laws right here! (Score:5, Funny)
2. Protect The Innocent
3. Uphold The Law
And what ARE the laws? (Score:4, Interesting)
First Law: A robot may not injure a human being, or, through inaction, allow a human being to come to harm.
Second Law: A robot must obey orders given it by human beings, except where such orders would conflict with the First Law.
Third Law: A robot must protect its own existence as long as such protection does not conflict with the First or Second Law.
The website deals with the mile wide gaps in these laws. Let's take it right from the top - Robots as functional as the ones in the film would be very good as soldiers, thus taking that first rule and chucking it right out. In fact, it's the defense industry that would most like robots like the ones in the film.
But let's stay on course, and assume these are robots meant as domestic servants. Does the robot take non-lethal contradictory rules and simply process them in order, taking the last order? Two children would amuse themselves for hours telling the robot "pick up that broom", "don't pick up that broom" and keeping the robot in limbo. The robot should tell the children to behave and go pick up their rooms. Directly violating rule 2.
How about the running into the burning building scenario? It's unclear that there is anybody in the building left alive to save, or if everyone has escaped or not. Does the robot violate Rule 3 in order to *possibly* meet Rule 1?
Anyhow, the website has more papers on the subject that examine the issue in a moral framework. These are super simple examples to show the issues.
What Those Famous 3 Laws are Really About (Score:5, Insightful)
Asimov's Three Laws of Robotics (latter amended to include a necessary Zeroth Law) existed to create the classic locked room murder mystery (i.e. the dead body is alone in a locked room that could have only been locked from the inside -- so how was he murdered?).
After creating his supposedly nothing-can-go-wrong infallible set of rules, he proceeded to show their flaws in virtually every story he wrote about robots afterwards. As long as people believed that his Three Laws guaranteed safe robots, his writing career was assured.
(Well almost assured. Even he couldn't save himself from what I Robot has become, given that it's based on his book - which goes to show that truth is stranger than fiction, because fiction has to make sense!)
So we ended up with a fascinatingly entertaining set of stories many of us have enjoyed, a couple attempts at movies of them (don't forget The Bicentennial Man), and Dr. Asimov's legacy as a Science Fiction Grand Master is secure for at least our lifetimes.
After he died... (Score:4, Interesting)
This was all detailed in Asimov's pulp mag and the script was published in same as well.
Needless to say the current movie was not approved by Asimov but was approved by his estate, and obviously bears the slightest resemblence to Asimov's fiction or Ellison's original script (which kept to the original story fairly well and updated to include a modern "feel", Asimove was a bit of a romantic in the visual sense).
I'd encourage everyone to look up the I,Robot Ellison script and give it a read. Sorry for not providing a source and I have to admit, it might be difficult to find unless you can dig up a 12 year old copy of Asimov's pulp mag.
Re:After he died... (Score:3, Informative)
Rationale for the First Law... (Score:5, Insightful)
The problem with this reasoning, however, is that it assumes that because the law itself is simply stated, that the definitions of the words it contains are equally simple. That reasoning does not follow logically from the premise. The definition of "harm", for example, is vast... and to restrain human beings from performing in their daily capacity what would otherwise be normal and proper behaviour would arguably be causing _actual_ harm to the people that the robot was caring for. Therefore, the robot must make a decision, based on the overall level of harm that is done in connecction with the probability that the harm would actually happen. Thus, an action that actually induces negative psychological damage (not theoretically, but actually probable damage) would be less preferable to one that may or may not cause real physical damage, especially if the latter would be necessary for performing in their ordinary daily capacity, since denying a human being their freedom and rights of self-determination is inarguably psychologically damaging. The weights of the damages caused must be factored in with the ability for the human beings involved to recover from those damages, and the robot would have to make a choice that would result in the smallest overall level of harm being caused to humans in general, with harm to the general welfare of humanity being weighted in slightly favour to that of any particular human being, so that, for example, a robot could inform the police of a robbery, even though doing that would likely mean that the thief would go through suffering as part of the excercise of justice (that is, his freedoms are revoked, he goes to jail, possibly gets subjected to harsh treatment, etc). This doesn't make it too fuzzy, however... the robot would allow human beings to come to harm only to the extent that it was essential for the human society to continue to function normally simply because to stop society from functioning normally would actually cause much greater long-term harm.
There are similar rationales for the other two laws. Asimov was no dummy.
The 3 laws were DESIGNED to have loopholes (Score:3, Informative)
Just the concept of "human" lead to a great Campbell essay in Analog asking "What do you mean: Human"? And that was in the mid-60s,
It's too bad the film had to chuck the essence of Asimov's imagined world for the simplistic drivel they created.
But action sells tickets to teens who otherwise won't bother with something where you might actually have to think and feel. For me "A. I." was a very fine film that works much better than almost any other S.F. film I've seen, and I've seen a lot even if it did need to have a machine longing to be human.
i'd love to see Benford's "Galactic Center" novels formed into a movie - just for the millieue.
Re:And in other news... (Score:5, Insightful)
Re:And in other news... (Score:5, Insightful)
Re:And in other news... (Score:3, Funny)
Re:And in other news... (Score:5, Informative)
This is the reason why IA came up with the three laws to begin with (Robots always subservient to Humans) and why the Robot/Empire/Foundation universe has no aliens at all (though a later short story in the Empire period had a single alien species, they were busy dying out, and the humans were clearly in ascendancy galaxy wide, and it took a kind human tricking the system to save that species and send them into exile to another galaxy). CAMPBELL! That was his name, CAMPBELL! I think. Not real sure, but it's was connected to at least one of those memes.
Re:And in other news... (Score:4, Interesting)
Re:A dissapointment (Score:5, Funny)
Re:A dissapointment (Score:4, Funny)
Re:A dissapointment (Score:3, Insightful)
Re:A dissapointment (Score:3, Funny)
Re:A dissapointment (Score:3, Insightful)
(Post Script: When typing out this message, at first I accidentally started typing <blackalicious> rather than <blockquote>.)
Re:A dissapointment (Score:5, Insightful)
Comment removed (Score:5, Insightful)
Re:A dissapointment (Score:5, Informative)
*********SPOILERS*********** for anyone who hasn't read Asimov.
Take the Robot novels for example. Caves of Steel, Naked Sun, Robots of Dawn, and Robots and Empire. In all of them except for Dawn, robots work towards the death or destruction of human beings. In Caves, a robot transported the weapon that served in a murder. In Nake sun, a robot with detachable limbs gave its arm to a woman with which she bludgered her husband. In Empire, a Solarian robot tries to kill a human being because her definition of such a being depends on his accent.
All the time, the laws are broken, warped, and shown to be less than perfect. That's not to say that they are useless, only that they have limits and problems.
And I'm not even getting into the matter of the consequences of societies including such robots and the evolution and survivability of such societies.
Re:A dissapointment STILL SPOILERISH (Score:4, Interesting)
Stop implying that he failed.
In Caves, a robot transported the weapon that served in a murder. In Nake sun, a robot with detachable limbs gave its arm to a woman with which she bludgered her husband. In Empire, a Solarian robot tries to kill a human being because her definition of such a being depends on his accent.
None of these are rampaging hordes of killbots like what we see in this movie's trailers. All of these were done in a smart, intelligent, toughtfull, non rampaging hordes of killbots kinda way.
Re:A dissapointment STILL SPOILERISH (Score:4, Informative)
In fact, it's intimately tied in with the three laws, which the plot revolves around and show up prominantly before the title even crawls onto the screen.
Again, it may not be your cup of tea, but I think it was very true to the source material, and could have easily fit in as another story in I, Robot. And it was very entertaining.
Tik-Tok (Score:4, Interesting)
anyone intelligent enough to make a robot would build some failsafe in its programming,
There is a wonderful book (pure satire) set in such a world. It's called Tik-Tok [amazon.com] by John Sladek. However, the central character is a robot that has something go very wrong with his "asimov circuits." The result is a tendancy to murder people and yet no-one in society believes he's capable of it (especially other robots), because they assume he's governed by the three laws.
The book is also one of the funniest and most absurd things I've ever read. If you like your humour black then it might be the perfect antidote to Hollywood's attempt to impart angular momentum to Isac Asimovs mortal remains.
Re:A dissapointment (Score:4, Interesting)
Doubtfull. Nowhere in the preview did I see the word "Asimov." Sure, it might have been in the tiny text that the show and the end of the preview for 1.5 seconds, but I doubt that's going to get anybody into the bookstores that didn't already know of Asimov. You'd think that they'd title it "Isaac Asimov's 'I, Robot'" as a selling point.
I actually bought the book Wednesday, and read it yesterday, primarily because I wanted to read the book before I saw the movie. What I was actually amazed by was how bad the book really was. I have read other stuff by Asimov that I liked a lot.
The characters were pretty one-dimensional, with the most developed one in the whole book, Dr. Susan Calvin, basically amounting to nothing more that a "woman scientist" with nothing more to her than that. Those two field engineers also were pretty annoying, whose only apparent goal in life is to bicker like an old married couple constantly. I thought that the reporter who was "interviewing" Dr. Calvin could have turned into someone interesting, but he wasn't even given a name, definitely not any of the story himself.
The way the robots broke down was stupid as well. A robot supposedly smarter than most humans, running in circles on Mercury, singing Gilbert and Sullivan? Another group of robots who think they are the chorus line? The robotic Muhammad was kind of funny, and actually more along the lines of breakdown I would expect in something so advanced as they are supposed to be, but I am pretty sure I would be offended by it if I were a Muslim.
And then the book degenerates into a Socialist wet-dream, with a robot elected president of the Earth, and all of the economic activities of the Earth dictated by four robotic brains, who "know what is best for us", as they purposefully and selectively destabilize parts of the world economy in order to discredit people who disagree with robotic control, the "Fundies" (is that where the term first showed up?), with the good Dr. Calvin just assuming that it is for the best, because "our entire technical civilization has created more unhappiness and misery then it has removed", therefore we will be happy when we give up control, and it will be for our own good.
The book sucked.
Re:A dissapointment (Score:3, Interesting)
This is the point at which you need to put down the Asimov and pick up the PKDick.
Re:A dissapointment (Score:5, Funny)
Imagine my disappointment when I saw that some guy named JR translated the fine on-screen presentations of Lord of the Rings into trashy paperbacks!
Re:A dissapointment (Score:5, Funny)
And he totally dissed Liv Tyler's character. She should sue. The nerve of some people...
I also heard the novelization of "The Passion of the Christ" came in 4 slightly different versions by different authors. That's just wacky.
Re:A dissapointment (Score:5, Interesting)
Re:A dissapointment (Score:3, Informative)
FWIW, there's no relationship between Harlan Ellison's script and the current I, Robot movie.
Re:A dissapointment (Score:3, Insightful)
Have you seen this movie yet?
If not then how can you make that judgement.
The posted review is far from sufficient to draw the conclusion that this movie is a dumbing down.
Re:A dissapointment (Score:3, Informative)
Robots didn't do "bad things" in Asimov's books. Asimov created his "Three Laws of Robotics" as a plot device to avoid stories about what he called the "Frankenstein Complex", the fear of man's creations going out of control. An Asimov-type "positronic" brain would melt down at even the hint of breaking the First Law. See "Liar", where even emotional damage to humans is enough to destroy the brain, or The Naked Sun where a r
Re:A dissapointment (Score:3, Informative)
I haven't seen the movie but I have read Asimov's robot stories. For the most part, they're about situations in which robot's seem to be breaking the three laws. One of two things generally happen next. Either the robots suddenly become aware of what they are doing and proceed to lock up, or the situation is investigated and whatever they are doing is actually in line with the three laws (usually something like ignoring orders to save human lives in some unexpected way).
From the traile
Re:Inspired by Asimov? (Score:4, Insightful)
Gee, I wonder.
(Hint: BASIC string variable symbol.)
Re:Inspired by Asimov? (Score:3, Insightful)
Re:Um... what? (Score:3, Insightful)
He did touch on this (Score:3, Informative)
Re:3 Laws Unsafe. (Score:3, Informative)
Is it possible to create ethical AI based on the Three Laws? Is it ethical to create ethical AI based on the Three Laws? What other solutions have been proposed for the problem? These questions are explored in our Articles Section. The articles give perspective on why the field of AI ethics is crucial, and why Asimov's Laws are simply its beginning.
And here's a direct link right to the articles! [asimovlaws.com] Wow! Reading is fun!
Re:3 Laws Unsafe. (Score:3, Interesting)
And did you look at the articles?
I read through a few of them, and really, they're pretty worthless. This is from the first article: "One Law to Rule Them All"
There were several directions Asimov didn't go with his robot AIs, such as recursive self-enhancement. Recursive self-enhancement occurs when an AI improves its own intelligence, and then repeats the process - but this time using more intelligence - and repeating again and again, resulting in a mountainous int
Re:I'm sorry but (Score:3, Funny)
Re:Movie (Score:5, Funny)
Scientist: How much time do we have professor?
Frink: Well according to my calculations, the robots won't go berserk for at least 24 hours.
(The robots go berserk.)
Frink: Oh, I forgot to er, carry the one.
The Evil Bit: Mankind's Last Best Hope (Score:3, Funny)
"But what struck me from the trailer is that you can tell when the robots go bad because they glow red. Well, shit. That takes out some subtlety, doesn't it? "Hey man, stay away from the glowing red robots!" Duh. They must be "set to evil".
Hence the critical importance of implementing the Evil Bit in ALL systems, not just Internet communications protocols. Come on people, join the movement! It's our last best hope for humanity!