"Within thirty years, we will have the technological means to create superhuman intelligence. Shortly after, the human era will be ended. Is such progress avoidable? If not to be avoided, can events be guided so that we may survive? These questions are investigated. Some possible answers (and some further dangers) are presented."
Here's the full essay:
Fascinating stuff! I hadn't read Vinge's essay before. Also read "Inside the unfathomable superhuman future after the 'singularity,'" by Bruce Sterling. A thought-provoking response to Vinge. Let's hope this refutation helps develop SF as a whole.