Over the past month, we’ve published a host of articles about artificial intelligence as part of our fourth installment of Futurography—a project from Future Tense in which we explore a different technological issue each month. We’ve seen experts worrying about autonomous weapons, talked with researchers about the complexities of value alignment, looked into the vagaries of sentience, and much more.
With all that behind us, we’re most interested in what you think. Where do these conversations leave us? Where do we go from here?
Come back next month for a roundup of your responses. And then stick around for the start of our fifth Futurography unit, which asks whether drones have to be so creepy.
This article is part of the artificial intelligence installment of Futurography, a series in which Future Tense introduces readers to the technologies that will define tomorrow. Each month from January through June 2016, we’ll choose a new technology and break it down. Read more from Futurography on artificial intelligence:
- What’s the Deal With Artificial Intelligence Killing Humans?
- Your Artificial Intelligence Cheat Sheet
- Killer Robots on the Battlefield
- The Wrong Cognitive Measuring Stick
- The Challenge of Determining Whether an A.I. Is Sentient
- An interview with A.I. expert Stuart Russell
- Why You Can’t Teach Human Values to Artificial Intelligence
- Let Artificial Intelligence Evolve
- Mika Model, a brand-new short story from sci-fi great Paolo Bacigalupi
- When a Robot Kills, Is It Murder or Product Liability?
- The Threats That Artificial Intelligence Researchers Actually Worry About