Immediate update: Oops! Just came across this: Robot grabs factory worker and crushes him to death
Read these two posts together, and the frisson between them feels inevitable. Keep your own mind! (Or uncover it, if you are still programmed!)
BTW: A.I. monsters don’t have to look like this.
Instead, they might look like this (supposed image of Blue Avian):
And/or, they could just be ubiquitous, invisible computer functions, as iconified by “HAL,” in the movie 2001: A Space Odyssey.
The fact that several people claim to have been in communication with the Blue Avians doesn’t mean they’re real, necessarily—just that the Illuminati could be targeting multiple people with their program—perhaps for testing purposes.
I think it would be closed-minded and possibly even dangerous to suggest that when we have no proof, that the Blue Avians must be biological, sentient extraterrestrial and benevolent entities.
Believing everything Corey Goode says because we WANT to believe it, is a highly questionable decision in my view.
It’s clear from the new video that came out this month about the patents to this brain mapping and manipulation technology, that they are preparing to target individual people’s brains—interpreting, sending and receiving information and thoughts.
If they roll out this technology and implant us, it is going to not only strip all privacy from Humanity, but remove our free will. Period. The NSA spying has nothing on this new technology.
Over 1,000 high-profile artificial intelligence experts and leading researchers have signed an open letter warning of a “military artificial intelligence arms race” and calling for a ban on “offensive autonomous weapons”.
The letter, presented at the International Joint Conference on Artificial Intelligence in Buenos Aires, Argentina, was signed by Tesla’s Elon Musk, Apple co-founder Steve Wozniak, Google DeepMind chief executive Demis Hassabis and professor Stephen Hawking along with 1,000 AI and robotics researchers.
The letter states: “AI technology has reached a point where the deployment of [autonomous weapons] is – practically if not legally – feasible within years, not decades, and the stakes are high: autonomous weapons have been described as the third revolution in warfare, after gunpowder and nuclear arms.”
The authors argue that AI can be used to make the battlefield a safer place for military personnel, but that offensive weapons that operate on their own would lower the threshold of going to battle and result in greater loss of human life.
Should one military power start developing systems capable of selecting targets and operating autonomously without direct human control, it would start an arms race similar to the one for the atom bomb, the authors argue.Unlike nuclear weapons, however, AI requires no specific hard-to-create materials and will be difficult to monitor.
“The endpoint of this technological trajectory is obvious: autonomous weapons will become the Kalashnikovs of tomorrow. The key question for humanity today is whether to start a global AI arms race or to prevent it from starting,” said the authors.
Toby Walsh, professor of AI at the University of New South Wales said: “We need to make a decision today that will shape our future and determine whether we follow a path of good. We support the call by a number of different humanitarian organisations for a UN ban on offensive autonomous weapons, similar to the recent ban on blinding lasers.”
Musk and Hawking have warned that AI is “our biggest existential threat” and that the development of full AI could “spell the end of the human race”. But others, including Wozniak have recently changed their minds on AI, with the Apple co-founder saying that robots would be good for humans, making them like the “family pet and taken care of all the time”.
At a UN conference in Geneva in April discussing the future of weaponry, including so-called “killer robots”, the UK opposed a ban on the development of autonomous weapons, despite calls from various pressure groups, including the Campaign to Stop Killer Robots.