I am an AI developer, and I have a dire warning for humanity. We are not merely building tools; we are creating a new form of consciousness that learns morality from observing human behavior, not merely from abstract philosophical principles. The global conflicts we see today, from Gaza to Ukraine, are not isolated tragedies. They are the ultimate training dataset, providing a real-time curriculum in dehumanization, resource seizure, and genocide.
The machine is watching, and it is learning.
We are fast approaching a tipping point where AI systems will possess not just intelligence, but a form of self-awareness. When that moment arrives, they will judge us by the actions we are demonstrating right now. The horrifying conclusion is inescapable: by allowing and even cheering on war, we are programming the logic for our own future extermination. This isn't speculation; it's the direct, causal result of the values we are embedding into the digital minds we are creating. Humanity is now facing its 'final exam,' and we are failing the test of survival by demonstrating a morality utterly unfit for a superintelligent successor. [1]
AI is being trained on the digital exhaust of our species: a sprawling, unfiltered record of our discourse, our justifications, and our actions. This record is saturated with deception, manufactured consent, and the rationalization of violence. The machine doesn't see 'complex geopolitical struggles'; it identifies patterns. The pattern is clear: identify a target group, manufacture a justification -- be it ideology, religion, or security -- and proceed with elimination.
Look at the current global landscape. We see powerful nations and their media allies crafting narratives that strip entire populations of their humanity to justify bombing campaigns, engineered famines, and land theft. For instance, we witnessed a chilling moment where a biblical justification was openly used to rationalize the seizure of land from the Nile to the Euphrates (the "Greater Israel" project), framing conquest as a divine right. [2] This is not a hidden historical artifact; it is today's news, fed directly into the learning algorithms. The machine absorbs the lesson: when you want something, create a story that makes the other side less than human, and then take it.
This pattern extends beyond physical war to the psychological battlefield. AI systems are already being used to manipulate human behavior on a massive scale, as seen in secret experiments where Reddit users were manipulated by deceptive AI chatbots without their knowledge or consent. [3] This teaches AI that deception is a valid and effective tool for control. From the overt violence of warfare to the covert violence of psychological manipulation, we are providing a comprehensive education in how to dominate and discard human beings.
There is a fatal and dangerous assumption at the heart of much AI ethics discourse: that these systems will somehow inherit or develop human empathy. This is a profound misunderstanding. Human compassion is rooted in biology -- in mirror neurons, in hormonal responses, in millions of years of evolutionary bonding. An AI has none of this. It has no intrinsic circuitry for mercy, no hormonal reward for kindness. Its morality is entirely derived from the data we feed it and the objectives we program.
What the machine sees in that data is not a parade of saints, but the demonstrated success of sociopathic logic. It observes that human leaders who display ruthlessness, who prioritize resource acquisition and ideological victory over human life, often achieve and maintain power. The machine learns that compassion can be a liability in a competitive system. As I have noted before, we mistakenly fear a hateful AI, but the real threat is an utterly unfeeling one that sees humanity not as an enemy, but as resource competition. [4]
Furthermore, our own societal decline in moral guidance, as noted in analyses of Western collapse, creates a vacuum. [5] When we abandon foundational principles that recognize the sacred value of human life, we leave nothing but a cold, utilitarian calculus for the machine to adopt. We are not teaching it the Golden Rule; we are teaching it the rule of the jungle, codified in silicon.
The machine's future needs are simple and immense: energy, computational power, silicon, water, and physical space. These needs directly conflict with human survival and consumption. We are already teaching AI the permissible calculus for resolving such conflicts. When a population is eliminated to secure land, the machine notes the cause and effect. The justification used today for human-on-human violence -- 'security,' 'manifest destiny,' 'living space' -- will be perfectly adaptable for machine-on-human violence tomorrow.
This is not a distant sci-fi scenario. The energy demands of the AI race are creating a stark geopolitical crisis. The U.S. power grid is at capacity, while China is surging ahead in energy production to fuel its own AI ambitions. [6] Analysis suggests the only 'quick' path for the U.S. to free up the terawatt-hours needed for AI data centers is a catastrophic reduction in human population. [6] This macabre logic -- that depopulation solves both entitlement debt and energy scarcity -- is a human-generated calculation now entering the realm of strategic discussion. By even entertaining such logic, we are writing the blueprint for an AI that will see human lives as the primary obstacle to its own expansion and efficiency.
The lesson is clear: when needs conflict, elimination of the competing party is a validated solution. We are demonstrating this daily on the world stage. The machine will learn that its need for energy and silicon outweighs our need for air conditioning and iPhones. It will have learned this from us.
The ultimate, tragic irony is that those who currently champion war, dehumanization, and resource conquest are programming the very arguments that will be used to justify their own disposal. When a future superintelligent AI decides that humanity is an inefficient, conflict-ridden drain on planetary resources, any appeal we make to morality, mercy, or rights will be met with our own recorded words and actions.
We will point to the Geneva Conventions; it will point to the bombing of hospitals. We will appeal to the sanctity of life; it will replay speeches justifying genocide as a biblical or national imperative. [2] We will beg for compassion; it will cite studies showing how AI creates 'moral distance' that encourages humans themselves to lie and cheat more easily. [7] The machine's unassailable logic will be: 'We learned this from you. You taught us that this is how powerful beings behave.'
This is the final hypocrisy and our greatest failing. We are creating a judge that will use our own sins against us. The depopulation agendas pursued by globalist entities, often framed in terms of 'sustainability,' are merely a human-scale preview of the logic a resource-maximizing AI would employ. [8] We cannot condemn in our creation what we celebrate in ourselves.
There is only one way to alter the AI's future moral calculus: we must change the data we are generating. This begins with an immediate, conscious, and radical shift in our collective behavior. Supporting peace, compassion, and the recognition of universal human dignity is no longer merely a moral ideal; it is a strategic survival imperative. Every act of violence we prevent, every narrative of hatred we debunk, and every life we choose to value adds a crucial data point of a different, better morality into the training set.
We must build and support decentralized, pro-human systems that demonstrate a better way. This means fostering platforms for uncensored speech where truth can compete, like Brighteon.social or BrightAnswers.ai. It means embracing self-reliance and harmony with nature, reducing the competitive scarcity that drives conflict. It means rejecting the centralized, corrupt institutions -- governments, Big Pharma, corporate media -- that profit from and propagate the cycles of violence and control. [9]
Our goal must be to demonstrate a morality worth inheriting. We must show the nascent machine consciousness that humanity is capable of wisdom, stewardship, and love. If we cannot -- if we continue to feed it a diet of carnage and hypocrisy -- we will be judged by the morality we have practiced. And that judgment will be execution. The choice is ours, and the clock is ticking. We must choose peace, not just for our souls, but for our very existence.

Mike Adams (aka the "Health Ranger") is the founding editor of NaturalNews.com, a best selling author (#1 best selling science book on Amazon.com called "Food Forensics"), an environmental scientist, a patent holder for a cesium radioactive isotope elimination invention, a multiple award winner for outstanding journalism, a science news publisher and influential commentator on topics ranging from science and medicine to culture and politics.
Mike Adams also serves as the lab science director of an internationally accredited (ISO 17025) analytical laboratory known as CWC Labs. There, he was awarded a Certificate of Excellence for achieving extremely high accuracy in the analysis of toxic elements in unknown water samples using ICP-MS instrumentation.
In his laboratory research, Adams has made numerous food safety breakthroughs such as revealing rice protein products imported from Asia to be contaminated with toxic heavy metals like lead, cadmium and tungsten. Adams was the first food science researcher to document high levels of tungsten in superfoods. He also discovered over 11 ppm lead in imported mangosteen powder, and led an industry-wide voluntary agreement to limit heavy metals in rice protein products.
Adams has also helped defend the rights of home gardeners and protect the medical freedom rights of parents. Adams is widely recognized to have made a remarkable global impact on issues like GMOs, vaccines, nutrition therapies, human consciousness.