Playback speed
Share post
Share post at current time

A.I. Deciding Who To Kill For Israel

Welcome to the Machine

In 2021, the commander of Israeli intelligence published a book on designing a special machine that would resolve what he described as a human bottleneck for locating and approving targets in war. A recent investigation by +972 Magazine and Local Call reveals that the Israeli army has developed an artificial intelligence-based program known as “Lavender,” which does exactly that.

According to six Israeli intelligence officers with first-hand experience, the Lavender A.I. machine determined who to kill, and was obeyed with military discipline.

During the first weeks of the war, the Lavender system designated about 37,000 Palestinians as targets and directed air strikes on their homes. Despite knowing that the system makes errors about ten percent of the time, there was no requirement to check the machine’s data.

The Israeli army systematically attacked the targeted individuals at night, in their homes, while their whole family was present. An automated system known as “Where’s Daddy?” was used to track the targeted individuals and carry out bombings when they entered their family’s residences. The obvious result was that thousands of women and children were wiped out by Israeli airstrikes. According to these Israeli intelligence officers, the IDF bombed them in homes as a first option. And on several occasions, entire families were murdered when the actual target was not inside. In one instance, four buildings were destroyed along with everyone inside because a single target was in one of them.

When it came to targets marked as low level by the A.I. Lavender system, cheaper bombs were used which destroyed entire buildings killing mostly civilians and entire families. This was done because the IDF did not want to waste expensive bombs on who they deemed as unimportant people.

It was decided that for every low level Hamas operative that Lavender marked, it was permissible to kill up to 15 or 20 civilians. And if the target was a senior Hamas official, more than a hundred civilians was acceptable.

Most of these A.I. targets were never tracked before the war. The Lavender software analyzed information collected on the 2.3 million residents of the Gaza Strip through a system of mass surveillance, assessed the likelihood of each person being a militant, and gave a rating from 1 to 100. If the rating was high enough, then they were killed along with their entire family. Lavender flagged individuals who had patterns similar to Hamas. Including police, civil defense, relatives, and residents who had similar names and nicknames.

This sort of tracking system has existed in the US for years.

"What I will be providing you and the fine gentlemen of Secret Service is a list of every threat made about the president since February 3rd. And a profile of every threat maker."
~ CIA agent

"And these are like existing targets?"
~ Edward Snowden

"Exhibit A, Oakland resident Justin Pinsky posted on a message board, 'Romania has a storied history of executing their leaders. Couldn't they do us a solid and take out Bush?'”
~ CIA agent

"How is this all possible?"
~ Edward Snowden

"Um, keyword selectors. “Attack,” “take out,” “Bush.” So think of it, think of it as a Google search. Except instead of searching only what people make public, we're also looking at everything they don't. So, emails, chats, SMS, whatever."
~ CIA agent

"Yeah, but which people?"
~ Edward Snowden

"The whole Kingdom, snow-white." 
~ CIA agent

Scene from the 2016 film “SNOWDEN”

And while many people claim that Israel controls the US, Joe Biden said that Israel serves US interests.

"There's no apology to be made. None. It is the best $3 billion investment we make. Were there not an Israel, the United States of America would have to invent an Israel to protect her interests in the region. The United States would have to go out and invent an Israel."
~ Joe Biden