Draft:Lavender (artificial intelligence)
Submission declined on 13 May 2024 by S0091 (talk). This submission is not adequately supported by reliable sources. Reliable sources are required so that information can be verified. If you need help with referencing, please see Referencing for beginners and Citing sources. This draft's references do not show that the subject qualifies for a Wikipedia article. In summary, the draft needs multiple published sources that are:
Where to get help
How to improve a draft
You can also browse Wikipedia:Featured articles and Wikipedia:Good articles to find examples of Wikipedia's best writing on topics similar to your proposed article. Improving your odds of a speedy review To improve your odds of a faster review, tag your draft with relevant WikiProject tags using the button below. This will let reviewers know a new draft has been submitted in their area of interest. For instance, if you wrote about a female astronomer, you would want to add the Biography, Astronomy, and Women scientists tags. Editor resources
|
- Comment: The Guardian article almost entirely is what those affiliated say so is a primary source. 972 is a little better but not enough to establish notability. Also, both were published on the same day which does not support sustained coverage and most of the content is unsourced. S0091 (talk) 17:03, 13 May 2024 (UTC)
Lavender is a human targeting AI system developed by the Israel Defense Forces' (IDF) elite intelligence division Unit 8200. First hypothesized in 2021 in a book titled "The Human-Machine Team: How to Create Synergy Between Human and Artificial Intelligence That Will Revolutionize Our World," Lavender is the first known implementation of its kind to actively employ artificial intelligence in identifying and approving military bombing targets.[1][2]
First deployed during the 2021 Israel–Palestine crisis, Lavender has since been relied on by the IDF at least during the first few weeks of the Israel-Hamas War. To date, Lavender is responsible for potentially ordering the bombing of at least 37,000 suspected militant targets in Gaza, comparable to a local health ministry's estimation that 33,000 Palestinians had been killed since the beginning of the conflict six months prior.
While the known role of Lavender is to broadly identify human targets in a manner similar to a hit list, Lavender works in conjunction with two other AI systems playing an active role in identifying targets in the war: Where's Daddy, which tracks the given human targets in order to signal an airstrike after they enter their home, and The Gospel, which targets militant-occupied buildings rather than the militants themselves.[1]
Operation
As an AI system, Lavender operates as a typical classification machine learning system. It is trained on data of known Hamas and Palestinian Islamic Jihad (PIJ) operatives, which allows it to potentially classify the likelihood that a given human is an operative. This produces a likelihood ranking of humans based on similarity to the training data, which is then passed to intelligence officers for human evaluation. Approved targets are then tracked by Where's Daddy, and an airstrike is called once the target enters their home.
According to testimony from six intelligence officers from the IDF, the human evaluation step was not observed in practice during the early stages of the Israel-Hamas War.[1] After checking that the target is male, IDF officers typically trusted Lavender's statistical judgements in order to save time while absolving personal responsibility by treating false positives as a statistical error.
Although it is estimated that 90% of the predictions made by Lavender are valid military targets by the IDF's definition, the remaining 10% might include people with similar names as Hamas operatives, or people who communicate in the same style as them.[2] This has led to an unusually large proportion of civilians being targeted by Lavender, including aid workers.
See also
- Lethal autonomous weapon – Autonomous military technology system
- Israel-Hamas war – Ongoing armed conflict in the Middle East
- Unit 8200 – Intelligence unit of the Israel Defense Forces
References
- ^ a b c Abraham, Yuval (3 April 2024). "'Lavender': The AI machine directing Israel's bombing spree in Gaza". 972 Magazine. Retrieved 12 May 2024.
- ^ a b McKernan, Bethan (3 April 2024). "'The machine did it coldly': Israel used AI to identify 37,000 Hamas targets". The Guardian. Retrieved 12 May 2024.