开源日报 每天推荐一个 GitHub 优质开源项目和一篇精选英文科技或编程文章原文,坚持阅读《开源日报》,保持每日学习的好习惯。
今日推荐开源项目:《来点音乐 awesome-music》
今日推荐英文原文:《A.I. Demilitarisation Won’t Happen》
2019年1月24日:开源日报第322期
今日推荐开源项目:《来点音乐 awesome-music》传送门:GitHub链接
推荐理由:这个项目是一个关于音乐的各种方面的列表——音乐的编辑器,音乐的库,音乐的作曲工具,音乐的理论教学……如果你正好对音乐感兴趣的话,这个项目里肯定会有你感兴趣的东西。不管是想要作曲,学习理论或者是找一些库来用都是做得到的。或者说你也可以在最后找到一些关于音乐的其他列表,比如音频可视化这类的,兴许它们也能满足你的需要。
今日推荐英文原文:《A.I. Demilitarisation Won’t Happen》作者:Josef Bajada
原文链接:https://medium.com/@josef.bajada/a-i-demilitarisation-wont-happen-92bcf6d4bc7d
推荐理由:相信大家都可能在各种科幻电影或者动漫中见过 AI 兵器了,兴许这样的事情会发生在未来

A.I. Demilitarisation Won’t Happen

Artificial Intelligence is already being integrated in next-generation defence systems, and its demilitarisation is highly unlikely. Restricting it from military use is probably anyway not the smartest strategy to pursue.
2019年1月24日:开源日报第322期
Photo by Rostislav Kralik on Public Domain Pictures

This year’s World Economic Forum’s annual meeting is about to start. While browsing through this year’s agenda, I couldn’t help but recall last year’s interview with Google’s CEO, Sundai Pichai, who was quite forthcoming and optimistic about how he envisages A.I. to change the fundamental ways of how humanity operates.

Indeed, the automation of thinking processes, combined with high speed communication networks, instant availability of data, and powerful computation resources, has the potential to affect all mechanisms of society. A.I. will play a key role in the decision making, day-to-day running and automation of health, education, energy provision, law enforcement, economics, and governance, to name just a few.

Like most other innovations, each technology advancement can be misused for evil purposes, and A.I. is no different. What scares most people is its elusive nature, being such a broad field, and playing a central role in so many science fiction plots that don’t end quite well.

In his interview, Sundai tried to quell such concerns by proposing that countries should demilitarise A.I. Of course, one should take such comments in context. But expecting such a thing from all countries is like expecting mankind to stop doing war altogether.

The reason I emphasise all, is because it just takes one country to create an imbalance. What if North Korea decides to integrate A.I. in its ICBMs for high precision autonomous guidance? Is it realistic to expect the U.S. to refrain from doing the same to counteract such threats? It is Mutual Assured Destruction all over again.

In such a situation you can only achieve a nash equilibrium if both parties have more to lose when they adopt a more aggressive dominant strategy and the opponent retaliates. One would expect the CEO of one of the world’s largest A.I. focused companies to be at least aware of such a fundamental concept of Game Theory.
2019年1月24日:开源日报第322期
Photo Credits Wikimedia Commons

But of course Sundai is fully aware of this. His utopic vision of a harmonious inter-country peaceful collaboration for the application of A.I. is just playing to the tune of his employer’s agenda. Ironically, this took place at the start of the same year when its “don’t do evil” motto was removed, its own employees walked out in protest over its participation in military projects, and people started to get seriously concerned about the availability and misuse of their private information such as emails and mobile phone data.

Maintaining a balance of power on opposing sides is already a good enough reason not to impose restraints on one side, without any concrete guarantees that the other side is actually doing the same. But here are some more reasons why A.I. demilitarisation will not happen, and why we probably do not really want it to happen in the first place.

Funding

Defence budgets drive a lot of research and innovation in western countries where having a strong military capability is considered critical. They don’t just fund the construction of missiles and stealth fighters. They support the development of ideas that often do not yet have any commercial business model yet, but are still worth having before competing nations do. Of course it is more likely that ideas with clear military application get funded from defence budgets. But technologies such as the Internet (originally ARPANET) and the Global Positioning System started as defence projects and eventually got released to the public.

Autonomous vehicles may have become more popular due to Google’s self-driving car project (now Waymo), but in actual fact this was the result of a grand challenge by the US Defense Advance Research Projects Agency (DARPA). Google then recruited people from the team that won the 2005 challenge. DARPA also held a robotics challenge, with the objective of developing humanoid robots that are capable of performing dangerous tasks in environments that are too harmful for humans. One only needs to recall the Chernobyl and Fukushima disasters, and the consequences of people working in radioactive environments to deal with their aftermath, to realise how useful such a technology could be.
2019年1月24日:开源日报第322期
JPL’s RoboSimian at the DARPA Robotics Challenge

DARPA takes A.I. very seriously, and similarly other defence organisations such as BAE Systems and Thales. Expecting the military to fold its hand and distance itself from A.I. is plain unrealistic.

A.I. Can Reduce Casualties (on both sides)

One of the main concerns about the use of A.I. in the military is the development of killer robots. This is a legitimate concern that needs to be addressed through the right policies and treaties, although I doubt that if the technology components were available, any country would publicly admit to having autonomous weapons (until it is constrained to bring them out and use them). Furthermore, if it becomes possible for one nation to obtain such a technology, it is probably a very bad idea to leave the playing field unbalanced.

Given that Artificial General Intelligence is still very far off, and a “Terminator-style” killer robot is unlikely, maybe we should focus more on the immediate benefits of A.I., such as Battlefield Intelligence. What if a soldier’s visor was equipped with augmented reality that uses A.I. to identify friend from foe, or civilian from threat, more accurately than an adrenaline pumped human fearing for his life? Autonomous armoured vehicles can be used for high risk missions, like delivering supplies through a hostile route, or extracting casualties from an active combat zone. Robots could accompany soldiers and assist them in investigating threats or carrying heavy equipment.
2019年1月24日:开源日报第322期
Photo by Sgt. Eric Keenan

What if missile guidance systems were equipped with on-board algorithms that verified their targets before detonation? “Surgical strikes” would just hit their intended targets, instead of causing collateral damage and innocent civilian deaths. What about smarter air defence systems that make use of A.I. to track and intercept incoming airborne threats, or even detect them before they are actually launched?

A.I. is not Well Defined

Most A.I. techniques involve crunching data streams, applying mathematical techniques such as calculus, probability and statistics, and putting them together using a handful of algorithms to arrive to a solution. There is nothing really cognitive, sentient or self-aware going on.

Unlike nuclear weapons, where you can clearly determine whether a country has stockpiles or not (assuming you can find them), there is no clear way to classify the software technology that is integrated in the various defence systems. Furthermore, one can’t really expect the military to refrain from using the data and computing resources it has available to obtain faster, smarter, and machine-verified decision support, while the general public has it at their finger tips. A.I. demilitarisation is just not happening.
下载开源日报APP:https://os.maxoyed.com/2579/
加入我们:https://os.maxoyed.com/about/join/
关注我们:https://os.maxoyed.com/about/love/