Image credit: Photo from Autodesk’s blog post
Our bi-weekly quick take on a bunch of the most important recent media stories about AI for the period November 5th 2018 to November 19th 2018.
Amazon is hiring 100,000 temporary workers this holiday season, 20,000 fewer than last year. The decrease may be due to Amazon’s increasing use of robots in its fulfillment centers. As the article quotes however, Amazon has hired more full-time fulfillment center employees this year, so it’s unclear to what extent the decrease in temp workers is actually due to automation, or perhaps even Amazon’s new minimum $15 per hour wage.
Holger Hansen, Reuters
Germany is planning to invest 3 billion euros in AI R&D until 2025, as outlined in an new “AI made in Germany” strategy paper. Governments around the world are paying more attention to the development of AI, noting its broad applications in many sectors across industry, healthcare, and the military. While the U.S. has yet to draft a national AI strategy, China released its own in 2017 and so did France in early 2018. An AI strategy is especially urgent for Germany as its car industry, an important engine for German economic growth, is falling behind in the development of self-driving (and electric) vehicles to their counterparts in the U.S. and China.
Loren Grush, The Verge
A refreshingly non neural-net based story about the uses of AI. NASA’s Jet Propulsion Laboratory (JPL) partnered with Autodesk after the latter showcased an impressive technique to design lighter spacecrafts:
“To make the lander, Autodesk used its own artificial intelligence software, which can develop hundreds of different designs in short periods of time. Known as generative design, it’s a technique that allows engineers to come up with computer-generated concepts for a project by inputting a set of constraints that the software must adhere to. For the lander, Autodesk and JPL input the types of temperatures and forces a lander might experience when traveling through deep space. They also input variables like the kinds of materials that the software should experiment with, such as titanium and aluminum. And they asked the software to explore different types of manufacturing methods, including casting and 3D printing.”
Joshua Rothman, The New Yorker
Image forgery has a long history. Will algorithmically assisted fakes change anything to our media landscape? The New Yorker delivers a level-headed analysis of what will probably change, in good and in bad.
Alex Hern, The Guardian
When DeepMind began its work with UK’s National Health Service (NHS) in 2017, gaining access to the NHS records of 1.6 million patients, DeepMind pledged to not share this data with Google, a sibling company under Alphabet. The recent restructure to move Streams, DeepMind’s flagship health app, under Google puts this pledge in question and rightfully raises concerns about patient privacy and inappropriate access to this data for the use by other products in Google. For now however, Google insists all data access guidelines and restrictions remain the same, and that this restructure is only for the purpose of scaling Streams globally beyond the UK.
James Vincent, The Verge
Following recent and controversial developments in machine-generated human voices and facial expressions (see DeepFakes), Xinhua, China’s state-run media outlet, has deployed an “AI” news anchor to read news. This anchor isn’t an “AI” but rather a synthesized video of a real-life anchor reading scripts written by a human. While the “AI” anchor is not completely realistic and can be easily distinguished by a human to be fake, the dissemination of this technology in Chinese media can be disconcerting to many, given the country’s records on press censorship.
Stephen Chen, South China Morning Post
“The 27 boys and four girls, all aged 18 and under, were selected for the four-year “experimental programme for intelligent weapons systems” at the Beijing Institute of Technology (BIT) from more than 5,000 candidates, the school said on its website.”
Chinese high school graduates have been selected highly competitive college program to work on AI-enabled weaponry. This event continues the trend of AI expertise becoming relevant to the national defense strategies of big countries across the globe.
Karen Hao, MIT Technology Review
The Trolley Problem is one discussed in philosophy and ethics classes for years. Now, when AI agents are the ones making decisions for us like in the case of self driving cars, these decisions become even more important. Researchers at MIT Media Lab designed an experiment to present various versions of these questions in a game like interface called the Moral Machine which received viral reception. On analysing these results, they observed that results vary wildly based on the country but they also correlate highly with culture and economics.
Sheer number of people was surprisingly not the only dominant factor in choosing which group should be spared. The choices of participants from individualistic countries differ greatly when compared to participants from collectivist countries. The researchers discuss various possible biases and how we need to discuss these issues now more than ever.
The authors of the study emphasized that the results are not meant to dictate how different countries should act. In fact, in some cases, the authors felt that technologists and policymakers should override the collective public opinion. The results should be used by industry and government as a foundation for understanding how the public would react to the ethics of different design and policy decisions.
Brad Smith, Microsoft Blog
The article highlights how Artificial Intelligence has embedded in our professional and personal lives today and presents both negative and positive possible implications of it. One such technology is facial recognition which on one hand, could help locate a missing child and on the other, invade the privacy of millions of people by tracking them at all times. It discusses the role of government in regulating these technologies and the responsibility of the tech sector to put effort into mitigating bias and developing ethical products.
Irving Wladawsky-Berger, The Wall Street Journal
“Artificial intelligence has the potential to incrementally add 16 percent or around $13 trillion by 2030 to current global economic output– an annual average contribution to productivity growth of about 1.2 percent between now and 2030, according to a September, 2018 report by the McKinsey Global Institute on the impact of AI on the world economy.”
A nice summary of recent projections about the worldwide economic impacts of AI over the next decade. The summary highlights key takeaways, the biggest of which are that AI will have a relatively large impact on the world economy but that it is also widen inequality among those with a starting lead and those who do not yet have much expertise in AI:
“The economic impact of AI is likely to be large, comparing well with other general-purpose technologies in history,” notes the report in conclusion. “At the same time, there is a risk that a widening AI divide could open up between those who move quickly to embrace these technologies and those who do not adopt them, and between workers who have the skills that match demand in the AI era and those who don’t. The benefits of AI are likely to be distributed unequally, and if the development and deployment of these technologies are not handled effectively, inequality could deepen, fueling conflict within societies.”
Melanie Mitchell, The New York Times
In this opinion piece, Pr. Mitchell points out that despite recent advances in fields like image recognition or machine translation, researchers have not yet solved the issue of making algorithms that understand the task that they are solving: this is why machine translation regularly gives nonsensical results, and why many algorithms cannot deal with trivial modifications in the data they are processing.
“The bareheaded man needed a hat” is transcribed by my phone’s speech-recognition program as “The bear headed man needed a hat.” Google Translate renders “I put the pig in the pen” into French as “Je mets le cochon dans le stylo” (mistranslating “pen” in the sense of a writing instrument).
Sebastian Ruder, AYLIEN blog
Sebastian Ruder succinctly provides the main trends and interesting research published in this year’s EMNLP, one of the most prominent conferences in natural language processing. He categorizes these trends into inductive bias, cross-lingual learning, word embeddings, latent variable models, language models, datasets, miscellaneous and shares insight to why these papers stand out. A must read for anyone looking to catch up on the latest NLP research.
Rachel Metz, CNN Business
CNN Business interviewed Janelle Shane, a scientist who has been delighting the internet by sharing the strange outputs of her neural networks trained to reproduce data like ice cream flavors, cooking recipes and halloween costumes. The results make for a good crash course in neural networks’ limitations.
Malay Haldar, Airbnb Engineering & Data Science
An introduction to the problems worked on in Airbnb’s recent Arxiv paper, in which they described methods for
Not quite a full curriculum, but a nice set of resources for people interested in teaching themselves the key concepts in Reinforcement Learning research.
That’s all for this digest! If you are not subscribed and liked this, feel free to subscribe below!