https://www.nytimes.com/2019/09/26/technology/ai-computer-expense.html
At Tech’s Leading Edge, Worry About a Concentration of Power
By Steve Lohr
Sept. 26, 2019, 3:00 a.m. ET
Each big step of progress in computing — from mainframe to personal computer to internet to smartphone — has opened opportunities for more people to invent on the digital frontier.
But there is growing concern that trend is being reversed at tech’s new leading edge, artificial intelligence.
Computer scientists say A.I. research is becoming increasingly expensive, requiring complex calculations done by giant data centers, leaving fewer people with easy access to the computing firepower necessary to develop the technology behind futuristic products like self-driving cars or digital assistants that can see, talk and reason.
The danger, they say, is that pioneering artificial intelligence research will be a field of haves and have-nots. And the haves will be mainly a few big tech companies like Google, Microsoft, Amazon and Facebook, which each spend billions a year building out their data centers.
Interested in All Things Tech?
The Bits newsletter will keep you updated on the latest from Silicon Valley and the technology industry.
In the have-not camp, they warn, will be university labs, which have traditionally been a wellspring of innovations that eventually power new products and services.
“The huge computing resources these companies have pose a threat — the universities cannot compete,” said Craig Knoblock, executive director of the Information Sciences Institute, a research lab at the University of Southern California.
The research scientists’ warnings come amid rising concern about the power of the big tech companies. Most of the focus has been on the current generation of technology — search, online advertising, social media and e-commerce. But the scientists are worried about a barrier to exploring the technological future, when that requires staggering amounts of computing.
The modern data centers of the big tech companies are sprawling and secretive. The buildings are the size of a football field, or larger, housing rack upon rack with hundreds of thousands of computers. The doors are bulletproof. The walls are fireproof. Outsiders are rarely allowed in.
These are the engine rooms of cloud computing. They help deliver a cornucopia of entertainment and information to smartphones and laptops, and they enable millions of developers to write cloud-based software applications.
But artificial intelligence researchers, outside the big tech companies, see a worrying trend in their field. A recent report from the Allen Institute for Artificial Intelligence observed that the volume of calculations needed to be a leader in A.I. tasks like language understanding, game playing and common-sense reasoning has soared an estimated 300,000 times in the last six years.
All that computing fuel is needed to turbocharge so-called deep-learning software models, whose performance improves with more calculations and more data. Deep learning has been the primary driver of A.I. breakthroughs in recent years.
“When it’s successful, there is a huge benefit,” said Oren Etzioni, chief executive of the Allen Institute, founded in 2014 by Paul Allen, the billionaire co-founder of Microsoft. “But the cost of doing research is getting exponentially higher. As a society and an economy, we suffer if there are only a handful of places where you can be on the cutting edge.”
The evolution of one artificial intelligence lab, OpenAI, shows the changing economics, as well as the promise of deep-learning A.I. technology.
Founded in 2015, with backing from Elon Musk, OpenAI began as a nonprofit research lab. Its ambition was to develop technology at the frontier of artificial intelligence and share the benefits with the wider world. It was a vision that suggested the computing tradition of an inspired programmer, working alone on a laptop, coming up with a big idea.
This spring, OpenAI used its technology to defeat the world champion team of human players at a complex video game called Dota 2. Its software learned the game by constant trial and error over months, the equivalent of more than 45,000 years of game play.
The OpenAI scientists have realized they are engaged in an endeavor more like particle physics or weather simulation, fields demanding huge computing resources. Winning at Dota 2, for example, required spending millions of dollars renting access to tens of thousands of computer chips inside the cloud computing data centers run by companies like Google and Microsoft.
Image“As a society and an economy, we suffer if there are only a handful of places where you can be on the cutting edge,” said Oren Etzioni, the chief executive of the Allen Institute.
“As a society and an economy, we suffer if there are only a handful of places where you can be on the cutting edge,” said Oren Etzioni, the chief executive of the Allen Institute.CreditKyle Johnson for The New York Times
Earlier this year, OpenAI morphed into a for-profit company to attract financing and, in July, announced that Microsoft was making a $1 billion investment. Most of the money, OpenAI said, would be spent on the computing power it needed to pursue its goals, which still include widely sharing the benefits of A.I., after paying off their investors.
As part of OpenAI’s agreement with Microsoft, the software giant will eventually become the lab’s sole source of computing.
“If you don’t have enough compute,you can’t make a breakthrough,” said Ilya Sutskever, chief scientist of OpenAI.
Academics are also raising concerns about the power consumed by advanced A.I. software. Training a large, deep-learning model can generate the same carbon footprint as the lifetime of five American cars, including gas, three computer scientists at the University of Massachusetts, Amherst, estimated in a recent research paper. (The big tech companies say they buy as much renewable energy as they can, reducing the environmental impact of their data centers.)
Mr. Etzioni and his co-authors at the Allen Institute say that perhaps both concerns — about power use and the cost of computing — could be at least partially addressed by changing how success in A.I. technology is measured.
The field’s single-minded focus on accuracy, they say, skews research along too narrow a path.
Efficiency should also be considered. They suggest that researchers report the “computational price tag” for achieving a result in a project as well.
Since their “Green A.I.” paper was published in July, their message has resonated with many in the research community.
Henry Kautz, a professor of computer science at the University of Rochester, noted that accuracy is “really only one dimension we care about in theory and in practice.” Others, he said, include how much energy is used, how much data is required and how much skilled human effort is needed for A.I. technology to work.
A more multidimensional view, Mr. Kautz added, could help level the playing field between academic researchers and computer scientists at the big tech companies, if research projects relied less on raw computing firepower.
Big tech companies are pursuing greater efficiency in their data centers and their artificial intelligence software, which they say will make computing power more available to the outside developers and academics.
John Platt, a distinguished scientist in Google’s artificial intelligence division, points to its recent development of deep-learning models, EfficientNets, which are 10 times smaller and faster than conventional ones. “That democratizes use,” he said. “We want these models to be trainable and accessible by as many people as possible.”
The big tech companies have given universities many millions over the years in grants and donations, but some computer scientists say they should do more to close the gap between the A.I. research haves and have-nots. Today, they say, the relationship that tech giants have to universities is largely as a buyer, hiring away professors, graduate students and even undergraduates.
The companies would be wise to also provide substantial support for academic research including much greater access to their wealth of computing — so the competition for ideas and breakthroughs extends beyond corporate walls, said Ed Lazowska, a professor at the University of Washington.
A more supportive relationship, Mr. Lazowska argues, would be in their corporate self-interest. Otherwise, he said, “We’ll see a significant dilution of the ability of the academic community to produce the next generation of computer scientists who will power these companies.”
At the Allen Institute in Seattle, Mr. Etzioni said, the team will pursue techniques to improve the efficiency of artificial intelligence technology. “This is a big push for us,” he said.
But Mr. Etzioni emphasized that what he was calling green A.I. should be seen as “an opportunity for additional ingenuity, not a restraint” — or a replacement for deep learning, which relies on vast computing power, and which he calls red A.I.
Indeed, the Allen Institute has just reached an A.I. milestone by correctly answering more than 90 percent of the questions on a standard eighth-grade science test. That feat was achieved with the red A.I. tools of deep learning.
Steve Lohr covers technology and economics. He was a foreign correspondent for a decade, and in 2013, he was part of the team awarded the Pulitzer Prize for Explanatory Reporting. @SteveLohr
Thursday, September 26, 2019
Thursday, September 12, 2019
Teaching AI to plan using language in a new open-source strategy game
https://ai.facebook.com/blog/-teaching-ai-to-plan-using-language-in-a-new-open-source-strategy-game/
When
humans face a complex challenge, we create a plan composed of
individual, related steps. Often, these plans are formed as natural
language sentences. This approach enables us to achieve our goal and
also adapt to new challenges, because we can leverage elements of
previous plans to tackle new tasks, rather than starting from scratch
each time.
Facebook
AI has developed a new method of teaching AI to plan effectively, using
natural language to break down complex problems into high-level plans
and lower-level actions. Our system innovates by using two AI models —
one that gives instructions in natural language and one that interprets
and executes them — and it takes advantage of the structure in natural
language in order to address unfamiliar tasks and situations. We’ve
tested our approach using a new real-time strategy game called
MiniRTSv2, and found it outperforms AI systems that simply try to
directly imitate human gameplay.
We’re now sharing our results which will be presented at NeurIPS 2019 later this year, and open-sourcing MiniRTSv2 so other researchers can use it to build and test their own imitation and reinforcement learning systems.
Previously,
the AI research community has found it challenging to bring this
hierarchical decision-making process to AI systems. Doing so has meant
that researchers had to manually specify how to break down a problem
into macro-actions, which is difficult to scale and requires expertise.
Alternatively, if the AI system has been trained to focus on the end
task, it is likely to learn how to achieve success through a single
composite action rather than a hierarchy of steps. Our work with
MiniRTSv2 shows that a different natural language-based method can make
progress against these challenges.
While
this is foundational research, it suggests that by using language to
represent plans, these systems can more efficiently generalize to a
variety of tasks and adapt to new circumstances. We believe this can
bring us closer to our long-term goal of building AI that can adapt and
generalize in real-world settings.
Building MiniRTSv2, an open source, NLP-ready game environment
MiniRTSv2
is a streamlined strategy game designed specifically for AI research.
In the game, a player commands archers, dragons, and other units in
order to defeat an opponent.
In
this sample MiniRTSv2 gameplay — recorded directly from the tool’s
interface — all the instructions that appear below the map field are
generated by an instructor model, while the corresponding in-game
actions, such as building and attacking units, are carried out by a
separate executor model.
Though
MiniRTSv2 is intentionally simpler and easier to learn than commercial
games such as DOTA 2 and StarCraft, it still allows for complex
strategies that must account for large state and action spaces,
imperfect information (areas of the map are hidden when friendly units
aren’t nearby), and the need to adapt strategies to the opponent’s
actions. Used as a training tool for AI, the game can help agents learn
effective planning skills, whether through NLP-based techniques or other
kinds of training, such as reinforcement and imitation learning.
Using language to generate high-level plans and assign low-level instructions
We
used MiniRTSv2 to train AI agents to first express a high-level
strategic plan as natural language instructions and then to act on that
plan with the appropriate sequence of low-level actions in the game
environment. This approach leverages natural language’s built-in
benefits for learning to generalize to new tasks. Those include the
expressive nature of language — different combinations of words can
represent virtually any concept or action — as well as its compositional
structure, which allows people to combine and rearrange words to create
new sentences that others can then understand. We applied these
features to the entire process of planning and execution, from the
generation of strategy and instructions to the interface that bridges
the different parts of the system’s hierarchical structure.
Our
AI agent plays a real-time strategy game using two models. The
instructor creates plans based on continually observing the game state
and issues instructions in natural language to the executor. The
executor grounds these instructions as actions, based on the current
state of the game.
The
AI agent that we built to test this approach consists of a two-level
hierarchy — an instructor model that decides on a course of action and
issues commands, and an executor model that carries out those
instructions. We trained both models using a data set collected from
human participants playing MiniRTSv2.
Those
participants worked in instructor-executor pairs, with designated
instructors issuing orders in the form of written text, and executors
accessing the game’s controls to carry those orders out. The commands
ranged from clear-cut directives, such as “build 1 dragon,” to general
instructions, such as “attack.” We used these natural language
interactions between players to generate a data set of 76,000 pairs of
instructions and executions across 5,392 games
Leveraging the versatility of natural language to learn more generalized plans
Though
MiniRTSv2 isn’t designed solely for NLP-related work, the game
environment’s text interface allows us to explore ambiguous and
context-dependent linguistic features that are relevant to building more
versatile AI. For example, given the instruction “make two more cavalry
and send them over with the other ones,” the executor model has to
grasp that “the other ones” are existing cavalry, an inference that’s
simple for most humans, but potentially challenging for AI. The agent
also has to account for the kind of potentially confusing nuances that
are common in natural language. The specific command “send idle peasant
to mine mineral” should lead to the same action as the comparatively
vague “back to mine,” which doesn’t specify which units should be moved.
At
each time step within a given MiniRTSv2 game, our system relies on
three encoders to turn inputs into feature vectors that the model can
use. The observation encoder focuses on spatial inputs (where game
objects appear on the map) and nonspatial inputs (such as the type of
unit or building that a given game object represents); the instruction
encoder generates vectors from a recent list of natural language
instructions; and the auxiliary encoder learns vectors for the remaining
global game attributes (such as the total amount of resources a player
has).
But
rather than clarifying phrasing or eliminating redundant permutations
of the same order, we intentionally leave the human instruction examples
(and corresponding executor actions) as they were delivered. The
instructor model can’t formulate original sentences and has to select
from examples from human play-throughs. This forces the agent to develop
pragmatic inference, learning how to plan and execute based on natural
language as humans actually use it, even when that usage is imprecise.
Training
our system to not only generate latent language commands but also
understand the context of those instructions resulted in a significant
boost in performance over more traditional agents. Using MiniRTSv2, we
pitted a number of different agents against an AI opponent that was
trained to directly imitate human actions, without taking language into
account. The results from these experiments showed that language
consistently improved agents’ win rates. For example, our most
sophisticated NLP-based agent, which uses a recurrent neural network
(RNN) encoder to help differentiate similar orders, beat the
non-language-based AI opponent 57.9 percent of the time. That’s
substantially better than the imitation-based agent’s 41.2 percent win
rate.
This
is the first model to show improvements in planning by generating and
executing latent natural language instructions. And though we employed a
video game to evaluate our agents, the implications of this work go far
beyond boosting the skills of game-playing AI bots, suggesting the
long-term potential of employing language to improve generalization. Our
evaluations showed that performance gains for NLP-based agents
increased with larger instruction sets, as the models were able to use
the compositional structure within language to better generalize across a
wide range of examples.
And
in addition to improving generalization, this approach has the
significant side benefit of demonstrating how decision-making AI systems
can be simultaneously high performance, versatile, and more
interpretable. If an agent’s planning process is based on natural
language, with sentences mapped directly to actions, understanding how a
system arrived at a given action could be as simple as reading its
internal transcript. The ability to quickly vet an AI’s behavior could
be particularly useful for AI assistants, potentially allowing a user to
fine-tune the system’s future actions.
Building language-based AI assistants through open science and collaboration
While
our results have focused on using language as an aid for hierarchical
decision-making, improving the ability of AI systems to utilize and
understand natural language could pave the way for an even wider range
of potential long-term benefits, such as assistants that are better at
adapting to unfamiliar tasks and surroundings. Progress in this area
might also yield systems that respond better to spoken or written
commands, making devices and platforms more accessible to people who
aren’t able to operate a touchscreen or mouse.
As
promising as our results have been, the experimental task that we’re
presenting, the NLP-based data set that we’ve created, and the MiniRTSv2
environment that we’ve updated are all novel contributions to the
field. Exploring their full potential will require a substantial
collective effort, which is why we’re inviting the wider AI community to use them.
And these resources aren’t limited to one task — for example, since the
MiniRTSv2 interface makes it easy to isolate the language activity from
the recorded games, our data set of sample commands could be valuable
for researchers training NLP systems, even if their work is unrelated to
game performance or hierarchical decision-making. We look forward to
seeing the results and insights that other researchers generate using
these tools, as we continue to advance the application of language to
improve the quality, versatility, and transparency of AI
decision-making.
Subscribe to:
Posts (Atom)