Informatica 41 (2017) 441–449 441 Mammalian Value Systems Gopal P. Sarma School of Medicine, Emory University, Atlanta, GA USA E-mail: gopal.sarma@emory.edu Nick J. Hay Vicarious FPC, San Francisco, CA USA E-mail: nnickhay@gmail.com Keywords: friendly AI, value alignment, human values, biologically inspired AI, human-mimetic AI Received: June 24, 2013 Characterizing human values is a topic deeply interwoven with the sciences, humanities, political phi- losophy, art, and many other human endeavors. In recent years, a number of thinkers have argued that accelerating trends in computer science, cognitive science, and related disciplines foreshadow the creation of intelligent machines which meet and ultimately surpass the cognitive abilities of human beings, thereby entangling an understanding of human values with future technological development. Contemporary rese- arch accomplishments suggest increasingly sophisticated AI systems becoming widespread and responsi- ble for managing many aspects of the modern world, from preemptively planning users’ travel schedules and logistics, to fully autonomous vehicles, to domestic robots assisting in daily living. The extrapolation of these trends has been most forcefully described in the context of a hypothetical “intelligence explosion,” in which the capabilities of an intelligent software agent would rapidly increase due to the presence of feedback loops unavailable to biological organisms. The possibility of superintelligent agents, or simply the widespread deployment of sophisticated, autonomous AI systems, highlights an important theoretical problem: the need to separate the cognitive and rational capacities of an agent from the fundamental goal structure, or value system, which constrains and guides the agent’s actions. The “value alignment problem” is to specify a goal structure for autonomous agents compatible with human values. In this brief article, we suggest that ideas from affective neuroscience and related disciplines aimed at characterizing neurological and behavioral universals in the mammalian kingdom provide important conceptual foundations relevant to describing human values. We argue that the notion of “mammalian value systems” points to a potential avenue for fundamental research in AI safety and AI ethics. Povzetek: Prispevek obravnava sistem vrednot sesalcev, ki so osnova za človeški sistem vrednot, pomem- ben za umetno inteligenco. 1 Introduction Artificial intelligence, a term coined in the 1950’s at the now famous Dartmouth Conference, has come to have a widespread impact on the modern world [1, 2]. If we bro- aden the phrase to include all software, and in particular, software responsible for the control and operation of phy- sical machinery, planning and operations management, or other tasks requiring sophisticated information processing, then it goes without saying that artificial intelligence has become a critical part of the infrastructure supporting mo- dern human society. Indeed, prominent venture capitalist Mark Andresseen famously wrote that “software is eating the world,” in reference to the ubiquitous deployment of software systems across all industries and organizations, and the corresponding growth of the financial investment into software companies [3]. Nonetheless, there is a fundamental gap between the abi- lities of the most sophisticated software-based control sy- stems today and the capacities of a human child or even many animals. Our AI systems have yet to display the capacity for learning, creativity, independent thought and discovery that define human intelligence. It is a near- consensus position, however, that at some point in the future, we will be able to create software-based agents whose cognitive capacities rival those of human beings. While there is substantial variability in researchers’ fore- casts about the time-horizons of the critical breakthroughs and the consequences of achieving human-level artificial intelligence, there it is little disagreement that it is an attai- nable milestone [4, 5].1 Some have argued that the creation of human-level ar- tificial intelligence would be followed by an “intelligence explosion,” whereby the intelligence of the software-based 1There have been a number of prominent thinkers who have expressed strongly conservative viewpoints about AI timelines. See, for example, commentaries by David Deutsch, Rodney Brooks, and Douglas Hofstadter [6–8]. 442 Informatica 41 (2017) 441–449 G.P. Sarma et al. system would rapidly increase due to its ability to analyze, model, and improve its cognition by re-writing its code- base, in a feat of self-improvement impossible for biolo- gical organisms. The net result would be a “superintelli- gence,” that is, an agent whose fundamental cognitive abi- lities vastly exceed our own [9–12]. To be more explicit, let us consider a superintelligence to be any agent which can surpass the sum total of human cognitive and emotional abilities. These abilities might in- clude intellectual tasks such as mathematical or scientific research, artistic invention in musical composition or poe- try, political philosophy and the crafting of public policy, or social skills and the ability to recognize and respond to human emotions. Many commentators in recent years and decades have predicted that convergent advances in compu- ter science, robotics, and related disciplines will give rise to the development of superintelligent machines during the 21st century [4]. If it is possible to create a superintelligence, then a num- ber of natural questions arise: What would such an agent choose to do? What are the constraints that would guide its actions and to what degree can these actions be shaped by the designers? If a superintelligence can reason about and influence the world to a substantially greater degree than human beings themselves, how can we design a system to be compatible with human values? Is it even possible to formalize the notion of human values? Are human values a monolithic, internally consistent entity, or are there in- trinsic conflicts and contradictions between the values of individuals and between the value systems of different cul- tures? [9, 12–16]. It is our belief that the value alignment problem is of fun- damental importance both for its relevance to near-term de- velopments likely to be realized by the computer and robo- tics industries and for longer- term possibilities of more so- phisticated AI systems leading to superintelligence. Furt- hermore, the broader set of problems posed by the realiza- tion of intelligent, autonomous, software-based agents may provide an important unifying framework that brings toget- her disparate areas of inquiry spanning computer science, cognitive science, philosophy of mind, behavioral neuros- cience, and anthropology, to name just a few. In this article, we set aside the question of how, when, and if AI systems will be developed that are of sufficient sophistication to require a solution to the value alignment problem. This is a substantial topic in its own right which has been analyzed elsewhere. We assume the feasibility of these systems as a starting point for further analysis of the goal structures of autonomous agents and propose the notion of “mammalian value systems” as providing a fra- mework for further research. 2 Goal structures for autonomous agents 2.1 The orthogonality thesis The starting point for discussing AI goal structures is the observation that the cognitive capacities of an intelligent agent are independent of the goal structure that constrains or guides the agents’ actions, what Bostrom calls the “ort- hogonality thesis:” We have seen that a superintelligence could have a great ability to shape the future according to its goals. But what will its goals be? What is the relation bet- ween intelligence and motivation in an artificial agent? Here we develop two theses. The orthogonality thesis holds (with some caveats) that intelligence and final goals are independent variables: any level of intelli- gence could be combined with any final goal. The in- strumental convergence thesis holds that superintelli- gent agents having any of a wide range of final goals will nevertheless pursue similar intermediary goals be- cause they have common instrumental reasons to do so. Taken together, these theses help us to think about what a superintelligent agent would do. [9] The orthogonality thesis allows us to illustrate the im- portance of autonomous agents being guided by human- compatible goal structures, whether they are truly super- intelligent as Bostrom envisions, or even more modestly intelligent but highly sophisticated AI systems likely to be developed in industry in the future. Consider the example of a domestic robot that is able to clean the house, moni- tor a security system, and prepare meals independently and without human intervention. A robot with a slightly incor- rect or inadequately specified goal structure might correctly infer that a household pet has high nutritional value to its owners, but not recognize its social and emotional relati- onship to the family. We can easily imagine the consequen- ces for companies involved in creating domestic robots if a family dog or cat ends up on the dinner plate [14]. Alt- hough such a scenario is unlikely without some amount of warning2—we may notice odd or annoying behavior in the robot in other tasks, for example—it highlights an impor- tant nuance about value alignment. For example, the exact difference between animals that we value for their emoti- onal role in our lives versus those that many have deemed ethically acceptable for food is far from obvious. Indeed for someone who lives on a farm, the line can be blurred and some creatures may play both roles. As the intelligent capabilities of an agent grows, the con- sequences for slight deviations from human values will be- come greatly magnified. The reason is that such an agent possesses increasing capacity to achieve its goals, however arbitrary those goals might be. It is for this reason that rese- archers concerned with the value alignment problem have 2What exactly counts as sufficient warning, and whether the warning is heeded or not, is another matter. Mammalian Value Systems Informatica 41 (2017) 441–449 443 distanced themselves from the fictitious and absurd scena- rios portrayed in Hollywood thrillers. These movies often depict outright malevolent agents whose explicit aim is to destroy or enslave humanity. What is implicit in these sto- ries is a goal structure that has been explicitly defined to be in opposition to human values. But as the simple ex- ample of the domestic robot illustrates, this is hardly the risk we face with sophisticated AI systems. The true risk is that if we incorrectly or inadequately specify the goals of a sufficiently capable agent, then it will devote its cog- nitive capacities to a task that is at odds with our values in ways that may be subtle or even bizarre. In the example given above, there was no malevolence or ulterior motive behind the robot making a nutritious meal out of the hou- sehold pet. Rather, it simply did not recognize—due to the failure of its human designers—that the pet was valued by its owners, not for nutritional reasons, but rather for social and emotional ones [13, 14]. 2.2 Anthropomorphic bias versus anthropomorphic design Before proceeding, we mention an important caveat with regards to the orthogonality thesis, namely, that it is not a free orthogonality. The particular goal structure of an agent will almost certainly constrain the necessary cogni- tive capabilities required for the agent to operate. In ot- her words, the orthogonality thesis does not suggest that one can pair an arbitrary set of algorithms with an arbitrary goal structure. For instance, if we are building an AI sy- stem to process a large number of photographs and videos so that families can efficiently find their most memorable moments amidst terabytes of data, we know that the under- lying algorithms will be those from computer vision and not computer algebra. The primary takeaway from the ort- hogonality thesis is that when reasoning about intelligence in the abstract, we should not assume that any particular goal structure is implied. In particular, there is no reason to believe that an arbitrary AI system having the cognitive capacity of humans will necessarily have a goal structure compatible with or in opposition to that of humans. It may very well be completely arbitrary from the perspective of human values. This observation about the orthogonality thesis brings to light an important point with regards to AI goal structu- res, namely the difference between anthropomorphic bias and anthropomorphic design. Anthropomorphic bias refers to the default assumption that an arbitrary AI system will behave in a manner possessing commonalities with human beings. In practice, instances of anthropomorphic bias al- most always go hand in hand with the assumption of ma- levolent intentions on behalf of an AI system—recall our previous dismissal of Hollywood thrillers depicting agents intent on destroying or enslaving humanity. On the other hand, it may very well be the case, per- haps even necessary, that solving the value alignment pro- blem requires us to build a specific AI system that posses- ses important commonalities with the human mind. This latter perspective is what we refer to as anthropomorphic design.3 2.3 Inferring human-compatible value systems An emerging train of thought among AI safety researchers is that a human-compatible goal structure will have to be in- ferred by the AI system itself, rather than pre-programmed by the designers. The reason is that human values are rich and complex, and in addition, often contradictory and con- flicting. Therefore, if we incorrectly specify what we think to be a safe goal structure, even slight deviations can be magnified and lead to detrimental consequences. On the other hand, if an AI system begins with an uncertain model of human values, and then begins to learn our values by ob- serving our behavior, then we can substantially reduce the risks of a misspecified goal structure. Furthermore, just as we are more likely to trust mathematical calculations per- formed by a computer than by humans, if we build an AI system that we know to have greater capacity than oursel- ves at performing those cognitive operations required to infer the values of other agents by observing their beha- vior, then we gain the additional benefit of knowing that these operations will be performed with greater certainty and accuracy than were they to be pre-programmed by hu- man AI researchers. There is context in contemporary research for this kind of indirect inference, such as Inverse Reinforcement Le- arning (IRL) [17, 18] or Bayesian Inverse Planning (BIP) [19]. In these approaches, an agent learns the values, or utility function, of another agent, whether it is a human, an animal, or software system, by observing its behavior. While these ideas are in their nascent stages, practical techniques have already been developed for designing AI systems [20–23]. Russell summarizes the notion of indirect inference of human values by stating three principles that should guide the development of AI systems [14]: 1. The machine’s purpose must be to maximize the rea- lization of human values. In particular, it has no pur- pose of its own and no innate desire to protect itself. 2. The machine must be initially uncertain about what those human values are. The machine may learn more about human values as it goes along, but it may never achieve complete certainty. 3. The machine must be able to learn about human values by observing the choices that we humans make. There are almost certainly many conceptual and practi- cal obstacles that lie ahead in designing a system that infers 3Anthropomorphic design refers to a more narrow class of systems than the term “human-compatible AI,” which has recently come into use. See, for example, The Berkeley Center for Human-Compatible AI (http://www.humancompatible.ai). 444 Informatica 41 (2017) 441–449 G.P. Sarma et al. the values of human beings from observing our behavior. In particular, human desires can often be masked by many lay- ers of conflicting emotions, they can often be inconsistent, and the desires of one individual may outright contradict the desires of another. In the context of a superintelligent agent capable of exerting substantial influence on the world (as opposed to a domestic robot), it is natural to ask about variations in the value systems of different cultures. It is often assumed that many human conflicts on a global scale stem from conflicts in the underlying value systems of the respective cultures or nation states. Is it even possible, the- refore, for an AI system, no matter how intelligent, to arrive at a consensus goal structure that respects the desires of all people and cultures? We make two observations in response to this important set of questions. The first is that when we say that cultures have conflicting values, implicit in this statement are our own limited cognitive capacities and ability to model the behavior and mental states of other individuals and groups. An AI system with capabilities vastly greater than oursel- ves may quickly perceive fundamental commonalities and avenues for conflict resolution that we are unable to envi- sion. To motivate this scenario, we give a highly simplified ex- ample from negotiation theory. A method known as “prin- cipled negotiation” distinguishes between values and posi- tions [24]. As an example, if two friends are deciding on a restaurant for dinner, and one wants Indian food and the other Italian, it may be that the first person simply likes spicy food and the second person wants noodles. These preferences are the values, spicy food and noodles, that the corresponding positions, Indian and Italian, instantiate. In this school of thought, when two parties are attempting to resolve a conflict, they should negotiate from values, rather than positions. That is, if we have some desire that is in conflict with another, we should ask ourselves—whether in the context of a business negotiation, family dispute, or ma- jor international conflict—what the underlying value is that the desire reflects. By understanding the underlying values, we may see that there is a mutually satisfactory set of out- comes satisfying all parties that we failed to see initially. In this particular instance, if the friends are able to state their true underlying preferences, they may recognize that Thai cuisine will satisfy both parties. We mention this example from negotiation theory to raise the possibility that what we perceive to be fundamentally conflicting values in hu- man society might actually be conflicting positions arising from distinct, but reconcilable values when viewed from the perspective of a higher level of intelligence. The second observation is that what we colloquially re- fer to as the values of a particular culture, or even collective human values, reflect not only innate features of the human mind, but also the development of human society. In ot- her words, to understand the underlying value system that guides human behavior, which would ultimately need to be modeled and inferred by an AI system, it may be helpful to disentangle those aspects of modern cultural values which were latent, but not explicitly evident during earlier periods of human history. Although an agent utilizing Inverse Reinforcement Lear- ning or Bayesian Inverse Planning will learn and refine its model of human values by observing our behavior, it must begin with some very rough or approximate initial assump- tions about the nature of the values it is trying to learn. By starting from a more accurate initial goal structure, an agent might learn from fewer examples, thus minimizing the li- kelihood of real-world actions having adverse affects. In the remainder of this article, we argue that the neurologi- cal substrate common to mammals and their corresponding behaviors may provide a framework for characterizing the structure of the initially uncertain value system of an auto- nomous, intelligent agent. 2.4 Mammalian value systems Our core thesis is the following: What we call human values can be informally decomposed into 1) mammalian values, 2) human cognition, and 3) several millennia of human social and cultural evolution. This decomposi- tion suggests that contemporary research broadly spanning the study of animal behavior, biological anthropology, and comparative neuroanatomy may be relevant to the value alignment problem, and in particular, in characterizing the initially uncertain goal structure which is refined through observation by an AI system. Additionally, in analyzing the subsequent behavioral trajectories of intelligent, auto- nomous agents, we can decompose the resulting dynamics as being guided by mammalian values merged with AI cog- nition. Aspects of contemporary human values which are the result of incidental historical processes—the third com- ponent of our decomposition above—might naturally arise in the course of the evolution of the AI system (though not necessarily), even though they were not directly program- med into the agent.4 There are many factors that might in- fluence the extent to which this third component of human values continues to be represented in the AI system. Ex- amples might include whether or not these values remain meaningful in a world where other problems had been sol- ved and the extent to which certain cultural values which were perceived to be in conflict with others could be re- solved with a more fundamental understanding stemming from the combination of mammalian values and AI cogni- tion.5 4Many human values communicated to children during the course of maturation and development are the result of incidental historical proces- ses. As an example, consider the rich set of cultural norms and social rituals surrounding food preparation. One does not need to have lived the entire history of a given culture to learn these norms. The same may be true of an AI system. 5Ethical norms can often vary depending on resource constraints which may also be the result of incidental historical processes. The norms of behavior may be different in a war zone where individuals are fighting for survival than in an affluent society during peacetime. If a family st- ruggling to survive in a war torn country is able to escape and move to a more stable region, these same behaviors may no longer be necessary. In a similar vein, imagine an AI system that has significantly impacted Mammalian Value Systems Informatica 41 (2017) 441–449 445 We want to emphasize that our claim is not that mamma- lian values are synonymous with human values. Rather, our thesis is that there are many aspects of human values which are the result of historical processes driven by human cog- nition. Consequently, many structural aspects of human ex- perience and human society which we colloquially refer to as “values” are derived entities, rather than features of the initial AI goal structure. As a thought experiment, consider a scenario whereby the fully digitized corpus of human lite- rature, cinema, and ongoing global developments commu- nicated via the Internet are analyzed and modeled by an AI system constructed around a core mammalian goal struc- ture. In the conceptual framework that we propose, this initially mammalian structure would gradually come to re- flect the more nuanced aspects of human society as the AI refines its model of human values via analysis and hypot- hesis generation. We also mention that as our aim in this article is to focus on the structure of the initial AI motiva- tional system and not other aspects of AI more broadly, we set aside the possible role human interaction and feedback may play in the subsequent development of the AI system’s cognition and instrumental values. 2.4.1 Neural correlates of values: behavioral and neurological foundations Our thesis about mammalian values is predicated on two converging lines of evidence, one primarily behavioral and the other primarily neuroscientific. Behaviorally, it is not difficult to characterize intuitively what human values are when viewed from the perspective of the mammalian king- dom. Like many other animals, humans are social cre- atures and many, if not most, of our fundamental drives originate from our relationships with others. Attachment, loss, anger, territoriality, playfulness, joy, anxiety, and love are all deeply rooted emotions that guide our behavior and which have been foundational elements in the emergence of human cognition, culture, and the structure of society6 [25–36]. The scientific study of behavior is largely the domain of the disciplines of ethology and behaviorism. As we are primarily concerned with emotions, we will focus on be- havioral insights and taxonomies originating from the sub- community of affective neuroscience, which also aims to correlate these behaviors with underlying neural architec- ture. More formally, Panksepp and Biven categorize the in- formal list given above into seven motivational and emoti- onal systems that are common to mammals: seeking, rage, fear, lust, care, panic/grief, and play [37]. We now give brief summaries of each of these systems: 1. SEEKING: This is the system that primarily mediates global affairs by solving major problems in food or energy production or by discovering novel insights into diplomatic strategy. Such an agent may find that previously necessary behaviors that have a rich human history are no longer needed. 6While we have mentioned several active areas of research, there are certainly others that we are simply not aware of. We apologize in advance to those scholars whose work we have not cited here. exploratory behavior and also enables the other sys- tems. The seeking system can give rise to both po- sitive and negative emotions. For instance, a mother who needs to feed her offspring will go in search of food, and the resulting maternal / child bonding (via the CARE system; see below) creates positive emotio- nal reinforcement. On the other hand, physical threats can generate negative emotions and prompt an animal to seek shelter and safety. The behaviors correspon- ding to SEEKING have been broadly associated with the dopaminergic systems of the brain, specifically re- gions interconnected with the ventral tegmental area and nucleus accumbens. 2. RAGE: The behaviors corresponding to rage are tar- geted and more narrowly focused than those governed by the seeking system. Rage compels animals towards specific threats and is generally accompanied by nega- tive emotions. However, it should be noted that in an adversarial scenario where rage can lead to victory, it can also be accompanied by the positive emotions of triumph or glory. The RAGE system involves medial regions of the amygdala, medial regions of the hypot- halamus, and the periaqueductal gray. 3. FEAR: The two systems described thus far are directly linked to externally directed, action-oriented beha- vior. In contrast, fear describes a system which pla- ces an animal in a negative affective state, one which it would prefer not to be in. In the early stages, fear tends to correspond to stationary states, after which it can transition to seeking or rage, and ultimately, at- tempts to flee from the offending stimulus. However, these are secondary effects, and the primary physical state of fear is typically considered to be an immobile one. The FEAR system involves central regions of the amygdala, anterior and medial regions of the hypotha- lamus, and dorsal regions of the periaqueductal gray. 4. LUST: Lust describes the system leading to behaviors of courtship and reproduction. Like fear, it will tend to trigger the seeking system, but can also lead to ne- gative affective states if satisfaction is not achieved. The LUST system involves anterior and ventromedial regions of the hypothalamus. 5. CARE: Care refers to acts of tenderness directed to- wards loved ones, and in particular, an animal’s off- spring. As we described in the context of seeking, the feelings associated with caring and nurturing can be profoundly positive and play a crucial component in the social behavior of mammals. CARE is associated with the ventromedial hypothalamus and the oxytocin system. 6. PANIC / GRIEF: Activation of the panic / grief sy- stem corresponds to profound psychological pain, and is generally not associated with external physical cau- ses. In young animals, this system is typically acti- 446 Informatica 41 (2017) 441–449 G.P. Sarma et al. vated by separation from caregivers, and is the under- lying network behind “separation anxiety.” Like care, the panic / grief system is a fundamental component of mammalian social behavior. It is the negative af- fective system which drives animals towards relations- hips with other animals, thereby stimulating the care system, generating feelings of love and affection, and giving rise to social bonding. This system is associ- ated with the periaqueductal gray, ventral septal area, and anterior cingulate. 7. PLAY: The play system corresponds to lighthearted behavior in younger animals and is a key component of social bonding, friendship, as well as the learning of survival-oriented skills. Although play can superfi- cially resemble aggression, there are fundamental dif- ferences between play and adult aggression. At an emotional level, it goes without saying that play corre- sponds to positive affective states, and unlike aggres- sive behavior, is typically part of a larger, orchestrated sequence of events. In play, for example, animals of- ten alternate between assuming dominant and submis- sive roles. The PLAY system is currently less neuro- anatomically localized, but involves midline thalamic regions. As we stated earlier, our thesis about mammalian values originates from two convergent lines of evidence, one be- havioral and the other neuroscientific. What we refer to as the “neural correlates of values,” or NCV, are the common mammalian neural structures which underlie the motivati- onal and emotional systems summarized above. To the ex- tent that human values are intertwined with our emotions, these architectural commonalities suggest that the shared mammalian neurological substrate is of importance to un- derstanding human value alignment in sophisticated lear- ning systems. Panksepp and Biven write, To the best of our knowledge, the basic biological va- lues of all mammalian brains were built upon the same basic plan, laid out in . . . affective circuits that are con- centrated in subcortical regions, far below the neocor- tical “thinking cap” that is so highly developed in hu- mans. Mental life would be impossible without this foundation. There, among the ancestral brain networks that we share with other mammals, a few ounces of brain tissue constitute the bedrock of our emotional lives, generating the many primal ways in which we can feel emotionally good or bad within ourselves. As we mature and learn about ourselves, and the world in which we live, these systems provide a solid founda- tion for further mental developments [37]. Latent in this excerpt is the decomposition that we have suggested earlier. The separation of the mammalian brain into subcortical and neocortical regions, roughly corre- sponding to emotions and cognition respectively, implies that we can attempt to reason by analogy what the archi- tecture of an AI system would look like with a human- compatible value system. In particular, the initially uncer- tain goal structure that the AI system refines via observa- tion may be much simpler than we might imagine by re- flecting on the complexities of human society and indivi- dual desires. As we have illustrated using our simple ex- ample from negotiation theory, our intuitive understanding of human values, and the conflicts that we regularly wit- ness between individuals and groups, may in fact represent conflicting positions stemming from a shared fundamen- tal value system, a value system that originates from the subcortical regions of the brain, and which other mammals share with us.7 Referring once again to the work of Panksepp, In short, many of the ancient, evolutionarily derived brain systems all mammals share still serve as the foundations for the deeply experienced affective pro- clivities of the human mind. Such ancient brain functi- ons evolved long before the emergence of the human neocortex with its vast cognitive skills. Among li- ving species, there is certainly more evolutionary di- vergence in higher cortical abilities than in subcortical ones [39]. The emphasis on the diversity in higher cortical abilities is of particular relevance to the decomposition that we have proposed. We might ask what the full spectrum of higher cortical abilities are that could be built on top of the com- mon mammalian substrate provided by the evolutionarily older parts of the brain. We need not confine ourselves to those manifestations of higher cognition that we see in nature, or that would even be hypothetical consequences of continued evolution by natural selection. Indeed, one resta- tement of our core thesis is to consider—in the abstract or as a thought experiment—the consequences of extending 7There is a contemporary and light-hearted social phenomenon which provides an evocative illustration of the universality of mammalian emo- tions, namely, the volume of animal videos posted to YouTube. From ordinary citizens with pets, to clips from nature documentaries, animal videos are regularly watched by millions of viewers worldwide. Indivi- dual videos and compilations of “animal odd couples,” “unlikely animal friends,” “dogs and babies,” and “animal friendship between different spe- cies” are commonly searched enough to be auto-completed by YouTube’s search capabilities. It is hardly surprising that these charming and heart- warming videos are so compelling to viewers of all age groups, genders, and ethnic backgrounds. Our relationships with other animals, whether home owners and their pets, or scientists and the wild animals that they study, tell us something deeply fundamental about ourselves [38]. The strong emotional bonds that humans form with other animals, in particu- lar, with our direct relatives in the mammalian kingdom, and the draw to simply watching this social behavior in other mammals, is a vivid illus- tration of the fundamental role that emotions play in our inner life and in guiding our behavior. In the future, the potential to apply inverse reinforcement learning (or related techniques) to large datasets of videos, including short clips from YouTube, movies, TV shows, documentaries, etc. opens up an interes- ting avenue to evaluate and further refine the hypothesis presented here. For instance, when such technology becomes available, we might ima- gine comparing the inferred goal structures when restricted to videos of human behavior versus those restricted to mammalian behavior. There are many other variations along these lines, for instance, restricting to vi- deos of non-mammalian behavior, mammals as well as humans, different cultures, etc. Mammalian Value Systems Informatica 41 (2017) 441–449 447 the diversity of brain architectures to include higher cor- tical abilities arising not from natural selection, but rather the de novo architectures of artificial intelligence. 2.5 Relationship to moral philosophy It is hardly a surprise that a vibrant area of research within AI safety is the relationship of contemporary and historical theories of moral philosophy to the problem of value alig- nment. Indeed, researchers have specifically argued for the relevance of moral philosophy in the context of the inverse reinforcement learning paradigm (IRL) that is the starting point for analysis in this article [40]. Is the framework we propose in opposition to those that are oriented towards moral philosophy? On the one hand, our perspective is that the field of AI safety is simply too young to make such judgments. At our present level of un- derstanding, we believe each of these agendas form solid foundations for further research and there seems little re- ason to pursue one to the exclusion of the other. On the other hand, we would also argue that this distinction is a false dichotomy. Indeed, there are active areas of rese- arch in the ethics community aimed at understanding the neurological and cognitive underpinning of human moral reasoning [41, 42]. Therefore, it is quite possible that a hybrid approach to value alignment emerges, bridging the “value primitives” perspective we advocate here with rese- arch from moral philosophy.8 3 Discussion The possibility of autonomous, software-based agents, whether self-driving cars, domestic robots, or the longer- term possibilities of superintelligence, highlights an impor- tant theoretical problem—the need to separate the intelli- gent capabilities of such a system from the fundamental values which guide the agents’ actions. For such an agent to exist in a human world and to act in a manner compatible 8In a recent article, Baum has argued that the normative basis for “so- cial choice” and “bottom-up” approaches to AI ethics must overcome strong obstacles that have been insufficiently explored by the AI safety community [43]. Although the approach we describe here decomposes values into more fundamental components, it is not a priori in opposi- tion to top-down ethics. In an extreme case, one could certainly imagine employing a purely predetermined approach to ethics within the context of mammalian values in which no value learning takes place. However, as we stated above, we suspect that an intermediate ground will be found when the issues are more thoroughly examined, and for that reason, we are reluctant to endorse either a bottom-up or a top-down approach too strongly. Given the intellectual youth of the field of AI safety, we see little reason to give strong preference to one set of approaches over the other. Moreover, an important observation that Baum makes in framing his ar- gument is that considerable work relevant to AI ethics already exists in the social choice literature, and yet none of this work has been discussed in any detail by the AI safety community. In our minds, this is a more fun- damental point, namely, that there is substantial scholarship in many areas of academic research relevant to AI safety. For this reason, we believe that where there is controversy, the first step should be to ensure that the best possible representations of given viewpoints have been made visible and adequately discussed before endorsing particular courses of action. with human values, these values would need to be explicitly modeled and formalized. An emerging train of thought in AI safety research is that this modeling process would need to be conducted by the AI system itself, rather than by the system’s designers. In other words, the agent would start off with an initially uncertain goal structure and infer hu- man values over time by observing our behavior. The question that motivates this article is to ask the fol- lowing: what can we say about the broad features of the initial goal structure that the agent then refines through ob- servation and hypothesis generation? The perspective we advocate is to view human values within the context of the broader mammalian kingdom, thereby providing implicit priors on the latent structure of the values we aim to in- fer. The shared neurological structures underlying mam- malian emotions and their corresponding social behaviors provide a starting point for formalizing an initial value sy- stem for autonomous, software-based agents. There are se- veral practical implications of having a more detailed un- derstanding of the structure of human values. By having more detailed prior information, it may be possible to learn from fewer examples. For an agent that is actively making decisions and having an impact on the world, learning an ethical framework more efficiently can minimize potential catastrophes. Furthermore, an informative prior may make approaches to AI safety which are otherwise computatio- nally intractable into practical options. From this vantage point, we argue that what we colloqui- ally refer to as human values can be informally decompo- sed into 1) mammalian values, 2) human cognition, and 3) several millennia of human social and cultural evolution. In the context of a de novo artificially intelligent agent, we can characterize desirable, human-compatible behavior as being described by mammalian values merged with AI cognition. It goes without saying that we have left out a considerable amount of detail in this description. The spe- cifics of Inverse Reinforcement Learning, the many neu- roscientific nuances underlying the comparative neuroana- tomy, physiology, and function of the mammalian brain, as well as the controversies and competing theories in the re- spective disciplines are all substantial topics on their own right. Our omission of these issues is not out of lack of re- cognition or belief that they are unimportant. Rather, our aim in this article has been to present a high-level over- view of a richly interdisciplinary set of questions whose broad outlines have only recently begun to take shape. We will tackle these issues and others in a subsequent series of manuscripts and invite interested researchers to join us. Our fundamental motivation in proposing this framework is to bring together scholars from diverse communities that may not be aware of each other’s research and their po- tential for synergy. We believe that there is a wealth of existing research which can be fruitfully re-examined and re-conceptualized from the perspective of artificial intelli- gence and the value alignment problem. We hope that ad- ditional interaction between these communities will help to 448 Informatica 41 (2017) 441–449 G.P. Sarma et al. refine and more precisely define research problems relevant to designing safe AI goal structures. Acknowledgements We would like to thank Adam Safron, Owain Evans, Daniel Dewey, Miles Brundage, and several anonymous reviewers for insightful discussions and feedback on the manuscript. We would also like to thank the guest editors of Informa- tica, Ryan Carey, Matthijs Maas, Nell Watson, and Roman Yampolskiy, for organizing this special issue. References [1] S. J. Russell and P. Norvig, Artificial Intelligence: A Modern Approach. Upper Saddle River, NJ, USA: Prentice Hall Press, 3rd ed., 2009. [2] N. J. Nilsson, The Quest for Artificial Intelligence. Cambridge University Press, 2009. [3] M. Andreessen, “Why Software Is Eating The World,” Wall Street Journal, vol. 20, 2011. [4] V. C. Müller and N. Bostrom, “Future Progress in Ar- tificial Intelligence: A survey of expert opinion,” in Fundamental issues of artificial intelligence, pp. 553– 570, Springer, 2016. [5] K. Grace, J. Salvatier, A. Dafoe, B. Zhang, and O. Evans, “When Will AI Exceed Human Perfor- mance? Evidence from AI Experts,” ArXiv e-prints, May 2017. [6] R. Brooks, “The Seven Deadly Sins of AI Predicti- ons,” MIT Technology Review, vol. 10, no. 6, 2017. [7] D. Deutsch, “How close are we to creating artificial intelligence?,” AEON Magazine, vol. 10, no. 3, 2012. [8] J. Somers, “The Man Who Would Teach Machines to Think,” The Atlantic, vol. 11, 2013. [9] N. Bostrom, Superintelligence: Paths, Dangers, Stra- tegies. OUP Oxford, 2014. [10] M. Shanahan, The Technological Singularity. MIT Press, 2015. [11] I. J. Good, “Speculations Concerning the First Ul- traintelligent Machine,” Advances In Computers, vol. 6, no. 99, pp. 31–83, 1965. [12] D. Chalmers, “The Singularity: A Philosophical Ana- lysis,” Journal of Consciousness Studies, vol. 17, no. 9-10, pp. 7–65, 2010. [13] E. Yudkowsky, “Artificial Intelligence as a Positive and Negative Factor in Global Risk,” in Global Ca- tastrophic Risks (Nick Bostrom and Milan Cirkovic, ed.), p. 303, Oxford University Press Oxford, UK, 2008. [14] S. Russell, “Should We Fear Supersmart Robots?,” Scientific American, vol. 314, no. 6, pp. 58–59, 2016. [15] S. Omohundro, “Autonomous technology and the gre- ater human good,” Journal of Experimental & Theo- retical Artificial Intelligence, vol. 26, no. 3, pp. 303– 315, 2014. [16] S. M. Omohundro, “The Basic AI Drives,” in AGI, vol. 171, pp. 483–492, 2008. [17] A. Y. Ng and S. J. Russell, “Algorithms For In- verse Reinforcement Learning,” in International Con- ference on Machine Learning, pp. 663–670, 2000. [18] D. Hadfield-Menell, A. Dragan, P. Abbeel, and S. Russell, “Cooperative inverse reinforcement lear- ning,” 2016. [19] C. L. Baker, R. R. Saxe, and J. B. Tenenbaum, “Bay- esian Theory of Mind: Modeling Joint Belief-Desire Attribution,” in Proceedings of the Thirty-Second An- nual Conference of the Cognitive Science Society, pp. 2469–2474, 2011. [20] O. Evans, A. Stuhlmüller, and N. D. Goodman, “Learning the Preferences of Ignorant, Inconsistent Agents,” arXiv:1512.05832, 2015. [21] O. Evans and N. D. Goodman, “Learning the Pre- ferences of Bounded Agents,” in NIPS Workshop on Bounded Optimality, 2015. [22] M. O. Riedl and B. Harrison, “Using Stories to Teach Human Values to Artificial Agents,” in Proceedings of the 2nd International Workshop on AI, Ethics and Society, Phoenix, Arizona, 2016. [23] M. O. Riedl, “Computational Narrative Intelligence: A Human-Centered Goal for Artificial Intelligence,” arXiv preprint arXiv:1602.06484, 2016. [24] R. Fisher and W. Ury, Getting to Yes. Simon & Schus- ter Sound Ideas, 1987. [25] I. Horswill, “Men Are Dogs (and Women Too),” in AAAI Fall Symposium: Naturally-Inspired Artificial Intelligence, pp. 67–71, 2008. [26] L. W. Swanson, “Cerebral Hemisphere Regulation of Motivated Behavior,” Brain Research, vol. 886, no. 1, pp. 113–164, 2000. [27] L. W. Swanson, Brain Architecture: Understanding the Basic Plan. Oxford University Press, 2012. [28] J. H. Barkow, L. Cosmides, and J. Tooby, The Adap- ted Mind: Evolutionary Psychology and the Genera- tion of Culture. Oxford University Press, 1995. [29] S. Dehaene and L. Cohen, “Cultural Recycling of Cortical Maps,” Neuron, vol. 56, no. 2, pp. 384–398, 2007. Mammalian Value Systems Informatica 41 (2017) 441–449 449 [30] C. Peterson and M. E. Seligman, Character Strengths and Virtues: A Handbook and Classification. Oxford University Press, 2004. [31] S. Schnall, J. Haidt, G. L. Clore, and A. H. Jordan, “Disgust as Embodied Moral Judgment,” Personality and social psychology bulletin, 2008. [32] J. B. Tenenbaum, C. Kemp, T. L. Griffiths, and N. D. Goodman, “How to Grow a Mind: Statistics, Struc- ture, and Abstraction,” Science, vol. 331, no. 6022, pp. 1279–1285, 2011. [33] J. Bowlby, Attachment and Loss, vol. 3. Basic books, 1980. [34] S. W. Porges, “Orienting in a Defensive World: Mam- malian Modifications of Our Evolutionary Heritage. A Polyvagal Theory,” Psychophysiology, vol. 32, no. 4, pp. 301–318, 1995. [35] J. Cassidy, Handbook of Attachment: Theory, Rese- arch, and Clinical Applications. Rough Guides, 2002. [36] M. Tomasello, The Cultural Origins of Human Cog- nition. Harvard University Press, 1999. [37] J. Panksepp and L. Biven, The Archaeology of Mind: Neuroevolutionary Origins of Human Emotions. WW Norton & Company, 2012. [38] R. List, “Why I Identify as a Mammal,” The New York Times, 10 2015. [39] J. Panksepp, Affective Neuroscience: The Foundati- ons of Human and Animal Emotions. Oxford univer- sity press, 1998. [40] S. Armstrong and J. Leike, “Towards Interactive In- verse Reinforcement Learning,” in NIPS, 2016. [41] J. Greene and J. Haidt, “How (and where) does mo- ral judgment work?,” Trends in Cognitive Sciences, vol. 6, no. 12, pp. 517–523, 2002. [42] J. D. Greene, “The cognitive neuroscience of mo- ral judgment,” The Cognitive Neurosciences, vol. 4, pp. 1–48, 2009. [43] S. D. Baum, “Social Choice Ethics in Artificial Intel- ligence,” AI & SOCIETY, pp. 1–12, 2017.