CITP Seminar: Machine Bullshit: Emergent Manipulative Behavior in Language Agents
Click here to join the seminar.
Our research group is currently trying to shed light on what we think is one of the most pressing dangers presaged by the increasing power and reach of AI technologies. The conjunction of large-scale language models like GPT-3 with advanced strategic decision-making systems like AlphaZero can bring about a plethora of extremely effective AI text-generation systems with the ability to produce compelling arguments in support of arbitrary ideas, whether true, false, benign or malicious.
Through continued interactions with many millions of users, such systems could quickly learn to produce statements that are highly likely to elicit the desired human response, belief or action. That is, these systems will reliably say whatever they need to say to achieve their goal: we call this Machine Bullshit, after Harry Frankfurt’s excellent 1986 philosophical essay “On Bullshit”. If not properly understood and mitigated, this technology could result in a large-scale behavior manipulation device far more effective than subliminal advertising, and far more damaging than “deep fakes” in the hands of malicious actors.
Our aim is to bring together insights from dynamic game theory, machine learning and human-robot interaction to better understand these risks and inform the design of safe language-enabled AI systems.”
Bio: Jaime Fernández Fisac is an assistant professor in Department of Electrical and Computer Engineering at Princeton. He is an associated faculty member in the Department of Computer Science and the Center for Statistics and Machine Learning as well as a co-director of the Princeton AI4ALL summer camp.
He is interested in ensuring the safe operation of robotic systems in the human space. Fernández Fisac’s work combines safety analysis from control theory with machine learning and artificial intelligence techniques to enable robotic systems to reason competently about their own safety in spite of using inevitably fallible models of the world and other agents. This is done by having robots monitor their own ability to understand the world around them, accounting for how the gap between their models and reality affects their ability to guarantee safety.
Much of his research uses dynamic game theory together with insights from cognitive science to enable robots to strategically plan their interaction with human beings in contexts ranging from human-robot teamwork to drone navigation and autonomous driving. His lab’s scope spans theoretical work, algorithm design, and implementation on a variety of robotic platforms.
Fernández Fisac completed his Ph.D. in electrical engineering and computer science at UC Berkeley in 2019; at the midpoint of his Ph.D., he spent six months doing R&D work at Apple. Before that, Fernández Fisac received his B.S./M.S. in electrical engineering at the Universidad Politécnica de Madrid in Spain and a master’s degree in aeronautics at Cranfield University in the UK. Before joining Princeton in fall 2020, he spent a year as a research scientist at Waymo (formerly known as Google’s Self-Driving Car project).
To request accommodations for a disability please contact Jean Butcher, firstname.lastname@example.org, at least one week prior to the event.
This seminar will be recorded.