6 big ethical questions about the future of AI Genevieve Bell
Let me tell you a story
about artificial intelligence.
There’s a building in Sydney
at 1 Bligh Street.
It houses lots of government apartments
and busy people.
From the outside, it looks like something
out of American science fiction:
all gleaming glass and curved lines,
and a piece of orange sculpture.
On the inside, it has excellent coffee
on the ground floor
and my favorite lifts in Sydney.
They’re beautiful;
they look almost alive.
And it turns out
I’m fascinated with lifts.
For lots of reasons.
But because lifts are one of the places
you can see the future.
In the 21st century, lifts are interesting
because they’re one of the first places
that AI will touch you
without you even knowing it happened.
In many buildings all around the world,
the lifts are running a set of algorithms.
A form of protoartificial intelligence.
That means before you even
walk up to the lift to press the button,
it’s anticipated you being there.
It’s already rearranging
all the carriages.
Always going down, to save energy,
and to know where
the traffic is going to be.
By the time you’ve actually
pressed the button,
you’re already part of an entire system
that’s making sense of people
and the environment
and the building and the built world.
I know when we talk about AI,
we often talk about a world of robots.
It’s easy for our imaginations
to be occupied with science fiction,
well, over the last 100 years.
I say AI and you think “The Terminator.”
Somewhere, for us, making the connection
between AI and the built world,
that’s a harder story to tell.
But the reality is AI is already
everywhere around us.
And in many places.
It’s in buildings and in systems.
More than 200 years of industrialization
suggest that AI will find its way
to systems-level scale relatively easily.
After all, one telling of that history
suggests that all you have to do
is find a technology,
achieve scale and revolution will follow.
The story of mechanization,
automation and digitization
all point to the role of technology
and its importance.
Those stories of technological
transformation
make scale seem, well, normal.
Or expected.
And stable.
And sometimes even predictable.
But it also puts the focus squarely
on technology and technology change.
But I believe that scaling a technology
and building a system
requires something more.
We founded the 3Ai Institute
at the Australian National University
in September 2017.
It has one deceptively simple mission:
to establish a new branch of engineering
to take AI safely, sustainably
and responsibly to scale.
But how do you build a new branch
of engineering in the 21st century?
Well, we’re teaching it into existence
through an experimental education program.
We’re researching it into existence
with locations as diverse
as Shakespeare’s birthplace,
the Great Barrier Reef,
not to mention one of Australia’s
largest autonomous mines.
And we’re theorizing it into existence,
paying attention to the complexities
of cybernetic systems.
We’re working to build something new
and something useful.
Something to create the next generation
of critical thinkers and critical doers.
And we’re doing all of that
through a richer understanding
of AI’s many pasts and many stories.
And by working collaboratively
and collectively
through teaching and research
and engagement,
and by focusing as much
on the framing of the questions
as the solving of the problems.
We’re not making a single AI,
we’re making the possibilities for many.
And we’re actively working
to decolonize our imaginations
and to build a curriculum and a pedagogy
that leaves room for a range of different
conversations and possibilities.
We are making and remaking.
And I know we’re always
a work in progress.
But here’s a little glimpse
into how we’re approaching
that problem of scaling a future.
We start by making sure
we’re grounded in our own history.
In December of 2018,
I took myself up to the town of Brewarrina
on the New South Wales-Queensland border.
This place was a meeting place
for Aboriginal people,
for different groups,
to gather, have ceremonies,
meet, to be together.
There, on the Barwon River,
there’s a set of fish weirs
that are one of the oldest
and largest systems
of Aboriginal fish traps in Australia.
This system is comprised
of 1.8 kilometers of stone walls
shaped like a series of fishnets
with the “Us” pointing down the river,
allowing fish to be trapped
at different heights of the water.
They’re also fish holding pens
with different-height walls for storage,
designed to change the way the water moves
and to be able to store
big fish and little fish
and to keep those fish
in cool, clear running water.
This fish-trap system was a way to ensure
that you could feed people
as they gathered there in a place
that was both a meeting of rivers
and a meeting of cultures.
It isn’t about the rocks
or even the traps per se.
It is about the system
that those traps created.
One that involves technical knowledge,
cultural knowledge
and ecological knowledge.
This system is old.
Some archaeologists
think it’s as old as 40,000 years.
The last time we have its recorded uses
is in the nineteen teens.
It’s had remarkable longevity
and incredible scale.
And it’s an inspiration to me.
And a photo of the weir
is on our walls here at the Institute,
to remind us of the promise
and the challenge
of building something meaningful.
And to remind us
that we’re building systems
in a place where people have built systems
and sustained those same systems
for generations.
It isn’t just our history,
it’s our legacy as we seek to establish
a new branch of engineering.
To build on that legacy
and our sense of purpose,
I think we need a clear framework
for asking questions about the future.
Questions for which there aren’t
ready or easy answers.
Here, the point is the asking
of the questions.
We believe you need to go
beyond the traditional approach
of problem-solving,
to the more complicated one
of question asking
and question framing.
Because in so doing, you open up
all kinds of new possibilities
and new challenges.
For me, right now,
there are six big questions
that frame our approach
for taking AI safely, sustainably
and responsibly to scale.
Questions about autonomy,
agency, assurance,
indicators, interfaces and intentionality.
The first question we ask is a simple one.
Is the system autonomous?
Think back to that lift on Bligh Street.
The reality is, one day,
that lift may be autonomous.
Which is to say it will be able
to act without being told to act.
But it isn’t fully autonomous, right?
It can’t leave that Bligh Street building
and wonder down
to Circular Quay for a beer.
It goes up and down, that’s all.
But it does it by itself.
It’s autonomous in that sense.
The second question we ask:
does this system have agency?
Does this system have controls
and limits that live somewhere
that prevent it from doing certain
kinds of things under certain conditions.
The reality with lifts,
that’s absolutely the case.
Think of any lift you’ve been in.
There’s a red keyslot
in the elevator carriage
that an emergency services person
can stick a key into
and override the whole system.
But what happens
when that system is AI-driven?
Where does the key live?
Is it a physical key, is it a digital key?
Who gets to use it?
Is that the emergency services people?
And how would you know
if that was happening?
How would all of that be manifested
to you in the lift?
The third question we ask
is how do we think about assurance.
How do we think about all of its pieces:
safety, security, trust, risk,
liability, manageability,
explicability, ethics,
public policy, law, regulation?
And how would we tell you
that the system was safe and functioning?
The fourth question we ask
is what would be our interfaces
with these AI-driven systems.
Will we talk to them?
Will they talk to us,
will they talk to each other?
And what will it mean to have
a series of technologies we’ve known,
for some of us, all our lives,
now suddenly behave
in entirely different ways?
Lifts, cars, the electrical grid,
traffic lights, things in your home.
The fifth question
for these AI-driven systems:
What will the indicators be
to show that they’re working well?
Two hundred years
of the industrial revolution
tells us that the two most important ways
to think about a good system
are productivity and efficiency.
In the 21st century,
you might want to expand
that just a little bit.
Is the system sustainable,
is it safe, is it responsible?
Who gets to judge those things for us?
Users of the systems
would want to understand
how these things are regulated,
managed and built.
And then there’s the final,
perhaps most critical question
that you need to ask
of these new AI systems.
What’s its intent?
What’s the system designed to do
and who said that was a good idea?
Or put another way,
what is the world
that this system is building,
how is that world imagined,
and what is its relationship
to the world we live in today?
Who gets to be part of that conversation?
Who gets to articulate it?
How does it get framed and imagined?
There are no simple answers
to these questions.
Instead, they frame what’s possible
and what we need to imagine,
design, build, regulate
and even decommission.
They point us in the right directions
and help us on a path to establish
a new branch of engineering.
But critical questions aren’t enough.
You also need a way of holding
all those questions together.
For us at the Institute,
we’re also really interested
in how to think about AI as a system,
and where and how to draw
the boundaries of that system.
And those feel like especially
important things right now.
Here, we’re influenced by the work
that was started way back in the 1940s.
In 1944, along with anthropologists
Gregory Bateson and Margaret Mead,
mathematician Norbert Wiener
convened a series of conversations
that would become known
as the Macy Conferences on Cybernetics.
Ultimately, between 1946 and 1953,
ten conferences were held
under the banner of cybernetics.
As defined by Norbert Wiener,
cybernetics sought
to “develop a language and techniques
that will enable us to indeed attack
the problem of control and communication
in advanced computing technologies.”
Cybernetics argued persuasively
that one had to think
about the relationship
between humans, computers
and the broader ecological world.
You had to think about them
as a holistic system.
Participants in the Macy Conferences
were concerned with how the mind worked,
with ideas about
intelligence and learning,
and about the role
of technology in our future.
Sadly, the conversations that started
with the Macy Conference
are often forgotten
when the talk is about AI.
But for me, there’s something
really important to reclaim here
about the idea of a system
that has to accommodate culture,
technology and the environment.
At the Institute, that sort
of systems thinking is core to our work.
Over the last three years,
a whole collection of amazing people
have joined me here
on this crazy journey to do this work.
Our staff includes anthropologists,
systems and environmental engineers,
and computer scientists
as well as a nuclear physicist,
an award-winning photo journalist,
and at least one policy
and standards expert.
It’s a heady mix.
And the range of experience
and expertise is powerful,
as are the conflicts and the challenges.
Being diverse requires
a constant willingness
to find ways to hold people
in conversation.
And to dwell just a little bit
with the conflict.
We also worked out early
that the way to build
a new way of doing things
would require a commitment to bringing
others along on that same journey with us.
So we opened our doors
to an education program very quickly,
and we launched our first
master’s program in 2018.
Since then, we’ve had two cohorts
of master’s students
and one cohort of PhD students.
Our students come from all over the world
and all over life.
Australia, New Zealand, Nigeria, Nepal,
Mexico, India, the United States.
And they range in age from 23 to 60.
They variously had backgrounds
in maths and music,
policy and performance,
systems and standards,
architecture and arts.
Before they joined us at the Institute,
they ran companies,
they worked for government,
served in the army, taught high school,
and managed arts organizations.
They were adventurers
and committed to each other,
and to building something new.
And really, what more could you ask for?
Because although I’ve spent
20 years in Silicon Valley
and I know the stories
about the lone inventor
and the hero’s journey,
I also know the reality.
That it’s never just a hero’s journey.
It’s always a collection of people
who have a shared sense of purpose
who can change the world.
So where do you start?
Well, I think you start where you stand.
And for me, that means
I want to acknowledge
the traditional owners of the land
upon which I’m standing.
The Ngunnawal and Ngambri people,
this is their land,
never ceded, always sacred.
And I pay my respects to the elders,
past and present, of this place.
I also acknowledge
that we’re gathering today
in many other places,
and I pay my respects
to the traditional owners and elders
of all those places too.
It means a lot to me
to get to say those words
and to dwell on what they mean and signal.
And to remember that we live in a country
that has been continuously occupied
for at least 60,000 years.
Aboriginal people built worlds here,
they built social systems,
they built technologies.
They built ways to manage this place
and to manage it remarkably
over a protracted period of time.
And every moment any one of us
stands on a stage as Australians,
here or abroad,
we carry with us a privilege
and a responsibility
because of that history.
And it’s not just a history.
It’s also an incredibly rich
set of resources,
worldviews and knowledge.
And it should run through all of our bones
and it should be the story we always tell.
Ultimately, it’s about
thinking differently,
asking different kinds of questions,
looking holistically
at the world and the systems,
and finding other people who want
to be on that journey with you.
Because for me,
the only way to actually think
about the future and scale
is to always be doing it collectively.
And because for me,
the notion of humans in it together
is one of the ways
we get to think about things
that are responsible, safe
and ultimately, sustainable.
Thank you.