Shared Futures: The A.I. Forum

How is AI reshaping the way we live, create, connect, and evolve?

On June 13, Shared Futures: The AI Forum will bring together the cultural architects of our time to explore.

Community-Aligned A.I. Benchmarks

Today’s AI development ecosystem is out of step with what many people want and need. Communities around the world are rightly concerned about their safety, livelihoods, and futures. Without adequate public input, AI developers may continue to pursue goals that are unrelated—or even counter—to the public interest. In order for AI systems to uphold the values and priorities of the public, we must proactively create mechanisms to meaningfully steer the frontier of AI research.

Developers of new AI systems are not adequately incentivized to center the public interest in their work. The proliferation of “AI for good” hackathons and projects demonstrates the AI community’s appetite to apply technical skills in the public interest, but these initiatives regularly fail to deliver durable impacts. Fundamentally, the capabilities prioritized in today’s AI systems are not the types of transformative capabilities that communities and organizations around the world need to make the world a better place.

Starting with the public’s number one goal for a better future—food security—we’re bringing together subject matter experts, community leaders with machine learning researchers to identify concrete challenges and opportunities for impact.

When people develop machine learning models for AI products and services, they iterate to improve performance. 

What it means to “improve” a machine learning model depends on what you want the model to do, like correctly transcribe an audio sample or generate a reliable summary of a long document.

Machine learning benchmarks are similar to standardized tests that AI researchers and builders can score their work against. Benchmarks allow us to both see if different model tweaks improve the performance for the intended task and compare similar models against one another.

Benchmarks are important, but their development and adoption has historically been somewhat arbitrary. The capabilities that benchmarks measure should reflect the priorities for what the public wants AI tools to be and do. 

We can build positive AI futures, ones that emphasize what the public wants out of these emerging technologies. As such, it’s imperative that we build benchmarks worth striving for.

Benchmarks are just one piece of the picture. There’s a whole ecosystem of AI evaluations dedicated to measuring AI systems and how they are used in the real world. Other organizations are already focused on evaluating the impacts of AI models after they are built. To complement this important work, Aspen Digital is approaching this space from early in the development cycle, when the goals and targets for a system are still being set.

Making the Most of the Global Digital Compact

Following adoption of the Global Digital Compact, I made remarks at an informal stakeholder consultation hosted by the United Nations.

This work is made possible thanks to generous support from Siegel Family Endowment.