1st post - be as gentle as you feel the need to be :-) https://open.substack.com/pub/blackerthanmirrors/p/exploring-the-ai-commons-questions?r=5v4urt&utm_medium=ios
It was a real pleasure imagining #AICommons as alternatives to extractive AI together with @rumya, Michelle Thorne and Solana Larsen at this year's @republica
A session co-instigated by Renata Avila from
@okfn and my colleague Deshni Govender from GIZ's FAIR Forward.
How can we build sustainable, participatory #AICommons?
Review the recording of the session at @republica where @rumya, @danbru8, Michelle Thorne and Solana Larsen discussed why data work matters in shaping alternatives to extractive AI.
#AICommons, one of 15 proposals for #SommetActionAI deliverables, aims to boost citizen participation in AI governance for more inclusive and equitable AI development.
Stay updated by following our #DigitalCommons for Public AI work: https://openfuture.eu/our-work/ai-and-the-commons/
Open Future participated in a consultation by The Future Society that gathered 200 leading voices on AI governance, and we are happy to see our ideas for #AICommons recognized as key proposals for the Summit.
Check the key findings on #AIGovernance: https://thefuturesociety.org/aiactionsummitconsultationreport/
"Transparency around the content used to train AI and information about how it was processed can support the legitimate interests of preventing discrimination and respecting cultural diversity." Learn more in this article by Maximilian Gahntz & Zuzanna Warso, published in Tech Policy Press.
https://www.techpolicy.press/how-the-eu-ai-act-can-increase-transparency-around-ai-training-data/
@mozilla and Open Future co-hosted an event that brought together policymakers, industry representatives and civil society.
The goal was to exchange views on what transparency for content used to train AI models should look like.
Read more: https://openfuture.eu/event/from-code-to-conduct-insights-from-a-mozilla-morning/
"The Open Source #AI Definition is an important step in defining the standard of openness in AI development. Still, it should be seen as just one position in a broader debate that needs to bridge positions of AI developers with those of other stakeholders."
– read our new analysis of the Open Source Initiative (OSI)'s definition of open source AI, by @tarkowski and @paulk
"Archival Images of AI", great new project from Sound & Vision.
It starts with a simple premise: you can usage heritage collections to make images of AI that are better than the ever-present illustrations of humanoid robots.
And turns out to be an opportunity to ask serious questions about heritage, the commons, and AI.
https://aixdesign.co/posts/archival-images-of-ai
#aicommons #ai #DigitalHeritage
Dan Cohen and Dave Hansen wrote recently a really good piece on books, libraries and AI training (the piece refers to the paper on Books Data Commons that I co-authored).
They start with a well-known argument about levelling the field: without offering public access to training resources, AI monopolies will benefit from information asymmetries. Google already has access to 40 million scanned books.
They add to this a key point about libraries' public interest stance - and suggest that libraries could actively govern / gatekeep access to books.
This reminds me of the recent paper by Melanie Dulong de Rosnay and Yaniv Benhamou, which for me is groundbreaking - it proposes that license-based approaches to sharing are combined with trusted institutions that offer more fine-grained access governance.
So it's good to see that this line of thinking is getting traction.
https://www.authorsalliance.org/2024/05/13/books-are-big-ais-achilles-heel/
Interesting data from a new edition of the Foundation Model Transaprency Index - collected six months after the initial index was released.
Overall, there's big improvement, with average score jumping from 37 to 58 point (out of a 100). That's a lot!
The interesting fact is that researchers contacted developers and solicited data - interactions count.
More importantly, there is little improvement, and little overall transparency in a category that researchers describe as "upstream": on data, labour and compute that goes into training. And "data access" gets the lowest score of all the parameters.
More at Tech Policy Press: https://www.techpolicy.press/the-foundation-model-transparency-index-what-changed-in-6-months/
The Think7 Italy Summit is happening this week, with the theme “The G7 and the World: Rebuilding Bridges”.
We have been invited to write a brief on “Democratic governance of AI systems and datasets”, which will be presented tomorrow by @tarkowski .
The brief has been a joint effort of three organizations: Open Future Foundation, Centro Politiche Europee and MicroSave Consulting (MSC), with contributions from Renata Avila, Lea Gimpel, and @savi.
https://think7.org/event/t7-italy-summit-the-g7-and-the-world-rebuilding-bridges/
Open Future's newest white paper, authored by @zwarso and myself, addresses the governance of data sets used for #AI training.
Over the past two years, it has become evident that shared datasets are necessary to create a level playing field and support AI solutions in the public interest. Without these shared datasets, companies with vast proprietary data reserves will always have the winning hand.
However, data sharing in the era of AI poses new challenges. Thus, we need to build upon established methods like #opendata refining them and integrating innovative ideas for data governance.
Our white paper proposes that data sets should be governed as commons, shared and responsibly managed collectively. We outline six principles for commons-based governance, complemented by real-life examples of these principles in action.
https://openfuture.eu/publication/commons-based-data-set-governance-for-ai/
I participated yesterday in an expert workshop on Public-Private Partnerships in Global Data Governance, organized by the United Nations University Centre for Policy Research (UNU-CPR) and the International Chamber of Commerce (ICC).
I was also invited to prepare a policy brief that presented how the Public Data Commons model, which we have been advocating for, could be applied at global level for dealing with emergencies, and the broader poly-crisis.
It is exciting to see UNU explore data sharing policies within the context of the policy debate on the UN Global Digital Compact.
Worth noting is also the recent report of the High-Level Advisory Board on Effective Multilateralism, "A Breakthrough for People and Planet". One of the transofrmative shifts, "the just digital transition", includes a recommendation for a global data impact hub.
In my brief, I show how this impact hub could be designed as a Public Data Commons. I also highly recommend other briefs presented at the event, by Alex Novikau, Isabel Rocha de Siqueira, Michael Stampfer and Stefaan Verhulst.
#aicommons #datacommons #datagovernance #ai
You can find the report and all the briefs on the UNU webpage: https://unu.edu/cpr/project/breakthrough-people-and-planet
In a month (7-8 December) I will be speaking at a conference on data governance and AI, organized in Washington, DC by the Digital Trade and Data Governance Hub. I am excited about this for two reasons:
first of all, we need to connect the policy debates on data governance and AI governance. The space of AI development offers new opportunities to develop, at scale, commons-based approaches that have been much theorized and advocated for, but not yet implemented.
and secondly, I am a deep believer in dialogue between the US and the EU. US is leading in terms of AI development itself, while EU will most probably be the first country to innovate in terms of AI regulation.
Please consider joining, either in-person or remotely (it's a hybrid event).
#aicommons #datacommons #datagovernance #ai
https://www.linkedin.com/events/datagovernanceintheageofgenerat7127306901125521408/comments/
Our October newsletter is out, with updates on our #AIcommons and #digitalpublicspace work. I'm especially proud of several publications that expand our policy ideas on Digital Public Space - check them out here: https://mailchi.mp/openfuture/digital_public_space_explained
The Chan-Zuckerberg Initiative announced that in order to support non-profit medical research they are building "computing infrastructure" - that is, purchasing over a 1000 state of the art GPUs.
This is super interesting, in an AI-powered world compute is not a commodity, but a currency.
So if a private foundation can do it, why can't governments do the same? Seems that providing public interest compute infrastructure is one of the simpler move that can be made, as the comples governance issues are solved in parallel.
New piece from @halcyene and Michael Birtwistle from Ada Lovelace argues for a more inclusive UK #AI Safety Summit.
The reason for this, they argue, is that "AI safety" is a very broad category. And since many risks are socio-technical, the governance debate needs to include the society, especially those affected by risk. "Nothing about us without us".
It's interesting to observe how UK-based civic actors are attempting to pry open a policy platform that currently is designed as a conversation between business and the state (with a sprinkling of just a few, selected, civic / academic actors). I hope it's succesful and sets a precedent.
And I like the way Ada Lovelace frames risks, and highlights that there are structural harms, risk of market concentration in particular.
This risk is often ignored, and it's the one that can be addressed by policies that support open, commons-based governance of AI.
Also, it's a risk that - since it's structural - affects the policy debate itself: there is a risk of regulatory capture by the largest players, in whose corporate hands power is concentrated. One more reason to make the AI policy debate more inclusive.
Next week, @opensource is running a series of webinars on open source #AI / #ML. Together with @zwarso we will be kicking off the series with a talk on the importance of data governance, and treating datasets as commons. #aicommons
https://opensource.org/events/deep-dive-ai-webinar-series-2023/