The New Foundation Models of AI Lack “Responsible” Paperwork

 If you’ve seen photos of a teapot shaped like avocado or read a well-written article that veers off on slightly weird tangents, you may have been exposed to a new trend in AI. Machine learning systems called DALL-E, GPT and PaLM are making a splash with their incredible ability to generate creative work. These systems are known as “foundation models” and are not all hype and party tricks.

Foundation models are models trained on broad data (generally using self-supervision at scale) that can be adapted to a wide range of downstream tasks. These models, which are based on standard ideas in transfer learning and recent advances in deep learning and computer systems applied at a very large scale, demonstrate surprising emergent capabilities and substantially improve performance on a wide range of downstream tasks. Given this potential, foundation models are seen as the subject of a growing paradigm shift, where many AI systems across domains will directly build upon or heavily integrate foundation models.

Foundation models incentivize homogenization: the same few models are repeatedly reused as the basis for many applications. Such consolidation is a double-edged sword: centralization allows researchers to concentrate and amortize their efforts (e.g., to improve robustness, to reduce bias) on a small collection of models that can be repeatedly applied across applications to reap these benefits (akin to societal infrastructure), but centralization also pinpoints these models as singular points of failure that can radiate harms (e.g., security risks, inequities) to countless downstream applications.

Comments

Popular posts from this blog

Agile Model: Accelerating AI Projects with Agile Methodologies

Individuals and interactions over processes and tools