The Artificiality of Alignment
AI

The Artificiality of Alignment

This essay first appeared in Reboot. Credulous, breathless coverage of “AI existential risk” (abbreviated “x-risk”) has reached the mainstream. Who could have foreseen that the smallcaps onomatopoeia “ꜰᴏᴏᴍ” — both evocative of and directly derived from children’s cartoons — might show up uncritically in the New Yorker? More than ever, the public discourse about AI […]

An Introduction to the Problems of AI Consciousness
AI

An Introduction to the Problems of AI Consciousness

Once considered a forbidden topic in the AI community, discussions around the concept of AI consciousness are now taking center stage, marking a significant shift since the current AI resurgence began over a decade ago. For example, last year, Blake Lemoine, an engineer at Google, made headlines claiming the large language model he was developing

Andrew Ng: Unbiggen AI – IEEE Spectrum
Technology

Andrew Ng: Unbiggen AI – IEEE Spectrum

Andrew Ng has serious street cred in artificial intelligence. He pioneered the use of graphics processing units (GPUs) to train deep learning models in the late 2000s with his students at Stanford University, cofounded Google Brain in 2011, and then served for three years as chief scientist for Baidu, where he helped build the Chinese

How AI Will Change Chip Design
Technology

How AI Will Change Chip Design

The end of Moore’s Law is looming. Engineers and designers can do only so much to miniaturize transistors and pack as many of them as possible into chips. So they’re turning to other approaches to chip design, incorporating technologies like AI into the process. Samsung, for instance, is adding AI to its memory chips to

Atomically Thin Materials Significantly Shrink Qubits
Technology

Atomically Thin Materials Significantly Shrink Qubits

Quantum computing is a devilishly complex technology, with many technical hurdles impacting its development. Of these challenges two critical issues stand out: miniaturization and qubit quality. IBM has adopted the superconducting qubit road map of reaching a 1,121-qubit processor by 2023, leading to the expectation that 1,000 qubits with today’s qubit form factor is feasible.

Understanding Convolutions on Graphs
AI

Understanding Convolutions on Graphs

Contents This article is one of two Distill publications about graph neural networks. Take a look at A Gentle Introduction to Graph Neural Networks for a companion view on many things graph and neural network related. Many systems and interactions – social networks, molecules, organizations, citations, physical models, transactions – can be represented quite naturally

A Gentle Introduction to Graph Neural Networks
AI

A Gentle Introduction to Graph Neural Networks

This article is one of two Distill publications about graph neural networks. Take a look at Understanding Convolutions on Graphs to understand how convolutions over images generalize naturally to convolutions over graphs. Graphs are all around us; real world objects are often defined in terms of their connections to other things. A set of objects,

Distill Hiatus
AI

Distill Hiatus

Over the past five years, Distill has supported authors in publishing artifacts that push beyond the traditional expectations of scientific papers. From Gabriel Goh’s interactive exposition of momentum, to an ongoing collaboration exploring self-organizing systems, to a community discussion of a highly debated paper, Distill has been a venue for authors to experiment in scientific

Adversarial Reprogramming of Neural Cellular Automata
AI

Adversarial Reprogramming of Neural Cellular Automata

Contents This article is part of the Differentiable Self-organizing Systems Thread, an experimental format collecting invited short articles delving into differentiable self-organizing systems, interspersed with critical commentary from several experts in adjacent fields. Self-Organising Textures This article makes strong use of colors in figures and demos. Click here to adjust the color palette. In a

Weight Banding
AI

Weight Banding

This article is part of the Circuits thread, an experimental format collecting invited short articles and critical commentary delving into the inner workings of neural networks. Branch Specialization Introduction Open up any ImageNet conv net and look at the weights in the last layer. You’ll find a uniform spatial pattern to them, dramatically unlike anything

Scroll to Top