There’s a new sheriff in town, kids.
Justs the other day, I learned of a new field of study, replete with PhD candidates and its own scholarly journal: Its called… wait for it:
“computational neuroscience for AI minds.”
The article was detailing the bizarre near-total opacity of the AI brain and how scientists were working hard to decode it and to understand how it works by using… well, essentially, the crude toolkits of biological neuroscience. [Read: When the AI dreams of spiders…]
What? Why?
Can’t they just open the source
and take a look at the code?
In short: No.
related article: Black Box AI, Opacity, and What it Really Means
UPDATE: 2024.06.27 [via WIRED]
Leading AI labs make progress in figuring out how their foundation AIs actually work
(do you sense the irony in that headline? that’s the very essence of Black Box AI: We seeded it, we spawned it, we trained it, and we have absolutely no idea how its internal logic works)
Looking into the black box? The opacity of neural networks has long presented limitations to understanding how certain AI models really work. While recent years have seen some limited headway in interpretability as a means of better understanding how these systems function, progress has remained limited. However, researchers at Anthropic and OpenAI have both reported recent successes in identifying features of large models that are interpretable by humans.
|
Enter the Black Box of AI
It all makes sense when you begin to comprehend how these modern AIs were spawned. And I do mean “spawned,” not “coded” or “created.” Here’s their origin story, as I understand it, in brief:
For a long time, more than 5 decades in fact (c. 1965-2015), AI candidate programs were “hand coded”… that is to say, their code was written by hand using human-readable computer languages (C++, Java, Python, etc) , by humans, in close consultation by a team of subject matter experts (chess grandmasters, factory automation engineers, linguistics professors, etc.). Sometimes these programs were written on pure theory.
Other times, they were given well-curated datasets to chew on, and to “draw their own conclusions from.” These included, for instance, optimized process flow diagrams, or databases containing what humans believed to be the absolute optimal paths through the first 10 moves of a chess game (opening databases) or the final 30 moves (endgames). But these databases were constrained by both the availability of well annotated datasets, and the expense of storage.
1990: The Internet
Then along came the internet… the end repository of all public and semi-private data (not to mention ostensibly “secure” intranets and extranets, which are private and organizational data repositories using common web data standards and syntax, the total data footprint of which potentially dwarfs what exists on the public internets).
2000: EDI
Then along came Electronic Data Interchange (EDI) and XML, which standardized data schemas, database formats and metadata annotation syntax for data across almost every field of human endeavor. In short, both the quantity, and the quality, the accessibility and the machine-readability of the available training data for any given AI program increased by several orders of magnitude from 1995-2005.
[SIDENOTE: at that same time, many initiatives were started, with the intent of digitizing / annotating the sum product of human creativity — books, music, television, film. The most notable project of this type was Project Gutenberg. Read here for more on how that data ocean feeds the AI.]
All this time Moore’s law was working its magic on storage systems. $500 of consumer hard drive capacity exploded from 20MB (1982) to 500MB (1994) to 500GB (2004) to 20TB (2022) (a 20TB drive stores the equivalent of the contents of one million 20MB drives).
2010: Big Data
And a bit later, along came the concept of Big Data. AS in, really big data. As in, grid-based databases that contained petabytes of structured information. And code didn’t care about size. Its appetite was insatiable. And all we knew was, the more it digested, for the most part, the smarter it got.
So, this brings us to today.
And the reason that computational neuroscience was born.
2020: AI
To the Age of Deep Learning.
To the exa-scale models of Neural Nets.
scientists, realising that these terabyte sized neural nets (token nodegraphs, multidimensional matrices, weights & biases) were in fact total black boxes, opaque to any traditional computer science analysis techniques or error tracking.
Realizing that they were far more like human wetware biological brains than any linear hand-coded computer program. With that realization, thinking: perhaps we can learn something from MRIs, from neuroscientists… region detection, trace mapping, n-dimensional analyses… ah, lets call it… computational neuroscience!
To the Days of the Computational AI Neuroscientists.
At this point, we might wish to ask: