What Is the Depth of Expertise of an AI Training Dataset?
| |

What Is the Depth of Expertise of an AI Training Dataset?

I use “depth of expertise” as a data quality dimension of AI training datasets. It describes how much a dataset reflects of expertise in a knowledge domain. This is not a common data quality dimension used in other contexts, and I haven’t seen it as such in discussions of, say, quality of data used for…

Black Box Approach to AI Governance
| |

Black Box Approach to AI Governance

As currently drafted (2024), the Algorithmic Accountability Act does not require the algorithms and training data used in an AI System to be available for audit. (See my notes on the Act, starting with the one here.) The way that an auditor learns about the AI System is from documented impact assessments, which involve descriptions…

Can Opacity Be Solved in an AI Derived from an LLM?
| |

Can Opacity Be Solved in an AI Derived from an LLM?

The short answer is “No”, and the reasons for it are interesting. An AI system is opaque if it is impossible or costly for it (or people auditing it) to explain why it gave some specific outputs. Opacity is undesirable in general – see my note here. So this question applies for both those outputs…

Why Specialized AI Should Be Certified by Expert Communities
|

Why Specialized AI Should Be Certified by Expert Communities

Should the explanations that an Artificial Intelligence system provides for its recommendations, or decisions, meet a higher standard than explanations for the same, that a human expert would be able to provide? I wrote separately, here, about conditions that good explanations need to satisfy. These conditions are very hard to satisfy, and in particular the…

AI Compliance at Scale via Embedded Data Governance
| |

AI Compliance at Scale via Embedded Data Governance

There are, roughly speaking, three problems to solve for an Artificial Intelligence system to comply with AI regulations in China (see the note here) and likely future regulation in the USA (see the notes on the Algorithmic Accountability Act, starting here):  Using available, large-scale crawled web/Internet data is a low-cost (it’s all relative) approach to…

Can an Artificial Intelligence Trained on Large-Scale Crawled Web Data Comply with the Algorithmic Accountability Act?
| | | |

Can an Artificial Intelligence Trained on Large-Scale Crawled Web Data Comply with the Algorithmic Accountability Act?

If an artificial intelligence system is trained on large-scale crawled web/Internet data, can it comply with the Algorithmic Accountability Act?  For the sake of discussion, I assume below that (1) the Act is passed, which it is not at the time of writing, and (2) the Act applies to the system (for more on applicability,…

Does the EU AI Act apply to most software?
| | |

Does the EU AI Act apply to most software?

Does the EU AI Act apply to most, if not all software? It is probably not what was intended, but it may well be the case.  The EU AI Act, here, applies to “artificial intelligence systems” (AI system), and defines AI systems as follows: ‘artificial intelligence system’ (AI system) means software that is developed with…

What is AI Governance for?
|

What is AI Governance for?

If an AI is not predictable by design, then the purpose of governing it is to ensure that it gives the right answers (actions) most of the time, and that when it fails, the consequences are negligible, or that it can only fail on inconsequential questions, goals, or tasks.