To begin to see how LT works, we start with a scenario that is close to home for Bringsjord and Ferrucci, given their sustained efforts to build story generation agents: Assume that Jones, a human AInik, attempts to build an artificial computational agent A that doesn't engage in conversation, but rather creates stories -- creates in the Lovelacean sense that this system originates stories. Assume that Jones activates A and that a stunningly belletristic story o is produced. We claim that if Jones cannot explain how o was gemerated by A, and if Jones has no reason whatever to believe that A succeeded on the strength of a fluke hardware error, etc. (which entails that A can produce other equally impressive stories), then A should at least provisionally be regarded genuinely creative. An artificial computational agent passes LT if and only if it stands to its creator as A stands to Jones.
LT relies on the special epistemic relationship that exists between Jones and A. But `Jones,' like `A,' is of course just an uninformative variable standing in for any human system designer. This yields the following rough-and-ready definition.
Notice that LT is actually what might be called a meta-test. The idea is that this scheme can be deployed for any partcular domain. If conversation is the kind of behavior wanted, then merely stipulate that o is an English sentence (or sequence of such sentences) in the context of a converation (as in, of course, TT). If the production of a mathematical proof with respect to a given conjecture is what's desired, then we merely set o to a proof. In light of this, we can focus LT on the particular kind of interaction appropriate for the digital entertainment involved.
Obvious questions arise at this point. Three are:
The answer to the third question is easy: H can have as long as he or she likes, within reason. The proffered explanation doesn't have to come immediately: H can take a month, months, even a year or two. Anything longer than a couple of years strikes us as perhaps unreasonable. We realize that these temporal parameters aren't exactly precise, but then again we should not be held to standards higher than those pressed against Turing and those who promote his test and variants thereof.10 The general point, obviously, is that H should have more than ample time to sort things out.
But what about Q1 and Q2? The answer to Q1 is that H is assumed to have at her disposal knowledge of the architecture of the agent in question, knowledge of the KB of the agent, knowledge of how the main functions in the agent are implemented (e.g., how TELL and ASK are implemented), and so on (recall the summary of intelligent agents above). H is also assumed to have resources sufficient to pin down these elements, to ``freeze" them and inspect them, and so on. I confess that this isn't exactly precise. To clarify things, I offer an example. This example is also designed to provide an answer to Q2.
To fix the context for the example, suppose that the output from our
artificial agent A' is a resolution-based proof which settles a problem
which human mathematicians and logicians have grappled unsuccessfully with
for decades. This problem, suppose, is to determine whether or not some
formula
can be derived from some (consistent) axiom set
.
Imagine that after many years of fruitless deliberation, a human H'
encodes
and
and gives both to OTTER (a well-known
theorem prover; it's discussed in (Bringsjord & Ferrucci, 2000)), and OTTER produces a proof
showing that this encoding is inconsistent, which establishes
,
and leads to an explosion of commentary in the media about
``brilliant" and ``creative" machines, and so on.11
In this case, A' doesn't pass LT. This is true because H, knowing
the KB, architecture, and central functions of A' will be able to give a
perfect explanation for the behavior in question.
I routinely give
explanations of this sort. The KB is simply the encoding of
,
the architecture consists in the search algorithms used by OTTER, and the main functions consist in the rules of inference used in a
resolution-based theorem prover.
Here, now, given the foregoing, is a better definition: