![]() |
Moderated by Stephen Muggleton. |
Nils J. NilssonTeleo-Reactive Programs and the Triple-Tower Architecture |
The
article
mentioned above has been submitted to the Electronic
Transactions on Artificial Intelligence, and the present page
contains the review discussion. Click here for
more
explanations and for the webpage of theauthor, Nils J. Nilsson.
Overview of interactions
Q1. Anonymous Referee 1 (15.10):
Recommendation: The paper can be accepted for ETAI after minor revisions. No further refereeing round is needed. This paper describes an agent architecture that permits a combination of reactive behaviour and planning. The author refers to related work using triple-tower architectures, notably Albus et al, but he does not make clear the novel contributions of this paper. My understanding is that this is a demonstration of the architecture using teleo-reactive programs. There is also a suggestion that this architecture lends itself to learning and more on this is available on the author's web page, but I'd like to have seen a more self-contained paper here. It would only be necessary to give summary of the ideas. I'd also be curious to know how this architecture would transfer to real robots operating in a non-deterministic world. Although the T-R formalism is described in detail elsewhere, a brief desciption here would be helpful to understanding the example diagrammed. A formatting note: the diagrams did not download correctly (either on Netscape or IE). I had to fetch the original Word document. Only IE correctly reproduces the symbol font characters since these are not HTML standard. So it would be better to have a postscript or PDF version. Q2. Anonymous Referee 2 (15.10):
Recommendation: The paper can be accepted for ETAI after minor revisions. No further refereeing round is needed. The paper is in fact excellent. Revision is only desirable in the sense of clarification or expansion of particular points. 1. Predicates corresponding to primitive percepts are associated with the "model tower". It is therefore confusing to find that the "perceptual tower" is labelled in the drawing with "Sensors". 2. A TMS is made a part of the perceptual and model towers. It is important to know whether each has its own TMS. If "yes", then is there a mediating mechanism for cases where facts stated in one of the models are left unasserted in the other, or where the two fact-bases are inconsistent? If "no", then a single unitary TMS presumably, by definition, maintains consistency everywhere and always. Whether there are two TMS's or one, questions of conflict-resolution are liable to crop up at some level, and deserve mention. 3. The paper does not make it clear whether "reporting actions" can in principle be fitted into the scheme. If so, is the reporting layer to be part of the perceptual tower? Or is it more expedient to have a separate "communications tower"? Mutual consistency of towers Inconsistency, mentioned above, can arise when relations qualifying one and the same object(s) enter the model tower from both perceptual and action towers. Some model assertions are updated directly from sensed environment changes. Others are derived from inferential rules of the perceptual tower. These latter, in a slightly more enriched world, might exploit knowledge of perspective etc. in handling pickup preconditions. Example Suppose that an enrichment of this blocks world allows "wide", "medium" and "narrow" blocks, and corresponding hand actions set-for-wide, set-for-medium and set-for-narrow for selecting in advance an appropriate width between the grippers. Suppose that the robot naively judges the width of an object by the width of the retinal projection. Suppose also that the perceptual tower has additional information, -- whether "given", or deduced or induced, -- that although blocks normally stand front-stage, occasionally one or more starts from a position a little set back, sufficiently for a set-back medium block to subtend the same optical angle as a set-front small block etc. The lower level of the model tower will presumably get populated with width-assertions from direct sensing, causing the robot to act accordingly, sometimes getting stuck through having the wrong gripper-width for the width of the next object to be picked up. One would like some mechanism which passes internal reports of failures to the perceptual layer, which could then perhaps plan up some action-on-failure heuristic by reasoning from everything it now knows. A proactive perceptual tower might even intervene before the start with a proposed revision to the model tower's pick-up routine before finalizing a plan which starts by picking up x:
Something analogous happens in human hand-eye pickup. Inconsistency can exist between the operationally perceived relative widths of two objects as implied by the "pre-grasp" posture of the hand and the epistemically perceived relative widths as measured by asking the subject to rank them. The latter is affected by the various visual illusions, the former not. In this case there are two completely separate visuo-action loops. The fast-track one doesn't use the conscious visual system at all, but a subcortical one left over from reptilian ancestry. I presume that when this primitive loop lets one down, some sort of default occurs to a more deliberative loop involving the visual cortex, among other higher centres. These referee's comments are not to be taken as mandatory. The paper deserves to be published as it stands. The author should only respond to any of this, if at all, according to his own good judgment. The only exception might be an obligation to make the labelling of the Figure clearer. |