Abstract |
Autonomous systems are expected to have a major impact in future coalition operations. These systems are enabled by a variety of Artificial Intelligence (AI) learning algorithms that contextualize and adapt in varying, possibly unforeseen situations to assist humans in achieving complex tasks. Moreover, these systems will be required to operate in dynamic and challenging environments that impose certain constraints such as task formation and collaboration, ad-hoc resource availability, rapidly changing environmental conditions and the requirement to abide by mission objectives. Therefore, such systems require the capability to adapt and evolve so that they can behave autonomously at the edge of the network in new situations. Crucially, autonomous systems have to understand the bounds in which they can operate based on their own capability and the constraints of the environment. Policies are typically used by systems to define their behavior and constraints and often these policies are manually configured and managed by humans. AI-enabled systems will require novel approaches to rapidly learn, create, augment, and model emerging policies to support real-time decision making. Recent work has shown that such policy model generations are possible through symbolic learning to shallow and deep learning approaches for different classes of problems. Motivated by this observation, in this paper, we propose to apply recent advances in explainable-AI to develop an approach which is agnostic to the learning algorithm, thus enabling seamless policy generation in the coalition environment. |