Not enough data to create a plot.
Try a different view from the menu above.
Minai, Ali A.
A Comparative Study of Sentence Embedding Models for Assessing Semantic Variation
Mistry, Deven M., Minai, Ali A.
Analyzing the pattern of semantic variation in long real-world texts such as books or transcripts is interesting from the stylistic, cognitive, and linguistic perspectives. It is also useful for applications such as text segmentation, document summarization, and detection of semantic novelty. The recent emergence of several vector-space methods for sentence embedding has made such analysis feasible. However, this raises the issue of how consistent and meaningful the semantic representations produced by various methods are in themselves. In this paper, we compare several recent sentence embedding methods via time-series of semantic similarity between successive sentences and matrices of pairwise sentence similarity for multiple books of literature. In contrast to previous work using target tasks and curated datasets to compare sentence embedding methods, our approach provides an evaluation of the methods 'in the wild'. We find that most of the sentence embedding methods considered do infer highly correlated patterns of semantic similarity in a given document, but show interesting differences.
DeepCPG Policies for Robot Locomotion
Deshpande, Aditya M., Hurd, Eric, Minai, Ali A., Kumar, Manish
Central Pattern Generators (CPGs) form the neural basis of the observed rhythmic behaviors for locomotion in legged animals. The CPG dynamics organized into networks allow the emergence of complex locomotor behaviors. In this work, we take this inspiration for developing walking behaviors in multi-legged robots. We present novel DeepCPG policies that embed CPGs as a layer in a larger neural network and facilitate end-to-end learning of locomotion behaviors in deep reinforcement learning (DRL) setup. We demonstrate the effectiveness of this approach on physics engine-based insectoid robots. We show that, compared to traditional approaches, DeepCPG policies allow sample-efficient end-to-end learning of effective locomotion strategies even in the case of high-dimensional sensor spaces (vision). We scale the DeepCPG policies using a modular robot configuration and multi-agent DRL. Our results suggest that gradual complexification with embedded priors of these policies in a modular fashion could achieve non-trivial sensor and motor integration on a robot platform. These results also indicate the efficacy of bootstrapping more complex intelligent systems from simpler ones based on biological principles. Finally, we present the experimental results for a proof-of-concept insectoid robot system for which DeepCPG learned policies initially using the simulation engine and these were afterwards transferred to real-world robots without any additional fine-tuning.
Robust Deep Reinforcement Learning for Quadcopter Control
Deshpande, Aditya M., Minai, Ali A., Kumar, Manish
Deep reinforcement learning (RL) has made it possible to solve complex robotics problems using neural networks as function approximators. However, the policies trained on stationary environments suffer in terms of generalization when transferred from one environment to another. In this work, we use Robust Markov Decision Processes (RMDP) to train the drone control policy, which combines ideas from Robust Control and RL. It opts for pessimistic optimization to handle potential gaps between policy transfer from one environment to another. The trained control policy is tested on the task of quadcopter positional control. RL agents were trained in a MuJoCo simulator. During testing, different environment parameters (unseen during the training) were used to validate the robustness of the trained policy for transfer from one environment to another. The robust policy outperformed the standard agents in these environments, suggesting that the added robustness increases generality and can adapt to non-stationary environments. Codes: https://github.com/adipandas/gym_multirotor
Developmental Reinforcement Learning of Control Policy of a Quadcopter UAV with Thrust Vectoring Rotors
Deshpande, Aditya M., Kumar, Rumit, Minai, Ali A., Kumar, Manish
In this paper, we present a novel developmental reinforcement learning-based controller for a quadcopter with thrust vectoring capabilities. This multirotor UAV design has tilt-enabled rotors. It utilizes the rotor force magnitude and direction to achieve the desired state during flight. The control policy of this robot is learned using the policy transfer from the learned controller of the quadcopter (comparatively simple UAV design without thrust vectoring). This approach allows learning a control policy for systems with multiple inputs and multiple outputs. The performance of the learned policy is evaluated by physics-based simulations for the tasks of hovering and way-point navigation. The flight simulations utilize a flight controller based on reinforcement learning without any additional PID components. The results show faster learning with the presented approach as opposed to learning the control policy from scratch for this new UAV design created by modifications in a conventional quadcopter, i.e., the addition of more degrees of freedom (4-actuators in conventional quadcopter to 8-actuators in tilt-rotor quadcopter). We demonstrate the robustness of our learned policy by showing the recovery of the tilt-rotor platform in the simulation from various non-static initial conditions in order to reach a desired state. The developmental policy for the tilt-rotor UAV also showed superior fault tolerance when compared with the policy learned from the scratch. The results show the ability of the presented approach to bootstrap the learned behavior from a simpler system (lower-dimensional action-space) to a more complex robot (comparatively higher-dimensional action-space) and reach better performance faster.
One-Shot Recognition of Manufacturing Defects in Steel Surfaces
Deshpande, Aditya M., Minai, Ali A., Kumar, Manish
Quality control is an essential process in manufacturing to make the product defect-free as well as to meet customer needs. The automation of this process is important to maintain high quality along with the high manufacturing throughput. With recent developments in deep learning and computer vision technologies, it has become possible to detect various features from the images with near-human accuracy. However, many of these approaches are data intensive. Training and deployment of such a system on manufacturing floors may become expensive and time-consuming. The need for large amounts of training data is one of the limitations of the applicability of these approaches in real-world manufacturing systems. In this work, we propose the application of a Siamese convolutional neural network to do one-shot recognition for such a task. Our results demonstrate how one-shot learning can be used in quality control of steel by identification of defects on the steel surface. This method can significantly reduce the requirements of training data and can also be run in real-time.
Graded Attractors: Configuring Context-Dependent Workspaces for Ideation
Minai, Ali A. (University of Cincinnati) | Iyer, Laxmi R. (University of Cincinnati) | Padur, Divyachapan (University of Cincinnati) | Doboli, Simona (Hofstra University) | Brown, Vincent R. (Hofstra University)
Thought is an essential aspect of mental function, but remains very poorly understood. In this paper, we take the view that thought is a response process — the emergent and dynamic configuration of structured response, i.e., ideas, by composing response elements, i.e., concepts, from a repertoire under the influence of afferent information, internal modulation and evaluative feedback. We hypothesize that the process of generating ideas occurs at two levels: 1) The identification of a context-specific subset — or workspace — of concepts from the larger repertoire; and 2) The configuration of plausible/useful ideas within this workspace. Workspace configuration is mediated by a dynamic selector network (DSN), which is an internal attention/working memory system. Each unit of the DSN selectively gates a subset of concepts, so that any pattern of activity in the DSN defines a workspace. The configuration of efficient and flexible workspaces is mediated by dynamical structures termed graded attractors — attractors where the set of active units can be varied in systematic order by inhibitory modulation. A graded attractor in the DSN can project a selective bias — a ``searchlight" — onto the concept repertoire to define a specific workspace, and inhibitory modulation can be used to vary the breadth of this workspace. As it experiences various contexts, the cognitive system can configure a set of graded attractors, each covering a domain of similar contexts. In this paper, we focus on a mechanism for configuring context-specific graded attractors, and evaluate its performance over a set of contexts with varying degrees of similarity. In particular, we look at whether contexts are clustered appropriately into a minimal number of workspaces based on the similarity of the responses they require. While the focus in this paper is on semantic workspaces, the model is broadly applicable to other cognitive response functions such as motor control or memory recall.