Knowledge-infused Learning

Friday, March 25, 2022 - 10:30 am
Seminar Room, AI Institute, 5th Floor

DISSERTATION DEFENSE

(will take place in hybrid fashion; both physical and virtual)

Author : Manas Gaur

Advisor : Dr. Amit P. Sheth

Date : March 25, 2022

Time 10:30 am

Location : Seminar Room, AI Institute, 5th Floor,

    1112 Greene Street (Science and Technology Building)
     Columbia, South Carolina-29208

Virtual Zoom Link

              Meeting ID: 844 013 9296
         Passcode: 12345

Abstract:

In DARPA’s view on three waves of AI, the first wave of AI termed symbolic AI, focused on explicit knowledge. The current second wave of AI is termed statistical AI. Deep learning techniques have been able to exploit large amounts of data and massive computational power to improve human levels of performance in narrowly defined tasks. Separately, knowledge graphs emerged as a powerful tool to capture and exploit an extensive amount and variety of explicit knowledge to make algorithms better understand the content, and enable the next generation of data processing, such as in semantic search. After initial hesitancy about the scalability of the knowledge creation process, the last decade has seen significant growth in developing and applying knowledge, usually in the form of knowledge graphs (KG). Examples range from the use of DBPedia in IBM’s Watson to Google Knowledge Graph in Google Semantic Search, and the application of ProteinBank in AlphaFold, recognized by many as the most significant AI breakthrough, as well as numerous domain-specific knowledge have been applied in improving AI methods in diverse domains such as medicine and healthcare, finance, manufacturing, and defense.


Now, we herald towards the third wave of AI built on what is termed as the neuro-symbolic approach that combines the strengths of statistical and symbolic AI. Combining the respective powers and benefits of using knowledge graphs and deep learning is particularly attractive. This has led to the development of an approach we have called knowledge-infused (deep) learning. This dissertation will advance the currently limited forms of combining the knowledge graphs and deep learning, called shallow and semi-infusion, with a more advanced form called deep-infusion, that will support stronger interleaving of more variety of knowledge at different levels of abstraction with layers in a deep learning architecture.

This dissertation will investigate the knowledge-infusion strategy in two important ways. The first is to infuse knowledge to make any classification task explainable. The second is to achieve explainability in any natural language generation tasks. I will demonstrate the effective strategies of knowledge infusion that bring five characteristic properties in any statistical AI model: (1) Context Sensitivity, (2) Handling Uncertainty and Risk, (3) Interpretable in learning, (4) User-level Explainability, and (5) Transferability across natural language understanding (NLU) tasks. Along with proven methodological contributions in AI made by the dissertation, I will show their applications for open-domain and close-domain NLU.
Furthermore, the dissertation will showcase the utility of incorporating diverse forms of knowledge: linguistic, commonsense, broad-based, and domain-specific. As the dissertation illustrates the success in various domains, achieving state-of-the-art in specific applications, and significant contributions towards improving the state of machine intelligence, I will walk through careful steps to prevent errors arising due to knowledge infusion. Finally, for future directions, I will discuss two exciting areas of research where knowledge infusion would be pivotal to propel machine understanding.

Concurrent identification, characterization, and reconstruction of Protein structure and mixed-mode dynamics from rdc data using redcraft 

Monday, March 21, 2022 - 09:30 am

DISSERTATION DEFENSE

Author : Hanin Omar

Advisor : Dr. Homayoun Valafar

Date : March 21, 2022

Time: 9:30 am

Place: Virtual Teams Link

Abstract

A complete understanding of the structure-function relationship of proteins requires an analysis of their dynamic behaviors in addition to the static structure. However, all current approaches to the study of dynamics in proteins have their shortcomings. A conceptually attractive and alternative approach simultaneously characterizes a protein's structure and its intrinsic dynamics⁠⁠. Ideally, such an approach could solely rely on RDC data-carrying both structural and dynamical information. The major bottleneck in the utilization of RDC data in recent years has been attributed to a lack of RDC analysis tools capable of extracting the pertinent information embedded within this complex source of data.  

Here we present a comprehensive strategy for structure calculation and reconstruction of discrete state dynamics from RDC data based on the SVD method of order tensor estimation. In addition to structure determination, we provide a mechanism of producing an ensemble of conformations for the dynamical regions of a protein from RDC data. The developed methodology has been tested on simulated RDC data with ±1Hz of error from an 83 residue α protein (PDB ID 1A1Z). In nearly all instances, our method reproduced the protein structure, including the conformational ensemble, to within less than 2Å. Based on our investigations, arc motions with more than 30° of rotation are recognized as internal dynamics and are reconstructed with sufficient accuracy.  Furthermore, states with relative occupancies above 20% are consistently recognized and reconstructed successfully. Arc motions with a magnitude of 15° or relative occupancy of less than 10% are consistently unrecognizable as dynamical regions within the context of ±1Hz of error.  

 We also introduce a computational approach named REDCRAFT that allows for uncompromised and concurrent characterization of protein structure and dynamics. We have subjected DHFR (PDB-ID 1RX2), a 159-residue protein, to a fictitious but plausible, mixed-mode internal dynamics model. In this simulation, DHFR was segmented into 7 regions. The two dynamical and rigid-body segments experienced an average orientational modification of 7˚ and 12˚, respectively. Observable RDC data for backbone C'-N, N-H, and C'-H were generated from 102 frames that described the molecular trajectory. The Dynamic Profile generated by REDCRAFT allowed for the recovery of individual fragments with bb-rmsd of less than 1Å and the identification of different dynamical regions of the protein. Following the recovery of fragments, structural assembly correctly assembled the four rigid fragments with respect to each other, categorized the two domains that underwent rigid-body dynamics, and identified one dynamical region for which no conserved structure can be defined. In conclusion, our approach successfully identified dynamical domains, recovery of structure where it is meaningful, and relative assembly of the domains when possible.  

How Smart City Infrastructure & Blockchain can Reduce Harmful Vehicle Emissions

Friday, March 18, 2022 - 02:20 pm
Swearingen Engineering Center in Room 2A31

Abstract

In 2020, Dr. Amari N. Lewis had the opportunity to conduct research at Aalborg University in Denmark. This talk on how smart city infrastructure and Blockchain can reduce harmful vehicle emissions revisits the research discoveries from Dr. Lewis' time in Denmark. As a result, we discovered two methodologies to reduce harmful vehicle emissions through the use of technology and were able to validate our theory through simulation. The first is a Blockchain Emissions Trading System (B-ETS). The second method involves traffic theory, smart infrastructure and dissuasion methods to improve air quality in residential areas.

 

Bio

Amari N. Lewis is currently a Postdoctoral scholar in the Department of Computer Science & Engineering at the University of California San Diego. Her current research is in the area of Computer Science Education studying retention and experiences of students especially focused on students from marginalized populations. Amari earned her PhD in Computer Science from the University of California, Irvine where her research focused on technological advancement in transportation systems. In the last year of her PhD she conducted research in Denmark at Aalborg University. At Aalborg University the research focused on the use of Blockchain in smart infrastructure.

 

Location:

In person

Swearingen Engineering Center in Room 2A31

 

Virtual MS Teams

Time

2:20-3:10pm

Graph Neural Network and Phylogenetic Tree Construction  

Friday, March 18, 2022 - 10:00 am

DISSERTATION DEFENSE

  

Author : Gaofeng Pan

Advisor : Dr. Jijun Tang

Date : March 18, 2022

Time 10:00 am

Place : Virtual (Teams link below)

 

Meeting Link

 

Abstract

Deep Learning had been widely used in computational biology research in past few years.  A great amount of deep learning methods was proposed to solve bioinformatics problems, such as gene function prediction, protein interaction classification, drug effects analysis, and so on; most of these methods yield better solutions than traditional computing methods.  However, few methods were proposed to solve problems encountered in evolutionary biology research. In this dissertation, two neural network learning methods are proposed to solve the problems of genome location prediction and median genome generation encountered in phylogenetic tree construction; the ability of neural network learning models on solving evolutionary biology problems will be explored. 

Phylogenetic tree represents the evolutionary relationships among genomes in intuitive ways, it could be constructed based on genomics phenotype and genomics genotype. The research of phylogenetic tree construction methods is meaningful to biology research. In the past decades, many methods had been proposed to analyze genome functions and predict genome subcellular locations in cell. However, these methods have low accuracy on multi-subcellular localization. In order to improve prediction accuracy, a neural network learning model is defined to extract genome features from genome sequences and evolution position matrix in this research. Experiment results on two widely used benchmark datasets show that this model has significant improvements than other currently available methods on multi-subcellular localization; deep neural network is effective on solving genome location prediction.  

Phylogenetic tree construction based on genomics genotype has more accurate results than construction based on genomics phenotype. The most famous phylogenetic tree construction framework utilizes median genome algorithms to filter tree topology structure and update phylogenetic ancestral genome. Currently, there are several median genome algorithms which could be applied on short genome and simple evolution pattern, however when genome length becomes longer and evolution pattern is complex these algorithms have unstable performance and exceptionally long running time. In order to lift these limitations, a novel median genome generator based on graph neural network learning model is proposed in this research. With graph neural network, genome rearrangement pattern and genome relation could be extracted out from internal gene connection. Experiment results show that this generator could obtain stable median genome results in constant time no matter how long or how complex genomes are; its outstanding performance makes it the best choice in GRAPPA framework for phylogenetic tree construction. 

CASY 2.2

Friday, March 18, 2022 - 08:00 am
1112 Greene St, Columbia, South Carolina 29208

CASY 2.2 is a hybrid physical/virtual conference on the theme of collaborative assistants for the society. The conference will take place on March 18, 2022 virtually over Zoom and physically, in a limited socially distant manner, on the campus on University of South Carolina, Columbia, SC, USA. The event will be free-to-attend and is intended to promote the ethical usage of digital assistants in society for daily life activities.

Please check out the speakers and register using Eventbrite here. Also, see the schedule.

Automata-theoretic approaches to planning in robotics

Wednesday, March 16, 2022 - 01:00 pm
Storey Innovation Center 2267, 550 Assembly St.

DISSERTATION DEFENSE

Automata-theoretic approaches to planning in robotics: combinatorial filter minimization, planning to chronicle, temporal logic planning with soft specifications, and sensor selection for detecting deviations from a planned itinerary

 

Author : Hazhar Rahmani

Advisor : Jason O'Kane

Date : March 16, 2022

Time: 1:00 pm

Place: Storey Innovation Center 2267, 550 Assembly St.

Virtual (Zoom link): https://us02web.zoom.us/j/83006866662

 

Abstract

In this dissertation, we present a collection of new planning algorithms that enable robots to achieve complex goals, beyond simple point-to-point path planning, using automata-theoretic methods, and we consider the filter minimization (FM) problem and a variant of it, filter partitioning minimization (FPM) problem, which aims to minimize combinatorial filters, used for filtering and automata-theoretic planning in systems with discrete sensor data. We introduce a new variant of bisimulation, compatibility, and using this notion we identify several classes of filters for which FM or FPM is solvable in polynomial time, and propose several integer linear programming (ILP) formulations of FM and FPM. Then, we consider a problem, planning to chronicle, in which a robot is tasked with observing an uncertain time-extended process to produce a `chronicle’ of occurrent events that meets a given specification. This problem is useful in applications where we deploy robots to autonomously make structured videos or documentaries from events occurring in an unpredictable environment. Next, we study two variants of temporal logic planning in which the objective is to synthesize a trajectory that satisfies an optimal selection of soft constraints while nevertheless satisfying a hard constraint expressed in linear temporal logic (LTL). We also extend planning to chronicle with the idea of this problem. Then, we consider the problem of planning where to observe the behavior of an agent to ensure that the agent’s execution within the environment fits a pre-disclosed itinerary. This problem arises in a range of contexts including in validating safety claims about robot behavior, applications in security and surveillance, and for both the conception and the (physical) design and logistics of scientific experiments.

Empowering translational data science through large-scale data harmonization

Friday, March 4, 2022 - 02:20 pm

Abstract

Data silos are a major problem in biomedical research. Our data commons and ecosystems framework helps to break down data silos and empower translational research. To do this, data must be harmonized through standardized workflows, references, annotations, and/or mapped to standardized ontologies. I helped lead the design and implementation of a petabyte-scale “omics” data harmonization system utilized in the National Cancer Institute’s Genomic Data Commons. I present some background in data commons, describe our automation system, and provide some highlights of other work my colleagues and I are doing in the Center for Translational Data Science at the University of Chicago. Our work showcases the importance of interdisciplinary collaboration and exciting opportunities for computer science, data science, and biomedical research.

 

Bio

Kyle Hernandez, Ph.D. is a Research Associate Professor of Medicine in the Section of Biomedical Data Science, a Co-PI of the VA Data Commons, and a Manager of Bioinformatics in the Center for Translational Data Science at the University of Chicago. He was a key contributor to the design and development of the large-scale workflow automation system used in the National Cancer Institute Genomic Data Commons (GDC) as well as the development of many of the workflows it uses. His research interests include workflow engines, the genetic architecture of complex phenotypes, the integration of multiple data types, reproducibility, and more recently the complex issues of EHR data curation and extraction. He earned his Ph.D. in Ecology, Evolution, and Population Biology at Purdue University and was an NSF Postdoctoral Fellow at the University of Texas. He joined the University in 2013 and has been affiliated with the Center since 2016.

 

Location:

In person

Swearingen Engineering Center in Room 2A31 

Virtual MS Teams

ACM Code-A-Thon

Friday, February 25, 2022 - 07:00 pm

The bi-annual Association of Computing Machinery Code-A-Thon is this week, February 25th-26th. It is a competition open to current and former USC students and alumni, regardless of field of study. It is divided into four divisions: 145, 146, 240, and 350. Challenges align with the concepts taught at the corresponding CSCE levels. For example lower-level division problems ask participants to use data structures or specific edge-case checking; upper-division might ask participants to process data in O(log n) time.

The ACM Code-A-Thon is a great social event for students to congregate while enjoying pizza, soft drinks, collaboration, and honestly, coding outside of class. It’s also rigorous. The social event starts on Feb 25th at 7pm and lasts for an hour until 8pm when the competition actually starts. The competition lasts for 24 hours. Each division has about four to seven problems with varying point totals that reflect the difficulty of the problem. Points are awarded by the number of test cases a submitted solution passes. Winners are determined by the most points awarded; in the event of a tie at any particular level the tie-break goes to the person who completed the event fastest. 

To make the event more enticing to the public gift cards are awarded to the first, second, and third place finishers at each level to the tune of $100, $75, and $50. Also, in the past instructors have awarded extra credit points to people who participated in the event. ACM will help verify by keeping an attendance sheet with the names of participants and for which instructor they want extra credit. It is ACMs suggestion that should any instructor offer extra credit, they tie the extra credit to points earned by the competing student to avoid a situation where dozens of students sign up to compete but don’t attempt a solution.

Again, this is meant to be collaborative. There will be channels for participants to ask questions and discuss potential solutions, or just throw stuff on the way and see what sticks. We encourage discussion.

Please reach out with any questions you may have about the Code-A-Thon. Blake Seekings (seekingj@email.sc.edu) is the best person to contact.

Students who wish to attend please fill out this form, or get more information here.

Towards Robot Learning for the Real World

Friday, February 25, 2022 - 02:30 pm
Swearingen Engineering Center in Room 2A31

Abstract

Even with the substantial progress we have seen in Robot Learning, we are nowhere near general purpose robots that can operate in the real world that we live in. There are two fundamental reasons for this. First, robots need to build concise representations in high-dimensional sensory observations often without access to explicit sources of supervision. Second, unlike standard supervised learning, they will need to solve long-horizon decision making problems. In this talk, I'll propose a recipe for general purpose robot learning that combines ideas of self-supervision for representation learning with ideas in RL, adaptation, and imitation for decision making.

 

Bio

Lerrel Pinto is an Assistant Professor of Computer Science at NYU. His research interests focus on machine learning and computer vision for robots. He received a PhD degree from CMU in 2019; prior to that he received an MS degree from CMU in 2016, and a B.Tech in Mechanical Engineering from IIT-Guwahati. His work on large-scale robot learning received the Best Student Paper award at ICRA 2016 and a Best Paper finalist award at IROS 2019. Several of his works have been featured in popular media such as The Wall Street Journal, TechCrunch, MIT Tech Review, Wired, and BuzzFeed among others. His recent work can be found on www.lerrelpinto.com.

 

Location:

In person

Swearingen Engineering Center in Room 2A31

 

Virtual MS Teams

Time

2:20-3:10pm

 

Deep Learning Based Generative Materials Design

Monday, February 21, 2022 - 12:00 pm

DISSERTATION DEFENSE

Author : Yong Zhao

Advisor : Dr. Jianjun Hu

Date : Feb 21, 2022

Time 12:00 pm

Place : Virtual (Zoom link below)

https://us02web.zoom.us/j/9031735360

 

Abstract

Discovery of novel functional materials is playing an increasingly important role in many key industries such as lithium batteries for electric vehicles and cell phones. However experimental tinkering of existing materials or Density Functional Theory (DFT) based screening of known crystal structures, two of the major current materials design approaches, are both severely constrained by the limited scale (around 250,000 in ICSD database) and diversity of existing materials and the lack of sufficient number of materials with annotated properties. How to generate a large number of physically feasible, stable, and synthesizable crystal materials and build accurate property prediction models for screening are the two major unsolved challenges in modern materials science.

This dissertation is focused on addressing these two fundamental tasks in material science using deep learning/machine learning models. Deep learning and machine learning have made tremendous progress in computer vision and natural language processing as shown by the autonomous driving cars and google translators, with their potential to greatly transform the research of materials science. Compared to conventional tinkering based materials discovery methods, data-driven approaches have been increasingly used in material informatics because of their much faster screening speed for new materials compared to DFT based calculations. In this dissertation, we design and develop novel deep learning-based algorithms to generate new crystal structures learning the hidden intricate chemical rules from known crystals that assemble atoms into stable crystal structures. We also explore and develop novel representation learning upon materials compositions and structures for high performance prediction of materials structural characteristics and elastic properties.

In the first topic, we propose CubicGAN, a generative adversarial network (GAN) based deep neural network model for large-scale generative design of novel cubic materials. When trained on 375 749 ternary materials from the OQMD database, we show that the model can not only rediscover most of the currently known cubic materials but also generate hypothetical materials of new structure prototypes. A total of 506 such materials have been verified by DFT based phonon dispersion calculation. Our technique allows to generate tens of thousands of new materials given sufficient computing resource.

In the second topic, we propose a Physics Guided Crystal Generative Model (PGCGM) for new materials generation, which significantly expands the structural scope of CubicGAN by bringing the capability of generate crystals of 20 space groups. This is achieved by capturing and exploiting the pairwise atomic distance constraints among neighbor atoms and symmetric geometric constraints and a novel data augmentation strategy using the base atom sites of materials. With atom clustering and merging on generated crystal structures, our method increases the generator's validity by 8 times compared to one of the baselines and by 143% compared to the previous CubicGAN along with its superiority in properties distribution and diversity. We further validated our generated candidates by DFT calculations, which successfully optimized/relaxed 1869 materials out of 2000 generated ones, of which 39.6% are with negative formation energy, indicating their stability.

In the third topic, we propose and evaluate machine-learning algorithms for determining the structure type of materials, given only their compositions. We couple random forest (RF) and multiple-layer perceptron (MLP) neural network models with three types of features: Magpie, atom vector, and one-hot encoding (atom frequency) for the crystal system and space group prediction of materials. Four types of models for predicting crystal systems and space groups are proposed, trained, and evaluated including one-versus-all binary classifiers, multiclass classifiers, polymorphism predictors, and multilabel classifiers. The synthetic minority over-sampling technique (SMOTE) is conducted to mitigate the effects of imbalanced data sets. Our results demonstrate that RF with Magpie features generally outperforms other algorithms for binary and multiclass prediction of crystal systems and space groups, while MLP with atom frequency features is the best one for structural polymorphism prediction.

Finally, we propose to use electronic charge density (ECD) as a generic unified 3D descriptor for materials property prediction with the advantage of possessing close relation with the physical and chemical properties of materials. We develop an ECD-based 3D convolutional neural network (CNN) for predicting the elastic properties of materials, in which CNNs can learn effective hierarchical features with multiple convolving and pooling operations. Our experiments show that our method can achieve good performance for elasticity prediction over 2170 Fm-3m materials.