Big Data Bridge

Monday, March 5, 2018 - 10:15 am
Innovation Center, Room 2277
COLLOQUIUM Justin Zhan Abstract Data has become the central driving force to new discoveries in science, informed governance, insight into society, and economic growth in the 21st century. Abundant data is a direct result of innovations including the Internet, faster computer processors, cheap storage, the proliferation of sensors, etc, and has the potential to increase business productivity and enable scientific discovery. However, while data is abundant and everywhere, people do not have a fundamental understanding of data. Traditional approaches to decision making under uncertainty are not adequate to deal with massive amounts of data, especially when such data is dynamically changing or becomes available over time. These challenges require novel techniques in data analytics, data-driven optimization, systems modeling and data mining. In this seminar, a number of recent funded data analytics projects will be presented to address various data analytics, mining, modeling, and optimization challenges. In particular, DataBridge, which is a novel data analytics system, will be illustrated. Dr. Justin Zhan is a professor at the Department of Computer Science, College of Engineering, Department of Radiology, School of Medicine, as well as Nevada Institute of Personalized Medicine. His research interests include Big Data, Information Assurance, Social Computing, Biomedical Computing and Health Informatics. He has been a steering chair of International Conference on Social Computing (SocialCom), and International Conference on Privacy, Security, Risk and Trust (PASSAT). He has been the editor-in-chief of International Journal of Privacy, Security and Integrity and International Journal of Social Computing and Cyber-Physical Systems. He has served as a conference general chair, a program chair, a publicity chair, a workshop chair, or a program committee member for over one-hundred and fifty international conferences and an editor-in-chief, an editor, an associate editor, a guest editor, an editorial advisory board member, or an editorial board member for about thirty journals. He has published more than two hundred articles in peer-reviewed journals and conferences and delivered thirty keynote speeches and invited talks. His research has been extensively funded by National Science Foundation, Department of Defense and National Institute of Health. Date: Mar. 5, 2018 Time: 10:15-11:15 am Place: Innovation Center, Room 2277

Bringing Millimeter-Wave Wireless to the Masses

Friday, March 2, 2018 - 10:15 am
Innovation Center, Room 2277
COLLOQUIUM Sanjib Sur University of Wisconsin- Madison Abstract: Many of the emerging IoT applications --- such as wireless virtual and augmented reality, autonomous vehicles, tactile internet --- demand multiple gigabits per second wireless throughput with sub-millisecond latency guarantees. Today’s wireless infrastructure --- such as LTE or Wi-Fi --- will unlikely handle such demand. Abundant opportunity, however, exists at millimeter-wave wireless, but with two key-barriers --- directional link alignment and link blockage --- that prevent the mass deployment of millimeter-wave in today’s network. In the first part of the talk, I will present my approach to addressing these two challenges by designing solutions that span across the wireless link, protocol, and system stack. Mass deployment of millimeter-wave devices also brings opportunity to enable new IoT applications, including designing new user-device interactions and ad-hoc imaging of objects hidden from the line-of-sight. In the second part of the talk, I will briefly go through my design to address the challenges of such ad-hoc applications. Finally, I will conclude this talk with a glimpse of my future works that are shaped by the emerging mass proliferation of cheap and ubiquitous wireless systems at millimeter-wave, sub-terahertz, and terahertz. Sanjib Sur is a Ph.D. candidate in the Electrical and Computer Engineering department at the University of Wisconsin-Madison. His research interests are in millimeter-wave networks, wireless and mobile systems, and IoT connectivity and sensing systems. His research works have appeared on multiple flagship conferences for wireless and mobile systems. Sanjib has been recently nominated for the Wisconsin Distinguished Graduate Fellowship for an outstanding graduate research work. He received a Bachelor’s degree with the highest distinction in Computer Science and Engineering from the Indian Institute of Engineering Science and Technology, where he was awarded the President of India Gold Medal for outstanding academic achievement. Location: Innovation Center, Room 2277 Date: Mar. 02 2018 Time: 10:15 - 11:15 AM

Towards Continual and Fine-Grained Learning for Robot Perception

Wednesday, February 28, 2018 - 10:15 am
Innovation Center, Room 2277
COLLOQUIUM Zsolt Kira Abstract A large number of robot perception tasks have been revolutionized by machine learning and deep neural networks in particular. However, current learning methods are limited in several ways that hinder their large-scale use for critical robotics applications: They are often focused on individual sensor modalities, do not attempt to understand semantic information in a fine-grained temporal manner, and are beholden to strong assumptions about the data (e.g. that the data distribution is the same when deployed in the real world as when trained). In this talk, I will describe work on novel deep learning architectures for moving beyond current methods to develop a richer multi-modal and fine-grained scene understanding from raw sensor data. I will also discuss methods we have developed that can use transfer learning to deal with changes in the environment or the existence of entirely new, unknown categories in the data (e.g. unknown object types). I will focus especially on this latter work, where we use neural networks to learn how to compare objects and transfer such learning to new domains using one of the first deep-learning based clustering algorithms, which we developed. I will show examples of real-world robotic systems using these methods, and conclude by discussing future directions in this area, towards making robots able to continually learn and adapt to new situations as they arise. Dr. Zsolt Kira received his B.S. in ECE at the University of Miami in 2002 and M.S. and Ph.D. in Computer Science from the Georgia Institute of Technology in 2010. He is currently a Senior Research Scientist and Branch Chief of the Machine Learning and Analytics group at the Georgia Tech Research Institute (GTRI). He is also an Adjunct at the School of Interactive Computing and Associate Director of Georgia Tech’s Machine Learning Center (ML@GT). He conducts research in the areas of machine learning for sensor processing and robot perception, with emphasis on feature learning for multi-modal object detection, video analysis, scene characterization, and transfer learning. He has over 25 publications in these areas, several best paper/student paper and other awards, and has been invited to speak at related workshops in both academia and government venues. Date: Feb. 28, 2018 Time: 10:15-11:15 am Place: Innovation Center, Room 2277

Improving Speech-related Facial Action Unit Recognition by Audiovisual Information Fusion

Tuesday, February 27, 2018 - 08:00 am
Meeting room 2267, Innovation Center
DISSERTATION DEFENSE Zibo Meng Advisor : Dr. Yan Tong Abstract In spite of great progress achieved on posed facial display and controlled image acquisition, performance of facial action unit (AU) recognition degrades significantly for spontaneous facial displays. Furthermore, recognizing AUs accompanied with speech is even more challenging since they are generally activated at a low intensity with subtle facial appearance/geometrical changes during speech, and more importantly, often introduce ambiguity in detecting other co-occurring AUs, e.g., producing non-additive appearance changes. All the current AU recognition systems utilized information extracted only from visual channel. However, sound is highly correlated with visual channel in human communications. Thus, we propose to exploit both audio and visual information for AU recognition. Specifically, a feature-level fusion method combining both audio and visual features is first introduced. Specifically, features are independently extracted from visual and audio channels. The extracted features are aligned to handle the difference in time scales and the time shift between the two signals. These temporally aligned features are integrated via feature-level fusion for AU recognition. Second, a novel approach that recognizes speech-related AUs exclusively from audio signals based on the fact that facial activities are highly correlated with voice during speech is developed. Specifically, dynamic and physiological relationships between AUs and phonemes are modeled through a continuous time Bayesian network (CTBN); then AU recognition is performed by probabilistic inference via the CTBN model. Third, a novel audiovisual fusion framework, which aims to make the best use of visual and acoustic cues in recognizing speech-related facial AUs is developed. In particular, a dynamic Bayesian network (DBN) is employed to explicitly model the semantic and dynamic physiological relationships between AUs and phonemes as well as measurement uncertainty. AU recognition is then conducted by probabilistic inference via the DBN model. To evaluate the proposed approaches, a pilot AU-coded audiovisual database was collected. Experiments on this dataset have demonstrated that the proposed frameworks yield significant improvement in recognizing speech-related AUs compared to the state-of-the-art visual-based methods. Furthermore, more impressive improvement has been achieved for those AUs, whose visual observations are impaired during speech.

Inference Framework for Model Update and Development

Monday, February 26, 2018 - 01:30 pm
Meeting room 2265, Innovation Center
DISSERTATION DEFENSE Xiao Lin Advisor : Dr. Gabriel Terejanu Abstract Computational models play an important role in scientific discovery and engineering design. However, developing computational models is challenging, since the process always follows a path contaminated with errors and uncertainties. The uncertainties and errors inherent in computational models are the result of many factors, including experimental uncertainties, model structure inadequacies, uncertainties in model parameters and initial conditions, as well as errors due to numerical discretiza- tions. To realize the full potential in applications it is critical to systematically and economically reduce the uncertainties inherent in all computational models. The update and development of computational models is a recursive process between data assimilation and data selection. In data assimilation, measurements are incorporated into computational simulations to reduce the uncertainties of the model and in reverse, the simulations help determine where to acquire data such that most information can be provided. Currently, data assimilation techniques are overwhelmed by data volume and velocity and increased complexity of computational models. In this work, we develop a novel data assimilation approach EnLLVM which is based on linear latent variable model. There are several advantages of this approach. First, it works well with high dimensional dynamic systems, but only requires a small number of samples. Second, it can absorb model structure error and reflect the error in the uncertainty of data assimilation results. In addition, data assimilation is performed without calculating likelihood of observation, thus it can be applied to data assimilation problems in which likelihood is intractable. Obtaining informative data is also crucial, as data collection is an expensive endeavor for a number of science and engineering fields. Mutual information, which naturally measures information provided about one quantity by knowing the other quantity, has become a major design metric and has fueled a large body of work on experimental design. However, estimating mutual information is challenging and results are not reliable in high dimensions. In this work, we derive a lower bound of mutual information, which is computed in much lower dimensions. This lower bound can be applied to experimental design as well as other problems that require comparison of mutual information. At last, we develop a general framework for building computational models. In this framework, hypotheses about unknown model structure are generated by using EnLLVM for data assimilation and lower bound of mutual information for finding relations between state variables and unknown structure function. Then, different hypotheses can be ranked with model selection technique. This framework not only provides a way to infer model discrepancy, but also could further contribute to scientific discoveries.

Transfer Learning for Performance Analysis of Highly-Configurable Software Systems

Monday, February 26, 2018 - 10:15 am
Innovation Center, Room 2277
COLLOQUIUM Pooyan Jamshidi Abstract A wide range of modern software-intensive systems (e.g., autonomous systems, big data analytics, robotics, deep neural architectures) are built configurable. These systems offer a rich space for adaptation to different domains and tasks. Developers and users often need to reason about the performance of such systems, making tradeoffs to change specific quality attributes or detecting performance anomalies. For instance, developers of image recognition mobile apps are not only interested in learning which deep neural architectures are accurate enough to classify their images correctly, but also which architectures consume the least power on the mobile devices on which they are deployed. Recent research has focused on models built from performance measurements obtained by instrumenting the system. However, the fundamental problem is that the learning techniques for building a reliable performance model do not scale well, simply because the configuration space is exponentially large that is impossible to exhaustively explore. For example, it will take over 60 years to explore the whole configuration space of a system with 25 binary options. In this talk, I will start motivating the configuration space explosion problem based on my previous experience with large-scale big data systems in industry. I will then present my transfer learning solution to tackle the scalability challenge: instead of taking the measurements from the real system, we learn the performance model using samples from cheap sources, such as simulators that approximate the performance of the real system, with a fair fidelity and at a low cost. Results show that despite the high cost of measurement on the real system, learning performance models can become surprisingly cheap as long as certain properties are reused across environments. In the second half of the talk, I will present empirical evidence, which lays a foundation for a theory explaining why and when transfer learning works by showing the similarities of performance behavior across environments. I will present observations of environmental changes‘ impacts (such as changes to hardware, workload, and software versions) for a selected set of configurable systems from different domains to identify the key elements that can be exploited for transfer learning. These observations demonstrate a promising path for building efficient, reliable, and dependable software systems. Finally, I will share my research vision for the next five years and outline my immediate plans to further explore the opportunities of transfer learning. Pooyan Jamshidi is a postdoctoral researcher at Carnegie Mellon University, where he works on transfer learning for building performance models to enable dynamic adaptation of mobile robotics software as a part of BRASS, a DARPA sponsored project. Prior to his current position, he was a research associate at Imperial College London, where he worked on Bayesian optimization for automated performance tuning of big data systems. He holds a Ph.D. from Dublin City University, where he worked on self-learning Fuzzy control for auto-scaling in the cloud. He has spent 7 years in industry as a developer and a software architect. His research interests are at the intersection of software engineering, systems, and machine learning, and his focus lies predominantly in the areas of highly-configurable and self-adaptive systems (more details: https://pooyanjamshidi.github.io/research/). Date: Feb. 26, 2018 Time: 10:15-11:15 am Place: Innovation Center, Room 2277

Internet of Acoustic Things (IoAT): Challenges, Opportunities, and Threats

Monday, February 19, 2018 - 10:15 am
Storey Innovation Center, Room 2277
Abstract: The recent proliferation of acoustic devices, ranging from voice assistants to wearable health monitors, is leading to a sensing ecosystem around us -- referred to as the Internet of Acoustic Things or IoAT. My research focuses on developing hardware-software building blocks that enable new capabilities for this emerging future. In this talk, I will sample some of my projects. For instance, (1) I will demonstrate carefully designed sounds that are completely inaudible to humans but recordable by all microphones. (2) I will discuss our work with physical vibrations from mobile devices, and how they conduct through finger bones to enable new modalities of short range, human-centric communication. (3) Finally, I will draw attention to various acoustic leakages and threats that arrive with sensor-rich environments. I will conclude this talk with a glimpse of my ongoing and future projects targeting a stronger convergence of sensing, computing, and communications in tomorrow’s IoT, cyber-physical systems, and healthcare technologies. Bio: Nirupam Roy is a Ph.D. candidate in Electrical and Computer Engineering at the University of Illinois, Urbana-Champaign (UIUC). His research interests are in mobile sensing, wireless networking, and embedded systems with applications to IoT, cyber-physical-systems, and security. Roy is the recipient of the Valkenburg graduate research award, the Lalit Bahl fellowship, and the outstanding thesis awards from both his Bachelor's and Master's institutes. His recent research on "Making Microphones Hear Inaudible Sounds" received the MobiSys'17 best paper award and was selected for the ACM SIGMOBILE research highlights of the year in 2017.