I2S Masters/ Doctoral Theses


All students and faculty are welcome to attend the final defense of I2S graduate students completing their M.S. or Ph.D. degrees. Defense notices for M.S./Ph.D. presentations for this year and several previous years are listed below in reverse chronological order.

Students who are nearing the completion of their M.S./Ph.D. research should schedule their final defenses through the EECS graduate office at least THREE WEEKS PRIOR to their presentation date so that there is time to complete the degree requirements check, and post the presentation announcement online.

Upcoming Defense Notices

Arin Dutta

Performance Analysis of Distributed Raman Amplification with Dual-Order Forward Pumping

When & Where:


Nichols Hall, Room 250 (Gemini Room)

Degree Type:

PhD Comprehensive Defense

Committee Members:

Rongqing Hui, Chair
Christopher Allen
Morteza Hashemi
Alessandro Saladrino
Hui Zhao

Abstract

As internet services like high-definition videos, cloud computing, and artificial intelligence keep growing, optical networks need to keep up with the demand for more capacity. Optical amplifiers play a crucial role in offsetting fiber loss and enabling long-distance wavelength division multiplexing (WDM) transmission in high-capacity systems. Various methods have been proposed to enhance the capacity and reach of fiber communication systems, including advanced modulation formats, dense wavelength division multiplexing (DWDM) over ultra-wide bands, space-division multiplexing, and high-performance digital signal processing (DSP) technologies. To sustain higher data rates while maximizing the spectral efficiency of multi-level modulated signals, a higher Optical signal-to-noise ratio (OSNR) is necessary. Despite advancements in coherent optical communication systems, the spectral efficiency of multi-level modulated signals is ultimately constrained by fiber nonlinearity.

Raman amplification is an attractive solution for wide-band amplification with low noise figures in multi-band systems. Distributed Raman Amplification (DRA) has been deployed in recent high-capacity transmission experiments to achieve a relatively flat signal power distribution along the optical path and offers the unique advantage of using conventional low-loss silica fibers as the gain medium, effectively transforming passive optical fibers into active or amplifying waveguides. Additionally, DRA provides gain at any wavelength by selecting the appropriate pump wavelength, enabling operation in signal bands outside the Erbium-doped fiber amplifier (EDFA) bands. Forward (FW) Raman pumping in DRA can be adopted to further improve the DRA performance as it is more efficient in OSNR improvement because the optical noise is generated near the beginning of the fiber span and attenuated along the fiber. Dual-order FW pumping helps to reduce the non-linear effect of the optical signal and improves OSNR by more uniformly distributing the Raman gain along the transmission span.

The major concern with Forward Distributed Raman Amplification (FW DRA) is the fluctuation in pump power, known as relative intensity noise (RIN), which transfers from the pump laser to both the intensity and phase of the transmitted optical signal as they propagate in the same direction. Additionally, another concern of FW DRA is the rise in signal optical power near the start of the fiber span, leading to an increase in the Kerr-effect-induced non-linear phase shift of the signal. These factors, including RIN transfer-induced noise and non-linear noise, contribute to the degradation of the system performance in FW DRA systems at the receiver. As the performance of DRA with backward pumping is well understood with a relatively low impact of RIN transfer, our study is focused on the FW pumping scheme.

Our research is intended to provide a comprehensive analysis of the system performance impact of dual-order FW Raman pumping, including signal intensity and phase noise induced by the RINs of both the 1st and the 2nd order pump lasers, as well as the impacts of linear and nonlinear noise. The efficiencies of pump RIN to signal intensity and phase noise transfer are theoretically analyzed and experimentally verified by applying a shallow intensity modulation to the pump laser to mimic the RIN. The results indicate that the efficiency of the 2nd order pump RIN to signal phase noise transfer can be more than 2 orders of magnitude higher than that from the 1st order pump. Then the performance of the dual-order FW Raman configurations is compared with that of single-order Raman pumping to understand the trade-offs of system parameters. The nonlinear interference (NLI) noise is analyzed to study the overall OSNR improvement when employing a 2nd order Raman pump.

Finally, a DWDM system with 16-QAM modulation is used as an example to investigate the benefit of DRA with dual order Raman pumping and with different pump RIN levels. We also consider a DRA system using a 1st order incoherent pump together with a 2nd order coherent pump. Although dual-order FW pumping corresponds to a slight increase of linear amplified spontaneous emission (ASE) compared to using only a 1st order pump, its major advantage comes from the reduction of nonlinear interference noise in a DWDM system. Because the RIN of the 2nd order pump has much higher impact than that of the 1st order pump, there should be more stringent requirement on the RIN of the 2nd order pump laser when dual order FW pumping scheme is used for DRA for efficient fiber-optic communication. Also, the result of system performance analysis reveals that higher baud rate systems, like those operating at 100Gbaud, are less affected by pump laser RIN due to the low-pass characteristics of the transfer of pump RIN to signal phase noise.


Babak Badnava

Joint Communication and Computation for Emerging Applications in Next Generation of Wireless Networks

When & Where:


Nichols Hall, Room 246 (Executive Conference Room)

Degree Type:

PhD Comprehensive Defense

Committee Members:

Morteza Hashemi, Chair
Taejoon Kim
Prasad Kulkarni
Shawn Keshmiri

Abstract

Emerging applications in next-generation wireless networks are driving the need for innovative communication and computation systems. Notable examples include augmented and virtual reality (AR/VR), autonomous vehicles, and mobile edge computing, all of which demand significant computational and communication resources at the network edge. These demands place a strain on edge devices, which are often resource-constrained. In order to incorporate available communication and computation resources, while enhancing user experience, this PhD research is dedicated to developing joint communication and computation solutions for next generation wireless applications that could potentially operate in high frequencies such as millimeter wave (mmWave) bands.

In the first thrust of this study, we examine the problem of energy-constrained computation offloading to edge servers in a multi-user multi-channel wireless network. To develop a decentralized offloading policy for each user, we model the problem as a partially observable Markov decision problem (POMDP). Leveraging bandit learning methods, we introduce a decentralized task offloading solution, where edge users offload their computation tasks to a nearby edge server using a selected communication channel. The proposed framework aims to meet user's requirements, such as task completion deadline and computation throughput (i.e., the rate at which computational results are produced).

The second thrust of the study emphasizes user-driven requirements for these resource-intensive applications, specifically the Quality of Experience (QoE) in 2D and 3D video streaming. Given the unique characteristics of mmWave networks, we develop a beam alignment and buffer predictive multi-user scheduling algorithm for 2D video streaming applications. This scheduling algorithm balances the trade-off between beam alignment overhead and playback buffer levels for optimal resource allocation across users. Next, we extend our investigation and develop a joint rate adaptation and computation distribution algorithm for 3D video streaming in mmWave-based VR systems. Our proposed framework balances the trade-off between communication and computation resource allocation to enhance the users’ QoE. Our numerical results using real-world mmWave traces and 3D video dataset, show promising improvements in terms of video quality, rebuffering time, and quality variation perceived by users.


Arman Ghasemi

Task-Oriented Communication and Distributed Control in Smart Grids with Time-Series Forecasting

When & Where:


Nichols Hall, Room 246

Degree Type:

PhD Comprehensive Defense

Committee Members:

Morteza Hashemi, Chair
Alex Bardas
Taejoon Kim
Prasad Kulkarni
Zsolt Talata

Abstract

Smart grids face challenges in maintaining the balance between generation and consumption at the residential and grid scales with the integration of renewable energy resources. Decentralized, dynamic, and distributed control algorithms are necessary for smart grids to function effectively. The inherent variability and uncertainty of renewables, especially wind and solar energy, complicate the deployment of distributed control algorithms in smart grids. In addition, smart grid systems must handle real-time data collected from interconnected devices and sensors while maintaining reliable and secure communication regardless of network failures. To address these challenges, our research models the integration of renewable energy resources into the smart grid and evaluates how predictive analytics can improve distributed control and energy management, while recognizing the limitations of communication channels and networks.

In the first thrust of this research, we develop a model of a smart grid with renewable energy integration and evaluate how forecasting affects distributed control and energy management. In particular, we investigate how contextual weather information and renewable energy time-series forecasting affect smart grid energy management. In addition to modeling the smart grid system and integrating renewable energy resources, we further explore the use of deep learning methods, such as the Long Short-Term Memory (LSTM) and Transformer models, for time-series forecasting. Time-series forecasting techniques are applied within Reinforcement Learning (RL) frameworks to enhance decision-making processes.

In the second thrust, we note that data collection and sharing across the smart grids require considering the impact of network and communication channel limitations in our forecasting models. As renewable energy sources and advanced sensors are integrated into smart grids, communication channels on wireless networks are overflowed with data, requiring a shift from transmitting raw data to processing only useful information to maximize efficiency and reliability. To this end, we develop a task-oriented communication model that integrates data compression and the effects of data packet queuing with considering limitation of communication channels, within a remote time-series forecasting framework. Furthermore, we jointly integrate data compression technique with age of information metric to enhance both relevance and timeliness of data used in time-series forecasting.


Neel Patel

Near-Memory Acceleration of Compressed Far Memory

When & Where:


Nichols Hall, Room 250 (Gemini Room)

Degree Type:

MS Thesis Defense

Committee Members:

Mohammad Allan, Chair
David Johnson
Prasad Kulkarni


Abstract

DRAM constitutes over 50% of server cost and 75% of the embodied carbon footprint of a server. To mitigate DRAM cost, far memory architectures have emerged. They can be separated into two broad categories: software-defined far memory (SFM) and disaggregated far memory (DFM). In this work, we compare the cost of SFM and DFM in terms of their required capital investment, operational expense, and carbon footprint. We show that, for applications whose data sets are compressible and have predictable memory access patterns, it takes several years for a DFM to break even with an equivalent capacity SFM in terms of cost and sustainability. We then introduce XFM, a near-memory accelerated SFM architecture, which exploits the coldness of data during SFM-initiated swap ins and outs. XFM leverages refresh cycles to seamlessly switch the access control of DRAM between the CPU and near-memory accelerator. XFM parallelizes near-memory accelerator accesses with row refreshes and removes the memory interference caused by SFM swap ins and outs.

We modify an open source far memory implementation to implement a full-stack, user-level XFM. Our experimental results use a combination of an FPGA implementation, simulation, and analytical modeling to show that XFM eliminates memory bandwidth utilization when performing compression and decompression operations with SFMs of capacities up to 1TB. The memory and cache utilization reductions translate to 5∼27% improvement in the combined performance of co-running applications.


Durga Venkata Suraj Tedla

Block chain based inter organization file sharing system

When & Where:


Eaton Hall, Room 2001B

Degree Type:

MS Project Defense

Committee Members:

David Johnson, Chair
Drew Davidson
Sankha Guria


Abstract

A coalition of companies collaborates collectively and shares information to improve their operations together. Distributed trust and transparency cannot be obtained with centralized file-sharing platforms. File sharing may be done transparently and securely with blockchain technology. This project suggests an inter-organizational secure file-sharing system based on blockchain technology. The group can use it to securely share files in a distributed manner. The creation of smart contracts and the configuration of blockchain networks are carried out by Hyperledger Fabric, an enterprise blockchain platform. Distributed file storage is accomplished through the usage of the Inter Planetary File System (IPFS).

The workflow for file-sharing and identity management procedures is provided in the paper. Using blockchain technology, the recommended approach enables a group of businesses to share files with availability, integrity, and confidentiality. The suggested method uses blockchain to enable safe file exchange amongst a group of enterprises. It offers shared file availability, confidentiality, and integrity. It guarantees complete file encryption. The blockchain provides tamper-resistant storage for the shared file's content ID. On the distributed storage and blockchain ledger, respectively, the encrypted file and file metadata are stored.


Dang Qua Nguyen

Hybrid Precoding Optimization and Private Federated Learning for Future Wireless Systems

When & Where:


Nichols Hall, Room 246 (Executive Conference Room)

Degree Type:

PhD Comprehensive Defense

Committee Members:

Taejoon Kim, Chair
Morteza Hashemi
Erik Perrins
Zijun Yao
KC Kong

Abstract

This PhD research addresses two challenges in future wireless systems: hybrid precoder design for sub-Terahertz (sub-THz) massive multiple-input multiple-output (MIMO) communications and private federated learning (FL) over wireless channels. The first part of the research introduces a novel hybrid precoding framework that combines true-time delay (TTD) and phase shifters (PS) precoders to counteract the beam squint effect - a significant challenge in sub-THz massive MIMO systems that leads to considerable loss in array gain.

Our research presents a novel joint optimization framework for the TTD and PS precoder design, incorporating realistic time delay constraints for each TTD device. We first derive a lower bound on the achievable rate of the system and show that, in the asymptotic regime, the optimal analog precoder that fully compensates for the beam squint is equivalent to the one that maximizes this lower bound. Unlike previous methods, our framework does not rely on the unbounded time delay assumption and optimizes the TTD and PS values jointly to cope with the practical limitations. Furthermore, we determine the minimum number of TTD devices needed to reach a target array gain using our proposed approach.

Simulations validate that the proposed approach demonstrates performance enhancement, ensures array gain, and achieves computational efficiency. In the second part, the research devises a differentially private FL algorithm that employs time-varying noise perturbation and optimizes transmit power to counteract privacy risks, particularly those stemming from engineering-inversion attacks. This method harnesses inherent wireless channel noise to strike a balance between privacy protection and learning utility. By strategically designing noise perturbation and power control, our approach not only safeguards user privacy but also upholds the quality of the learned FL model. Additionally, the number of FL iterations is optimized by minimizing the upper bound on the learning error. We conduct simulations to showcase the effectiveness of our approach in terms of DP guarantee and learning utility.


Past Defense Notices

Dates

Mohsen Nayebi Kerdabadi

Contrastive Learning of Temporal Distinctiveness for Survival Analysis in Electronic Health Records

When & Where:


Nichols Hall, Room 250 (Gemini Room)

Degree Type:

MS Project Defense

Committee Members:

Zijun Yao, Chair
Fengjun Li
Cuncong Zhong


Abstract

Survival analysis plays a crucial role in many healthcare decisions, where the risk prediction for the events of interest can support an informative outlook for a patient's medical journey. Given the existence of data censoring, an effective way of survival analysis is to enforce the pairwise temporal concordance between censored and observed data, aiming to utilize the time interval before censoring as partially observed time-to-event labels for supervised learning. Although existing studies mostly employed ranking methods to pursue an ordering objective, contrastive methods which learn a discriminative embedding by having data contrast against each other, have not been explored thoroughly for survival analysis. Therefore, we propose a novel Ontology-aware Temporality-based Contrastive Survival (OTCSurv) analysis framework that utilizes survival durations from both censored and observed data to define temporal distinctiveness and construct negative sample pairs with adjustable hardness for contrastive learning. Specifically, we first use an ontological encoder and a sequential self-attention encoder to represent the longitudinal EHR data with rich contexts. Second, we design a temporal contrastive loss to capture varying survival durations in a supervised setting through a hardness-aware negative sampling mechanism. Last, we incorporate the contrastive task into the time-to-event predictive task with multiple loss components. We conduct extensive experiments using a large EHR dataset to forecast the risk of hospitalized patients who are in danger of developing acute kidney injury (AKI), a critical and urgent medical condition. The effectiveness and explainability of the proposed model are validated through comprehensive quantitative and qualitative studies.


Jarrett Zeliff

An Analysis of Bluetooth Mesh Security Features in the Context of Secure Communications

When & Where:


Eaton Hall, Room 1

Degree Type:

MS Thesis Defense

Committee Members:

Alexadnru Bardas, Chair
Drew Davidson
Fengjun Li


Abstract

Significant developments in communication methods to help support at-risk populations have increased over the last 10 years. We view at-risk populations as a group of people present in environments where the use of infrastructure or electricity, including telecommunications, is censored and/or dangerous. Security features that accompany these communication mechanisms are essential to protect the confidentiality of its user base and the integrity and availability of the communication network.

In this work, we look at the feasibility of using Bluetooth Mesh as a communication network and analyze the security features that are inherent to the protocol. Through this analysis we determine the strengths and weaknesses of Bluetooth Mesh security features when used as a messaging medium for at risk populations and provide improvements to current shortcomings. Our analysis includes looking at the Bluetooth Mesh Networking Security Fundamentals as described by the Bluetooth Sig: Encryption and Authentication, Separation of Concerns, Area isolation, Key Refresh, Message Obfuscation, Replay Attack Protection, Trashcan Attack Protection, and Secure Device Provisioning.  We look at how each security feature is implemented and determine if these implementations are sufficient in protecting the users from various attack vectors. For example, we examined the Blue Mirror attack, a reflection attack during the provisioning process which leads to the compromise of network keys, while also assessing the under-researched key refresh mechanism. We propose a mechanism to address Blue-Mirror-oriented attacks with the goal of creating a more secure provisioning process.  To analyze the key refresh mechanism, we implemented our own full-fledged Bluetooth Mesh network and implemented a key refresh mechanism. Through this we form an assessment of the throughput, range, and impacts of a key refresh in both lab and field environments that demonstrate the suitability of our solution as a secure communication method.


Daniel Johnson

Probability-Aware Selective Protection for Sparse Iterative Solvers

When & Where:


Nichols Hall, Room 246

Degree Type:

MS Thesis Defense

Committee Members:

Hongyang Sun, Chair
Perry Alexander
Zijun Yao


Abstract

With the increasing scale of high-performance computing (HPC) systems, transient bit-flip errors are now more likely than ever, posing a threat to long-running scientific applications. A substantial portion of these applications involve the simulation of partial differential equations (PDEs) modeling physical processes over discretized spatial and temporal domains, with some requiring the solving of sparse linear systems. While these applications are often paired with system-level application-agnostic resilience techniques such as checkpointing and replication, the utilization of these techniques imposes significant overhead. In this work, we present a probability-aware framework that produces low-overhead selective protection schemes for the widely used Preconditioned Conjugate Gradient (PCG) method, whose performance can heavily degrade due to error propagation through the sparse matrix-vector multiplication (SpMV) operation. Through the use of a straightforward mathematical model and an optimized machine learning model, our selective protection schemes incorporate error probability to protect only certain crucial operations. An experimental evaluation using 15 matrices from the SuiteSparse Matrix Collection demonstrates that our protection schemes effectively reduce resilience overheads, often outperforming or matching both baseline and established protection schemes across all error probabilities.


Javaria Ahmad

Discovering Privacy Compliance Issues in IoT Apps and Alexa Skills Using AI and Presenting a Mechanism for Enforcing Privacy Compliance

When & Where:


LEEP2, Room 2425

Degree Type:

PhD Dissertation Defense

Committee Members:

Bo Luo, Chair
Alex Bardas
Tamzidul Hoque
Fengjun Li
Michael Zhuo Wang

Abstract

The growth of IoT and voice assistant (VA) apps poses increasing concerns about sensitive data leaks. While privacy policies are required to describe how these apps use private user data (i.e., data practice), problems such as missing, inaccurate, and inconsistent policies have been repeatedly reported. Therefore, it is important to assess the actual data practice in apps and identify the potential gaps between the actual and declared data usage. We find that app stores lack in regulating the compliance between the app practices and their declaration, so we use AI to discover the compliance issues in these apps to assist the regulators and developers. For VA apps, we also develop a mechanism to enforce the compliance using AI. In this work, we conduct a measurement study using our framework called IoTPrivComp, which applies an automated analysis of IoT apps’ code and privacy policies to identify compliance gaps. We collect 1,489 IoT apps with English privacy policies from the Play Store. IoTPrivComp detects 532 apps with sensitive external data flows, among which 408 (76.7%) apps have undisclosed data leaks. Moreover, 63.4% of the data flows that involve health and wellness data are inconsistent with the practices disclosed in the apps’ privacy policies. Next, we focus on the compliance issues in skills. VAs, such as Amazon Alexa, are integrated with numerous devices in homes and cars to process user requests using apps called skills. With their growing popularity, VAs also pose serious privacy concerns. Sensitive user data captured by VAs may be transmitted to third-party skills without users’ consent or knowledge about how their data is processed. Privacy policies are a standard medium to inform the users of the data practices performed by the skills. However, privacy policy compliance verification of such skills is challenging, since the source code is controlled by the skill developers, who can make arbitrary changes to the behaviors of the skill without being audited; hence, conventional defense mechanisms using static/dynamic code analysis can be easily escaped. We present Eunomia, the first real-time privacy compliance firewall for Alexa Skills. As the skills interact with the users, Eunomia monitors their actions by hijacking and examining the communications from the skills to the users, and validates them against the published privacy policies that are parsed using a BERT-based policy analysis module. When non-compliant skill behaviors are detected, Eunomia stops the interaction and warns the user. We evaluate Eunomia with 55,898 skills on Amazon skills store to demonstrate its effectiveness and to provide a privacy compliance landscape of Alexa skills.


Xiangyu Chen

Toward Efficient Deep Learning for Computer Vision Applications

When & Where:


Nichols Hall, Room 246

Degree Type:

PhD Dissertation Defense

Committee Members:

Cuncong Zhong, Chair
Prasad Kulkarni
Bo Luo
Fengjun Li
Honguo Xu

Abstract

Deep learning leads the performance in many areas of computer vision. However, after a decade of research, it tends to require larger datasets and more complex models, leading to heightened resource consumption across all fronts. Regrettably, meeting these requirements proves challenging in many real-life scenarios. First, both data collection and labeling processes entail substantial labor and time investments. This challenge becomes especially pronounced in domains such as medicine, where identifying rare diseases demands meticulous data curation. Secondly, the large size of state-of-the-art models, such as ViT, Stable Diffusion, and ConvNext, hinders their deployment on resource-constrained platforms like mobile devices. Research indicates pervasive redundancies within current neural network structures, exacerbating the issue. Lastly, even with ample datasets and optimized models, the time required for training and inference remains prohibitive in certain contexts. Consequently, there is a burgeoning interest among researchers in exploring avenues for efficient artificial intelligence.

This study endeavors to delve into various facets of efficiency within computer vision, including data efficiency, model efficiency, as well as training and inference efficiency. The data efficiency is improved from the perspective of increasing information brought by given image inputs and reducing redundancies of RGB image formats. To achieve this, we propose to integrate both spatial and frequency representations to finetune the classifier. Additionally, we propose explicitly increasing the input information density in the frequency domain by deleting unimportant frequency channels. For model efficiency, we scrutinize the redundancies present in widely used vision transformers. Our investigation reveals that trivial attention in their attention modules covers useful non-trivial attention due to its large amount. We propose mitigating the impact of accumulated trivial attention weights. To increase training efficiency, we propose SuperLoRA, a generation of LoRA adapter, to fine-tune pretrained models with few iterations and extremely-low parameters. Finally, a model simplification pipeline is proposed to further reduce inference time on mobile devices. By addressing these challenges, we aim to advance the practicality and performance of computer vision systems in real-world applications.


Grace Young

Quantum Polynomial-Time Reduction for the Dihedral Hidden Subgroup Problem

When & Where:


Nichols Hall, Room 246

Degree Type:

PhD Dissertation Defense

Committee Members:

Perry Alexander, Chair
Esam El-Araby
Matthew Moore
Cuncong Zhong
KC Kong

Abstract

The last century has seen incredible growth in the field of quantum computing. Quantum computation offers the opportunity to find efficient solutions to certain computational problems which are intractable on classical computers. One class of problems that seems to benefit from quantum computing is the Hidden Subgroup Problem (HSP). The HSP includes, as special cases, the problems of integer factoring, discrete logarithm, shortest vector, and subset sum - making the HSP incredibly important in various fields of research.                               

The presented research examines the HSP for Dihedral groups with order 2^n and proves a quantum polynomial-time reduction to the so-called Codomain Fiber Intersection Problem (CFIP). The usual approach to the HSP relies on harmonic analysis in the domain of the problem and the best-known algorithm using this approach is sub-exponential, but still super-polynomial. The algorithm we will present deviates from the usual approach by focusing on the structure encoded in the codomain and uses this structure to direct a “walk” down the subgroup lattice terminating at the hidden subgroup.                               

Though the algorithm presented here is specifically designed for the DHSP, it has potential applications to many other types of the HSP. It is hypothesized that any group with a sufficiently structured subgroup lattice could benefit from the analysis developed here. As this approach diverges from the standard approach to the HSP it could be a promising step in finding an efficient solution to this problem.


Daniel Herr

Information Theoretic Physical Waveform Design with Application to Waveform-Diverse Adaptive-on-Transmit Radar

When & Where:


Nichols Hall, Room 246

Degree Type:

PhD Comprehensive Defense

Committee Members:

James Stiles, Chair
Chris Allen
Shannon Blunt
Carl Leuschen
Chris Depcik

Abstract

Information theory provides methods for quantifying the information content of observed signals and has found application in the radar sensing space for many years. Here, we examine a type of information derived from Fisher information known as Marginal Fisher Information (MFI) and investigate its use to design pulse-agile waveforms. By maximizing this form of information, the expected error covariance about an estimation parameter space may be minimized. First, a novel method for designing MFI optimal waveforms given an arbitrary waveform model is proposed and analyzed. Next, a transformed domain approach is proposed in which the estimation problem is redefined such that information is maximized about a linear transform of the original estimation parameters. Finally, informationally optimal waveform design is paired with informationally optimal estimation (receive processing) and are combined into a cognitive radar concept. Initial experimental results are shown and a proposal for continued research is presented.


Rachel Chang

Designing Pseudo-Random Staggered PRI Sequences

When & Where:


Nichols Hall, Room 246

Degree Type:

MS Thesis Defense

Committee Members:

Shannon Blunt, Chair
Chris Allen
James Stiles


Abstract

In uniform pulse-Doppler radar, there is a well known trade-off between unambiguous Doppler and unambiguous range. Pulse repetition interval (PRI) staggering, a technique that involves modulating the interpulse times, addresses this trade-space allowing for expansion of the unambiguous Doppler domain with little range swath incursion. Random PRI staggering provides additional diversity, but comes at the cost of increased Doppler sidelobes. Thus, careful PRI sequence design is required to avoid spurious sidelobe peaks that could result in false alarms.

In this thesis, two random PRI stagger models are defined and compared, and sidelobe peak mitigation is discussed. First, the co-array concept (borrowed from the intuitively related field of sparse array design in the spatial domain) is utilized to examine the effect of redundancy on sidelobe peaks for random PRI sequences. Then, a sidelobe peak suppression technique is introduced that involves a gradient-based optimization of the random PRI sequences, producing pseudo-random sequences that are shown to significantly reduce spurious Doppler sidelobes in both simulation and experimentally.


Fatima Al-Shaikhli

Fiber Property Characterization based on Electrostriction

When & Where:


Nichols Hall 250 | Gemini Room

Degree Type:

MS Thesis Defense

Committee Members:

Ron Hui, Chair
Shannon Blunt
Shima Fardad


Abstract

Electrostriction in an optical fiber is introduced by the interaction between the forward propagated optical signal and the acoustic standing waves in the radial direction resonating between the center of the core and the cladding circumference of the fiber. The response of electrostriction is dependent on fiber parameters, especially the mode field radius. A novel technique is demonstrated to characterize fiber properties by means of measuring their electrostriction response under intensity modulation. As the spectral envelope of electrostriction-induced propagation loss is anti-symmetrical, the signal-to-noise ratio can be significantly increased by subtracting the measured spectrum from its complex conjugate. It is shown that if the transversal field distribution of the fiber propagation mode is Gaussian, the envelope of the electrostriction-induced loss spectrum closely follows a Maxwellian distribution whose shape can be specified by a single parameter determined by the mode field radius. 


Venkata Nadha Reddy Karasani

Implementing Web Presence For The History Of Black Writing

When & Where:


LEEP2, Room 1415

Degree Type:

MS Thesis Defense

Committee Members:

Drew Davidson, Chair
Perry Alexander
Hossein Saiedian


Abstract

The Black Literature Network Project is a comprehensive initiative to disseminate literature knowledge to students, academics, and the general public. It encompasses four distinct portals, each featuring content created and curated by scholars in the field. These portals include the Novel Generator Machine, Literary Data Gallery, Multithreaded Literary Briefs, and Remarkable Receptions Podcast Series. My significant contribution to this project was creating a standalone website for the Current Archives and Collections Index that offers an easily searchable index of black-themed collections. Additionally, I was exclusively responsible for the complete development of the novel generator tool. This application provides customized book recommendations based on user preferences. As a part of the History of Black Writing (HBW) Program, I had the opportunity to customize an open-source annotation tool called Hypothesis. This customization allowed for its use on all websites related to the Black Literature Network Project by the end users. The Black Book Interactive Project (BBIP) collaborates with institutions and groups nationwide to promote access to Black-authored texts and digital publishing. Through BBIP, we plan to increase black literature’s visibility in digital humanities research.