Browsing by Subject "Machine Learning"
Results Per Page
Sort Options
Item Embargo Advancing Polyhydroxyalkanoate Biopolymer Material Design: Integrating Machine Learning and Experimental Validation(2024) Lalonde, Jessica NicoleVirtually every consumer product available on the market today contains some form of fossil fuel-based polymer. However, these materials pose environmental, human health, and economic concerns due to their enduring presence in the global ecosystem and their degradation products. Addressing this crisis necessitates scalable production of biodegradable alternatives, such as polyhydroxyalkanoates (PHAs). PHAs are presented as promising substitutes due to their biodegradability, biocompatibility, and the potential for complete renewable utilization post-degradation, but a current challenge to widespread use of these materials lies in understanding the quantitative relationship between the structural characteristics of PHAs, their environmental interactions, and their degradation rates to enhance their industrial production and distribution. To bridge this knowledge gap, the dissertation outlines a comprehensive approach involving the development of a specialized dataset, the application of machine learning (ML) models to predict degradation rates based on structural and environmental factors, and the experimental validation of these predictions. The first part of this research focuses on assembling a manually curated dataset from the extensive, available open-access literature, aimed at understanding the effects of structural and environmental features on PHA degradation. The second part leverages this dataset through ML modeling, employing techniques like random forest regression to predict degradation profiles with over 80% accuracy. This methodology enables a deeper understanding of the complex interplay between chemical structures and degradation properties, surpassing traditional trial-and-error approaches. The final part of this research aims to complete an iterative workflow for dataset development by validating ML model predictions through physical experiments, enriching the original dataset with comprehensive experimental data on PHA degradation in hydrolytic environments with contact angle, molecular weight, and thermal property characterizations. The incorporation of experimental findings into the ML dataset, particularly through expanded ML techniques that emphasize pairwise feature importance such as explainable boosting machines (EBM), helps in pinpointing critical factors influencing PHA degradation, such as environmental temperature and material properties. The model performances indicate a strong performance of manually assembled literature-based datasets when predicting degradation rate for PHAs. In conclusion, a data science-based framework has been developed for exploring PHA biopolyester degradation and explores the combination of features of the material and its environment that integrates the structure, properties, and experimentally verified degradation profiles of the material. This workflow will be a useful and generalizable pipeline for PHAs and other polymers to expand the biopolymer design space with degradation in mind.
Item Open Access ADVANCING VISION INTELLIGENCE THROUGH THE DEVELOPMENT OF EFFICIENCY, INTERPRETABILITY AND FAIRNESS IN DEEP LEARNING MODELS(2024) Kong, FanjieDeep learning has demonstrated remarkable success in developing vision intelligence across a variety of application domains, including autonomous driving, facial recognition, medical image analysis, \etc.However, developing such vision systems poses significant challenges, particularly in relation to ensuring efficiency, interpretability, and fairness. Efficiency requires a model to leverage the least possible computational resources while preserving performance relative to more computationally-demanding alternatives, which is essential for the practical deployment of large-scale models in real-time applications. Interpretability demands a model to align with the domain-specific knowledge of the task it addresses while having the capability for case-based reasoning. This characteristic is especially crucial in high-stakes areas such as healthcare, criminal justice, and financial investment. Fairness ensures that computer vision models do not perpetuate or exacerbate societal biases in downstream applications such as web image search, text-guided image generation, \etc. In this dissertation, I will discuss the contributions that I have made in advancing vision intelligence regarding to efficiency, interpretability and fairness in computer vision models.
The first part of this dissertation will focus on how to design computer vision models to efficiently process very large images.We propose a novel CNN architecture termed { \em Zoom-In Network} that leverages a hierarchical attention sampling mechanisms to select important regions of images to process. Such approach without processing the entire image yields outstanding memory efficiency while maintaining classification accuracy on various tiny object image classification datasets.
The second part of this dissertation will discuss how to build post-hoc interpretation method for deep learning models to obtain insights reasoned from the predictions.We propose a novel image and text insight-generation framework based on attributions from deep neural nets. We test our approach on an industrial dataset and demonstrate our method outperforms competing methods.
Finally, we study fairness in large vision-language models.More specifically, we examined gender and racial bias in text-based image retrieval for neutral text queries. In an attempt to address bias in the test-time phase, we proposed post-hoc bias mitigation to actively balance the demographic group in the image search results. Experiments on multiple datasets show that our method can significantly reduce bias while maintaining satisfactory retrieval accuracy at the same time.
My research in enhancing vision intelligence via developments in efficiency, interpretability, and fairness, has undergone rigorous validation using publicly available benchmarks and has been recognized at leading peer-reviewed machine learning conferences.This dissertation has sparked interest within the AI community, emphasizing the importance of improving computer vision models through these three critical dimensions, namely, efficiency, interpretability and fairness.
Item Open Access Automated Detection of P. falciparum Using Machine Learning Algorithms with Quantitative Phase Images of Unstained Cells.(PloS one, 2016-01) Park, Han Sang; Rinehart, Matthew T; Walzer, Katelyn A; Chi, Jen-Tsan Ashley; Wax, AdamMalaria detection through microscopic examination of stained blood smears is a diagnostic challenge that heavily relies on the expertise of trained microscopists. This paper presents an automated analysis method for detection and staging of red blood cells infected by the malaria parasite Plasmodium falciparum at trophozoite or schizont stage. Unlike previous efforts in this area, this study uses quantitative phase images of unstained cells. Erythrocytes are automatically segmented using thresholds of optical phase and refocused to enable quantitative comparison of phase images. Refocused images are analyzed to extract 23 morphological descriptors based on the phase information. While all individual descriptors are highly statistically different between infected and uninfected cells, each descriptor does not enable separation of populations at a level satisfactory for clinical utility. To improve the diagnostic capacity, we applied various machine learning techniques, including linear discriminant classification (LDC), logistic regression (LR), and k-nearest neighbor classification (NNC), to formulate algorithms that combine all of the calculated physical parameters to distinguish cells more effectively. Results show that LDC provides the highest accuracy of up to 99.7% in detecting schizont stage infected cells compared to uninfected RBCs. NNC showed slightly better accuracy (99.5%) than either LDC (99.0%) or LR (99.1%) for discriminating late trophozoites from uninfected RBCs. However, for early trophozoites, LDC produced the best accuracy of 98%. Discrimination of infection stage was less accurate, producing high specificity (99.8%) but only 45.0%-66.8% sensitivity with early trophozoites most often mistaken for late trophozoite or schizont stage and late trophozoite and schizont stage most often confused for each other. Overall, this methodology points to a significant clinical potential of using quantitative phase imaging to detect and stage malaria infection without staining or expert analysis.Item Embargo Benign and Malignant Lymph Nodes Classification in Non-Small Cell Lung Cancer via Machine Learning Model(2024) Ge, JingyuObjective:To develop a machine learning model that integrates deep learning image features and radiomics features to classify lymph nodes as benign or malignant in Non-Small Cell Lung Cancer (NSCLC). Methods: The dataset comprises contrast-enhanced CT scans from 541 lung cancer patients before surgery, collected at a Shanghai Hospital between July 2015 and December 2017 under an IRB study. It includes 1,237 lymph nodes, identified from preoperative CT scans due to enlargement and confirmed as non-small cell lung cancer (NSCLC) via surgical pathology. Lymph node classification into malignant or benign categories utilized in postoperative pathological reports. Our method employs a dual radiomic feature extraction strategy. The deep image features (DIF) were derived from the final convolutional layer of a pre-trained VGG-16 encoder network to characterize the lymph node’s image texture. A total of nine 2D shape-based radiomic features (RF) are extracted based on the Py-radiomics calculation toolbox to characterize lymph node morphological information. And ninety-two handcrafted radiomic features (HRF) are extracted. The extracted DIF, RF, and HRF were combined and fed into a Random Forest classifier for the benign and malignant lymph node classification. The random forest classifier was trained following an 8:2 train/test split ratio and evaluated using Area Under the Curve (AUC), Receiver Operating Characteristic (ROC), and p-value, and 5-fold cross-validation was also employed to objectively evaluate model performance.
Results: The mean AUC for the Random Forest classifier using only 2D shape features is 0.691, while mean AUC for the classifier employing only DIF is 726. Utilizing both DIF and HRF for classification resulted in an average AUC of 0.724, whereas integrating RF with DIF achieved superior classification efficacy, boasting the highest average AUC of 0.746. All results were considered statistically significant with a p-value of less than 0.05. Conclusion: The combination of image texture analysis refers to DIF with morphological information offers an enhanced characterization ability to classify lymph nodes as benign or malignant from CT images for lung NSCLC patients.
Item Open Access Combining adult with pediatric patient data to develop a clinical decision support tool intended for children: leveraging machine learning to model heterogeneity.(BMC medical informatics and decision making, 2022-03) Sabharwal, Paul; Hurst, Jillian H; Tejwani, Rohit; Hobbs, Kevin T; Routh, Jonathan C; Goldstein, Benjamin ABackground
Clinical decision support (CDS) tools built using adult data do not typically perform well for children. We explored how best to leverage adult data to improve the performance of such tools. This study assesses whether it is better to build CDS tools for children using data from children alone or to use combined data from both adults and children.Methods
Retrospective cohort using data from 2017 to 2020. Participants include all individuals (adults and children) receiving an elective surgery at a large academic medical center that provides adult and pediatric services. We predicted need for mechanical ventilation or admission to the intensive care unit (ICU). Predictor variables included demographic, clinical, and service utilization factors known prior to surgery. We compared predictive models built using machine learning to regression-based methods that used a pediatric or combined adult-pediatric cohort. We compared model performance based on Area Under the Receiver Operator Characteristic.Results
While we found that adults and children have different risk factors, machine learning methods are able to appropriately model the underlying heterogeneity of each population and produce equally accurate predictive models whether using data only from pediatric patients or combined data from both children and adults. Results from regression-based methods were improved by the use of pediatric-specific data.Conclusions
CDS tools for children can successfully use combined data from adults and children if the model accounts for underlying heterogeneity, as in machine learning models.Item Open Access Developing nonlinear k-nearest neighbors classification algorithms to identify patients at high risk of increased length of hospital stay following spine surgery.(Neurosurgical focus, 2023-06) Shahrestani, Shane; Chan, Andrew K; Bisson, Erica F; Bydon, Mohamad; Glassman, Steven D; Foley, Kevin T; Shaffrey, Christopher I; Potts, Eric A; Shaffrey, Mark E; Coric, Domagoj; Knightly, John J; Park, Paul; Wang, Michael Y; Fu, Kai-Ming; Slotkin, Jonathan R; Asher, Anthony L; Virk, Michael S; Michalopoulos, Giorgos D; Guan, Jian; Haid, Regis W; Agarwal, Nitin; Chou, Dean; Mummaneni, Praveen VObjective
Spondylolisthesis is a common operative disease in the United States, but robust predictive models for patient outcomes remain limited. The development of models that accurately predict postoperative outcomes would be useful to help identify patients at risk of complicated postoperative courses and determine appropriate healthcare and resource utilization for patients. As such, the purpose of this study was to develop k-nearest neighbors (KNN) classification algorithms to identify patients at increased risk for extended hospital length of stay (LOS) following neurosurgical intervention for spondylolisthesis.Methods
The Quality Outcomes Database (QOD) spondylolisthesis data set was queried for patients receiving either decompression alone or decompression plus fusion for degenerative spondylolisthesis. Preoperative and perioperative variables were queried, and Mann-Whitney U-tests were performed to identify which variables would be included in the machine learning models. Two KNN models were implemented (k = 25) with a standard training set of 60%, validation set of 20%, and testing set of 20%, one with arthrodesis status (model 1) and the other without (model 2). Feature scaling was implemented during the preprocessing stage to standardize the independent features.Results
Of 608 enrolled patients, 544 met prespecified inclusion criteria. The mean age of all patients was 61.9 ± 12.1 years (± SD), and 309 (56.8%) patients were female. The model 1 KNN had an overall accuracy of 98.1%, sensitivity of 100%, specificity of 84.6%, positive predictive value (PPV) of 97.9%, and negative predictive value (NPV) of 100%. Additionally, a receiver operating characteristic (ROC) curve was plotted for model 1, showing an overall area under the curve (AUC) of 0.998. Model 2 had an overall accuracy of 99.1%, sensitivity of 100%, specificity of 92.3%, PPV of 99.0%, and NPV of 100%, with the same ROC AUC of 0.998.Conclusions
Overall, these findings demonstrate that nonlinear KNN machine learning models have incredibly high predictive value for LOS. Important predictor variables include diabetes, osteoporosis, socioeconomic quartile, duration of surgery, estimated blood loss during surgery, patient educational status, American Society of Anesthesiologists grade, BMI, insurance status, smoking status, sex, and age. These models may be considered for external validation by spine surgeons to aid in patient selection and management, resource utilization, and preoperative surgical planning.Item Open Access Effect of machine learning methods on predicting NSCLC overall survival time based on Radiomics analysis.(Radiation oncology (London, England), 2018-10-05) Sun, Wenzheng; Jiang, Mingyan; Dang, Jun; Chang, Panchun; Yin, Fang-FangBACKGROUND:To investigate the effect of machine learning methods on predicting the Overall Survival (OS) for non-small cell lung cancer based on radiomics features analysis. METHODS:A total of 339 radiomic features were extracted from the segmented tumor volumes of pretreatment computed tomography (CT) images. These radiomic features quantify the tumor phenotypic characteristics on the medical images using tumor shape and size, the intensity statistics and the textures. The performance of 5 feature selection methods and 8 machine learning methods were investigated for OS prediction. The predicted performance was evaluated with concordance index between predicted and true OS for the non-small cell lung cancer patients. The survival curves were evaluated by the Kaplan-Meier algorithm and compared by the log-rank tests. RESULTS:The gradient boosting linear models based on Cox's partial likelihood method using the concordance index feature selection method obtained the best performance (Concordance Index: 0.68, 95% Confidence Interval: 0.62~ 0.74). CONCLUSIONS:The preliminary results demonstrated that certain machine learning and radiomics analysis method could predict OS of non-small cell lung cancer accuracy.Item Open Access FlowKit: A Python Toolkit for Integrated Manual and Automated Cytometry Analysis Workflows.(Frontiers in immunology, 2021-01) White, Scott; Quinn, John; Enzor, Jennifer; Staats, Janet; Mosier, Sarah M; Almarode, James; Denny, Thomas N; Weinhold, Kent J; Ferrari, Guido; Chan, CliburnAn important challenge for primary or secondary analysis of cytometry data is how to facilitate productive collaboration between domain and quantitative experts. Domain experts in cytometry laboratories and core facilities increasingly recognize the need for automated workflows in the face of increasing data complexity, but by and large, still conduct all analysis using traditional applications, predominantly FlowJo. To a large extent, this cuts domain experts off from the rapidly growing library of Single Cell Data Science algorithms available, curtailing the potential contributions of these experts to the validation and interpretation of results. To address this challenge, we developed FlowKit, a Gating-ML 2.0-compliant Python package that can read and write FCS files and FlowJo workspaces. We present examples of the use of FlowKit for constructing reporting and analysis workflows, including round-tripping results to and from FlowJo for joint analysis by both domain and quantitative experts.Item Open Access Identifying treatment effects of an informal caregiver education intervention to increase days in the community and decrease caregiver distress: a machine-learning secondary analysis of subgroup effects in the HI-FIVES randomized clinical trial.(Trials, 2020-02) Shepherd-Banigan, Megan; Smith, Valerie A; Lindquist, Jennifer H; Cary, Michael Paul; Miller, Katherine EM; Chapman, Jennifer G; Van Houtven, Courtney HBackground
Informal caregivers report substantial burden and depressive symptoms which predict higher rates of patient institutionalization. While caregiver education interventions may reduce caregiver distress and decrease the use of long-term institutional care, evidence is mixed. Inconsistent findings across studies may be the result of reporting average treatment effects which do not account for how effects differ by participant characteristics. We apply a machine-learning approach to randomized clinical trial (RCT) data of the Helping Invested Family Members Improve Veteran's Experiences Study (HI-FIVES) intervention to explore how intervention effects vary by caregiver and patient characteristics.Methods
We used model-based recursive partitioning models. Caregivers of community-residing older adult US veterans with functional or cognitive impairment at a single VA Medical Center site were randomized to receive HI-FIVES (n = 118) vs. usual care (n = 123). The outcomes included cumulative days not in the community and caregiver depressive symptoms assessed at 12 months post intervention. Potential moderating characteristics were: veteran age, caregiver age, caregiver ethnicity and race, relationship satisfaction, caregiver burden, perceived financial strain, caregiver depressive symptoms, and patient risk score.Results
The effect of HI-FIVES on days not at home was moderated by caregiver burden (p < 0.001); treatment effects were higher for caregivers with a Zarit Burden Scale score ≤ 28. Caregivers with lower baseline Center for Epidemiologic Studies Depression Scale (CESD-10) scores (≤ 8) had slightly lower CESD-10 scores at follow-up (p < 0.001).Conclusions
Family caregiver education interventions may be less beneficial for highly burdened and distressed caregivers; these caregivers may require a more tailored approach that involves assessing caregiver needs and developing personalized approaches.Trial registration
ClinicalTrials.gov, ID:NCT01777490. Registered on 28 January 2013.Item Open Access Improved AlphaFold modeling with implicit experimental information.(Nature methods, 2022-11) Terwilliger, Thomas C; Poon, Billy K; Afonine, Pavel V; Schlicksup, Christopher J; Croll, Tristan I; Millán, Claudia; Richardson, Jane S; Read, Randy J; Adams, Paul DMachine-learning prediction algorithms such as AlphaFold and RoseTTAFold can create remarkably accurate protein models, but these models usually have some regions that are predicted with low confidence or poor accuracy. We hypothesized that by implicitly including new experimental information such as a density map, a greater portion of a model could be predicted accurately, and that this might synergistically improve parts of the model that were not fully addressed by either machine learning or experiment alone. An iterative procedure was developed in which AlphaFold models are automatically rebuilt on the basis of experimental density maps and the rebuilt models are used as templates in new AlphaFold predictions. We show that including experimental information improves prediction beyond the improvement obtained with simple rebuilding guided by the experimental data. This procedure for AlphaFold modeling with density has been incorporated into an automated procedure for interpretation of crystallographic and electron cryo-microscopy maps.Item Open Access Improved Detection of Invasive Pulmonary Aspergillosis Arising during Leukemia Treatment Using a Panel of Host Response Proteins and Fungal Antigens.(PloS one, 2015-01) Brasier, Allan R; Zhao, Yingxin; Spratt, Heidi M; Wiktorowicz, John E; Ju, Hyunsu; Wheat, L Joseph; Baden, Lindsey; Stafford, Susan; Wu, Zheng; Issa, Nicolas; Caliendo, Angela M; Denning, David W; Soman, Kizhake; Clancy, Cornelius J; Nguyen, M Hong; Sugrue, Michele W; Alexander, Barbara D; Wingard, John RInvasive pulmonary aspergillosis (IPA) is an opportunistic fungal infection in patients undergoing chemotherapy for hematological malignancy, hematopoietic stem cell transplant, or other forms of immunosuppression. In this group, Aspergillus infections account for the majority of deaths due to mold pathogens. Although early detection is associated with improved outcomes, current diagnostic regimens lack sensitivity and specificity. Patients undergoing chemotherapy, stem cell transplantation and lung transplantation were enrolled in a multi-site prospective observational trial. Proven and probable IPA cases and matched controls were subjected to discovery proteomics analyses using a biofluid analysis platform, fractionating plasma into reproducible protein and peptide pools. From 556 spots identified by 2D gel electrophoresis, 66 differentially expressed post-translationally modified plasma proteins were identified in the leukemic subgroup only. This protein group was rich in complement components, acute-phase reactants and coagulation factors. Low molecular weight peptides corresponding to abundant plasma proteins were identified. A candidate marker panel of host response (9 plasma proteins, 4 peptides), fungal polysaccharides (galactomannan), and cell wall components (β-D glucan) were selected by statistical filtering for patients with leukemia as a primary underlying diagnosis. Quantitative measurements were developed to qualify the differential expression of the candidate host response proteins using selective reaction monitoring mass spectrometry assays, and then applied to a separate cohort of 57 patients with leukemia. In this verification cohort, a machine learning ensemble-based algorithm, generalized pathseeker (GPS) produced a greater case classification accuracy than galactomannan (GM) or host proteins alone. In conclusion, Integration of host response proteins with GM improves the diagnostic detection of probable IPA in patients undergoing treatment for hematologic malignancy. Upon further validation, early detection of probable IPA in leukemia treatment will provide opportunities for earlier interventions and interventional clinical trials.Item Open Access Machine learning functional impairment classification with electronic health record data.(Journal of the American Geriatrics Society, 2023-09) Pavon, Juliessa M; Previll, Laura; Woo, Myung; Henao, Ricardo; Solomon, Mary; Rogers, Ursula; Olson, Andrew; Fischer, Jonathan; Leo, Christopher; Fillenbaum, Gerda; Hoenig, Helen; Casarett, DavidBackground
Poor functional status is a key marker of morbidity, yet is not routinely captured in clinical encounters. We developed and evaluated the accuracy of a machine learning algorithm that leveraged electronic health record (EHR) data to provide a scalable process for identification of functional impairment.Methods
We identified a cohort of patients with an electronically captured screening measure of functional status (Older Americans Resources and Services ADL/IADL) between 2018 and 2020 (N = 6484). Patients were classified using unsupervised learning K means and t-distributed Stochastic Neighbor Embedding into normal function (NF), mild to moderate functional impairment (MFI), and severe functional impairment (SFI) states. Using 11 EHR clinical variable domains (832 variable input features), we trained an Extreme Gradient Boosting supervised machine learning algorithm to distinguish functional status states, and measured prediction accuracies. Data were randomly split into training (80%) and test (20%) sets. The SHapley Additive Explanations (SHAP) feature importance analysis was used to list the EHR features in rank order of their contribution to the outcome.Results
Median age was 75.3 years, 62% female, 60% White. Patients were classified as 53% NF (n = 3453), 30% MFI (n = 1947), and 17% SFI (n = 1084). Summary of model performance for identifying functional status state (NF, MFI, SFI) was AUROC (area under the receiving operating characteristic curve) 0.92, 0.89, and 0.87, respectively. Age, falls, hospitalization, home health use, labs (e.g., albumin), comorbidities (e.g., dementia, heart failure, chronic kidney disease, chronic pain), and social determinants of health (e.g., alcohol use) were highly ranked features in predicting functional status states.Conclusion
A machine learning algorithm run on EHR clinical data has potential utility for differentiating functional status in the clinical setting. Through further validation and refinement, such algorithms can complement traditional screening methods and result in a population-based strategy for identifying patients with poor functional status who need additional health resources.Item Open Access Machine learning risk prediction of mortality for patients undergoing surgery with perioperative SARS-CoV-2: the COVIDSurg mortality score.(The British journal of surgery, 2021-11) COVIDSurg CollaborativeTo support the global restart of elective surgery, data from an international prospective cohort study of 8492 patients (69 countries) was analysed using artificial intelligence (machine learning techniques) to develop a predictive score for mortality in surgical patients with SARS-CoV-2. We found that patient rather than operation factors were the best predictors and used these to create the COVIDsurg Mortality Score (https://covidsurgrisk.app). Our data demonstrates that it is safe to restart a wide range of surgical services for selected patients.Item Open Access Pharmacogenomics-Driven Prediction of Antidepressant Treatment Outcomes: A Machine-Learning Approach With Multi-trial Replication.(Clinical pharmacology and therapeutics, 2019-10) Athreya, Arjun P; Neavin, Drew; Carrillo-Roa, Tania; Skime, Michelle; Biernacka, Joanna; Frye, Mark A; Rush, A John; Wang, Liewei; Binder, Elisabeth B; Iyer, Ravishankar K; Weinshilboum, Richard M; Bobo, William VWe set out to determine whether machine learning-based algorithms that included functionally validated pharmacogenomic biomarkers joined with clinical measures could predict selective serotonin reuptake inhibitor (SSRI) remission/response in patients with major depressive disorder (MDD). We studied 1,030 white outpatients with MDD treated with citalopram/escitalopram in the Mayo Clinic Pharmacogenomics Research Network Antidepressant Medication Pharmacogenomic Study (PGRN-AMPS; n = 398), Sequenced Treatment Alternatives to Relieve Depression (STAR*D; n = 467), and International SSRI Pharmacogenomics Consortium (ISPC; n = 165) trials. A genomewide association study for PGRN-AMPS plasma metabolites associated with SSRI response (serotonin) and baseline MDD severity (kynurenine) identified single nucleotide polymorphisms (SNPs) in DEFB1, ERICH3, AHR, and TSPAN5 that we tested as predictors. Supervised machine-learning methods trained using SNPs and total baseline depression scores predicted remission and response at 8 weeks with area under the receiver operating curve (AUC) > 0.7 (P < 0.04) in PGRN-AMPS patients, with comparable prediction accuracies > 69% (P ≤ 0.07) in STAR*D and ISPC. These results demonstrate that machine learning can achieve accurate and, importantly, replicable prediction of SSRI therapy response using total baseline depression severity combined with pharmacogenomic biomarkers.Item Open Access Practical Solutions to Neural Architecture Search on Applied Machine Learning(2024) Zhang, TunhouThe advent of Artificial Intelligence (AI) propels the real world into a new era characterized by remarkable design innovations and groundbreaking design automation, primarily fueled by Deep Neural Networks (DNN). At the heart of this transformation is the progress in Automated Machine Learning (AutoML), notably Neural Architecture Search (NAS). NAS lays a robust foundation for developing algorithms capable of automating design processes to determine the optimal architecture for academic benchmarks. However, the real challenge emerges when adapting NAS for Applied Machine Learning (AML) scenarios: navigating the complex terrain of design space exploration and exploitation. This complexity arises due to the heterogeneity of data and architectures required by real-world AML problems, an aspect that traditional NAS approaches struggle to address fully.
To bridge this gap, our research emphasizes creating a flexible search space that reduces reliance on human-derived architectural assumptions. We introduce innovative techniques aimed at refining search algorithms to accommodate greater flexibility. By carefully examining and enhancing search spaces and methodologies, we empower NAS solutions to cater to practical AML problems. This enables the exploration of broader search spaces, better performance potential, and lower search process costs.
We start by challenging homogeneous search space design for multi-modality 3D representations, proposing ``PIDS'' to enable joint dimension and interaction search for 3D point cloud segmentation. We consider two axes on adapting point cloud operators toward multi-modality data with density, geometry, and order varieties, achieving significant mIOU improvement on segmentation benchmarks over the state-of-the-art 3D models.To implement our approach efficiently in recommendation systems, we develop ``NASRec'' to support heterogeneous building operators and propose practical solutions to improve the quality of NAS on Click-Through Rates (CTR) prediction. We propose an end-to-end full architecture search with minimal human priors. We provide practical solutions to tackle scalability and heterogeneity challenges in NAS, outperforming manually designed models and existing NAS models on various CTR benchmarks. Finally, we pioneer our effort on industry-scale CTR benchmarks and propose DistDNAS to optimize search and serving efficiency, producing smaller and better recommendation models on a large-scale CTR benchmark. Intuited by the discoveries in NAS, we additionally uncover the underlying theoretical foundations of residual learning on computer vision foundation research and envision the prospects of our research on Artificial Intelligence, including Large Language Models, Generative AI, and beyond.
Item Open Access Predicting Student Performance Using Discussion Forums' Participation Data(2024) Gray, McCullough JosephA significant gap in education lies in the need for mechanisms that enable early detection of potentially at-risk students. Through access to an earlier prediction of student performance, instructors are given ample time to meet with and assist under-achieving students. As with any prediction modeling problem, there are many predictors to choose from when formulating a model. Previous related works have shown limited success in predicting course performance using students' personal and socioeconomic traits. Students learn by asking clarifying questions. Therefore, discussion boards have been a staple of learning at the university level for years.
This research aims to utilize participation in discussion forums to predict final student performance. Using students' course grades at roughly the halfway point in the term and various discussion forum predictors, our model predicts the students' final percentage score. Using the model's prediction, instructors can speak with at-risk students and discuss ways to improve. The student grades and discussion board participation datasets are gathered from graduate-level Electrical and Computer Engineering (ECE) courses at Duke University. Various classical machine learning models are explored, with random forest yielding the highest accuracy. This random forest model, trained on discussion forum participation data, surpasses other similarly trained state-of-the-art models.
Furthermore, related research attempts the classification problem of predicting what discrete letter grade a student will earn. This is not an accurate representation of a student's performance, and therefore, we attempt the regression problem of predicting the exact percentage a student will earn. A significant finding of this research is that our random forest model can predict student performance with an average error of approximately 2.3%. Additionally, our random forest model can generalize to a different graduate-level course and make performance predictions with an average error of 3.3%.
The final important finding is that a model including discussion board predictors outperforms another whose sole predictor is the students' halfway point grade. This indicates that discussion forums hold significant value in determining final performance. We envision that the knowledge from our findings and our optimal random forest model can enable instructors to identify and support potentially at-risk students preemptively.
Item Open Access Prediction of Stock Market Price Index using Machine Learning and Global Trade Information(2020-10-30) Wong, Eugene Lu XianGlobalization has led to an increasingly integrated global economy, one with less trade barriers and more capital mobility between countries. Consequently, no country is an island of its own. In this paper, it aims to investigate how global trade affects a country's stock market and also determine if such information with the use of machine learning techniques can predict a country's stock market index.Item Embargo Quantifying Radiomic Texture Characterization Performance on Image Resampling and Discretization(2024) Sang, WeiweiPurpose: To develop a novel radiomic quantification framework to quantify the impact of image resampling and discretization on radiomic texture characterization performance.
Methods: The study employed 251 CT scans of a Credence Cartridge phantom (consisting of 10 texture materials) with different image acquisition parameters. Each material was segmented using a pre-defined cylindrical mask. Different image pre-processing workflows including 5 resampling methods (no resampling, trilinear, and nearest resampling to both 1mm³ and 5mm³) and 8 discretization methods (fixed bin size of 25,50,75,100 and fixed bin counts of 8,16,32,64) were randomly applied. 75 radiomic texture features (including 24GLCM-based, 16GLRLM-based, 16GLSZM-based, 14GLDM-based, and 5NGTDM-based) were extracted from each material to characterize its textural attributes. Three machine learning models including logistic regression (LR), random forest (RF), and supporting vector machine (SVM) were developed to identify 10 materials based on the extracted features, and grid search was adopted to optimize the model hyperparameters. The model performance was evaluated on 10-class macro-AUC with 5-fold cross-validation.
Results: Three models successfully classified 10 materials with macro-AUC=0.9941±0.0081, 0.9979±0.0040, and 0.9957±0.0067 for LR, RF, and SVM, respectively. Across 8 different discretization methods, an increasing trend in performance can be observed when the original CT was discretized to a larger gray level range: performance improved by 0.0038 with bin sizes decreasing from 100-25, and by 0.0074 with bin counts increasing from 8 to 64. Among 5 resampling methods, resampling CT to an isotropic voxel spacing showed an improved prediction performance (0.9942±0.0075/0.9944±0.0073 for trilinear/nearest resampling to 1mm³ and 5mm³, respectively) over no interpolation (0.9862±0.0228), with minimal performance discrepancies observed among two different interpolation algorithms. In addition, no statistically significant differences were observed across five folds.
Conclusion: The proposed framework successfully quantified the dependence of radiomics texture characterization on image resampling and discretization.
Item Open Access Radiomics analysis using stability selection supervised component analysis for right-censored survival data.(Computers in biology and medicine, 2020-09) Yan, Kang K; Wang, Xiaofei; Lam, Wendy WT; Vardhanabhuti, Varut; Lee, Anne WM; Pang, Herbert HRadiomics is a newly emerging field that involves the extraction of massive quantitative features from biomedical images by using data-characterization algorithms. Distinctive imaging features identified from biomedical images can be used for prognosis and therapeutic response prediction, and they can provide a noninvasive approach for personalized therapy. So far, many of the published radiomics studies utilize existing out of the box algorithms to identify the prognostic markers from biomedical images that are not specific to radiomics data. To better utilize biomedical images, we propose a novel machine learning approach, stability selection supervised principal component analysis (SSSuperPCA) that identifies stable features from radiomics big data coupled with dimension reduction for right-censored survival outcomes. The proposed approach allows us to identify a set of stable features that are highly associated with the survival outcomes in a simple yet meaningful manner, while controlling the per-family error rate. We evaluate the performance of SSSuperPCA using simulations and real data sets for non-small cell lung cancer and head and neck cancer, and compare it with other machine learning algorithms. The results demonstrate that our method has a competitive edge over other existing methods in identifying the prognostic markers from biomedical imaging data for the prediction of right-censored survival outcomes.Item Open Access The Press and Peace(2024-05-10) Bussey, JakobeThis study utilizes state-of-the-art BERT (Bidirectional Encoder Representations from Transformers) models to perform sentiment analysis on Wall Street Journal and New York Times articles about the Iraq War published between 2002 and 2012 and further categorize them using advanced unsupervised machine learning techniques. By utilizing statistical analysis and quartic regression models, this paper concludes that the two newspapers report on the Iraq War differently, with both exhibiting a predominantly negative-neutral tone overall. Additionally, the analysis reveals significant fluctuations in negativity from both outlets over time as the war progresses. Furthermore, this study examines the objectivity of reporting between editorial and non-editorial articles, finding that non-editorials tend to report more objectively, and the neutrality of editorials remains relatively constant while the objectivity of non-editorials fluctuates in response to war events. Finally, the paper investigates variations in sentiment across different topics, uncovering substantial variations in positive, neutral, and negative sentiments across topics and their evolution over time.