The goal was to detect if an application was malware based on sequence of API calls made to the OS. We explored several supervised methods (logistic regression, deep neural networks, SVM, LSTM and decision trees) as well as unsupervised models (clustering, autoencoders, DAGMM) but none of them showed reasonable results in a high-imbalance scenario (<5% malware). Finally, we developed a technique based on BERT, which leverages cross-domain transfer learning and shows an F-1 score of 0.92 with only 2% malware in the training data.
We developed a method called unlearning which facilitates model update without retraining, and can improve model performance by maximizing loss for false negatives. We propose a new loss function which incorporates learning rate shrinkage and Elastic Weight Consolidation (EWC) so that the model parameters change according to false positive and negative information. We validated the unlearning approach with three models: LSTM, regression and autoencoders to show that our method is model agnostic. We were able to reduce up to 77.3% false positives and 76.6% false negatives.
The goal was to design a system to classify DARPA emails into various nefarious categories (phishing, spam, malware) and develop generative models to automatically respond to such emails. We scraped email text and metadata from corpus provided by JPL/DARPA, and developed models for spam detection. We Leveraged statistical analysis of email text, authorship analysis, as well as advanced language models such as BERT, ELMO and XLNet.
[Paper] [Code] [Slides]
We leveraged language models to contribute an open source dataset consisting of 100k AI-generated news articles. We then explored a variety of semantic (function words, readability) and textual features (word embeddings, contextual embeddings) for classification. Applied classifiers such as SVM, Random Forest, Regression and deep neural networks and anomaly detection methods such as autoencoders to identify AI generated content. Our results show that AI generated text can be detected with high F-1 scores of over 90%.
[Paper] [Code] [Slides]
We investigate to what extent the increasing performance of deep neural networks is impacted by background features. In particular, we focus on background invariance (accuracy unaffected by switching background features) and background influence (predictive power of background features itself when foreground is masked). We perform experiments with 32 different neural networks including state-of-the-art models. Our investigations reveal that increasing expressive power of DNNs leads to higher influence of background features, while simultaneously, increases their ability to make correct predictions when background features are altered.