Machine Learning for Security and Security for Machine Learning with Nicole Nichols
EPISODE 252
|
APRIL
16,
2019
Watch
Follow
Share
About this Episode
Today we're joined by Nicole Nichols, a senior research scientist at the Pacific Northwest National Lab.
Nicole joined me to discuss her recent presentation at GTC, which was titled "Machine Learning for Security, and Security for Machine Learning." Our conversation explores the two use cases she presented, insider threat detection, and software fuzz testing. We discuss the effectiveness of standard and bidirectional RNN language models for detecting malicious activity within the Los Alamos National Laboratory cybersecurity dataset, the augmentation of software fuzzing techniques using deep learning, and light-based adversarial attacks on image classification systems. I'd love to hear your thoughts on these use cases!
About the Guest
Nicole Nichols
Pacific Northwest National Lab
Resources
- Paper: Recurrent Neural Network Language Models for Open Vocabulary Event-Level Cyber Anomaly Detection
- Paper: Deep Learning for Unsupervised Insider Threat Detection in Structured Cybersecurity Data Streams
- Paper: Faster Fuzzing: Reinitialization with Deep Neural Models
- Paper: Projecting Trouble: Light Based Adversarial Attacks on Deep Learning Classifiers
- Code: SafeKit
- Check out the rest of our GTC 2019 Series!
