The 2016 ACM Workshop on Artificial Intelligence and Security will be co-located with CCS - the premier computer security conference. As the 9th workshop in the series, AISec 2016 calls for papers on topics related to both AI/learning and security/privacy.
Artificial Intelligence (AI), and Machine Learning (ML) in particular, provide a set of useful analytic and decision-making techniques that are being leveraged by an ever-growing community of practitioners, including applications with security-sensitive elements. However, while security researchers often utilize such techniques to address problems and AI/ML researchers develop techniques for big-data analytics applications, neither community devotes much attention to the other. Within security research, AI/ML components are often regarded as black-box solvers. Conversely, the learning community seldom considers the security/privacy implications entailed in the application of their algorithms when designing them. While these two communities generally focus on different issues, where these two fields do meet, interesting problems appear. Researchers working in the intersection have already raised many novel questions for both communities and created a new branch of research known as secure learning.
The past few years have particularly seen increasing interest within the AISec / Secure Learning community. There are several reasons for this surge. Firstly, machine learning, data mining, and other artificial intelligence technologies play a key role in extracting knowledge, situational awareness, and security intelligence from Big Data. Secondly, companies like Google, Amazon and Splunk are increasingly exploring and deploying learning technologies to address Big Data problems for their customers. Finally, these trends are increasingly exposing companies and their customers/users to intelligent technologies. As a result, these learning technologies are both being explored by researchers as potential solutions to security/privacy problems, and also are being investigated as a potential source of new privacy/security vulnerabilities that need to be secured to prevent them from misbehaving or leaking information to an adversary. The AISec Workshop meets this need and serves as the sole long-running venue for this topic.
AISec serves as the primary meeting place for diverse researchers in security, privacy, AI, and machine learning, and as a venue to develop the fundamental theory and practical applications supporting the use of machine learning for security and privacy. The workshop addresses on this burgeoning community who are especially focused on (among other topics) learning in game-theoretic adversarial environments, privacy-preserving learning, or use of sophisticated new learning algorithms in security.
Topics of interest include, but are not limited to:
Theoretical topics related to security
Adversarial Learning
Robust Statistics
Online Learning
Learning in games
Economics of security
Differential privacy
Security applications
Computer Forensics
Spam detection
Phishing detection and prevention
Botnet detection
Intrusion detection and response
Malware identification
Authorship Identification
Security-related AI problems
Distributed inference and decision making for security
Secure multiparty computation and cryptographic approaches
Privacy-preserving data mining
Adaptive side-channel attacks
Design and analysis of CAPTCHAs
AI approaches to trust and reputation
Vulnerability testing through intelligent probing (e.g. fuzzing)
Content-driven security policy management & access control
Techniques and methods for generating training and test sets
Anomalous behavior detection (e.g. for the purposes of fraud prevention, authentication)
Special topic: Machine learning and security at massive scale
High-throughput abuse detection systems
Large-scale active learning
Big data analytics for security
Systems defending against multiple attack vectors
Techniques dealing with well-resourced adversaries
10月28日
2016
会议日期
初稿截稿日期
初稿录用通知日期
终稿截稿日期
注册截止日期
留言