📍🔗515 W Hastings St, Vancouver, BC V6B 4N6
[right beside the Harbour Centre, Downtown Vancouver]
(10-min walk from Vancouver Convention Centre)
Researchers across Simon Fraser University (SFU) are dedicated to advancing the field of machine learning (ML), spanning from innovative applications to foundational theory. SFU’s work includes groundbreaking research in areas such as visual computing, language processing, reinforcement learning, and beyond. By publishing extensively, collaborating across disciplines, and engaging with the global ML community, SFU aims to foster a rich, collaborative ecosystem that pushes the boundaries of what's possible in ML.
SFU VINCI Institute (Visual and INteractive Computing): https://vinci.sfu.ca/
Machine Learning Team @ SFU: https://ml.cs.sfu.ca/
GrUVi (Graphics U Vision) Team @ SFU: https://gruvi.cs.sfu.ca/
This year, SFU is proud to host a parallel event along with NeurIPS. This special SFU event will feature invited talks and social events designed to complement the NeurIPS experience. We warmly invite NeurIPS attendees to join us for these sessions, connecting with our speakers and researchers, and engaging in thought-provoking discussions about the latest advances in AI/ML.
See our schedule below. You can find more information about SFU’s contributions to NeurIPS 2024 at "SFU@NeurIPS".
Wed Dec. 11
USC
Towards Robust Al: Advances in Outlier and Out-of-Distribution Detection
TBD
Wed Dec. 11
Stanford
Concept Learning Across Domains and Modalities
I will discuss a concept-centric paradigm for building agents that can learn continually and reason flexibly across multiple domains and input modalities. The concept-centric agent utilizes a vocabulary of neuro-symbolic concepts. These concepts, including object, relation, and action concepts, are grounded on sensory inputs and actuation outputs. They are also compositional, allowing for the creation of novel concepts through their structural combination. To facilitate learning and reasoning, the concepts are typed and represented using a combination of symbolic programs and neural network representations. Leveraging such neuro-symbolic concepts, the agent can efficiently learn and recombine them to solve various tasks across different domains and data modalities, ranging from 2D images, videos, 3D scenes, temporal data, and robotic manipulation data.
Wed Dec. 11
UW Madison
Voyager: LLM-powered Lifelong Learning Agent
Thu Dec. 12
Towards Data-efficient Training of Large Language Models (LLMs)
High quality data is crucial for training LLMs with superior performance. In this talk, I will present two theoretically-rigorous approaches to find smaller subsets of examples that can improve the performance and efficiency of training LLMs. First, I will present a one-shot data selection method for supervised fine-tuning of LLMs. Then, I'll talk about an iterative data selection strategy to pretrain or fine-tune LLMs on imbalanced mixtures of language data. I'll conclude by showing empirical results confirming that the above data selection strategies can effectively improve the performance of various LLMs during fine-tuning and pretraining.
Thu Dec. 12
Generative World Modeling for Embodied Agents
Generative models have transformed content creation, and the next frontier may be simulating realistic experiences in response to actions by humans and agents. In this talk, I will talk about a line of work that involves learning a real-world simulator (i.e., a world model) to emulate interactions through generative modeling of video content. I will then talk about the applications of using this world model to train embodied agents through reinforcement learning (RL) and planning, which have demonstrated zero-shot real-world transfer. Lastly, I will talk about how to improve generative world models from real-world feedback.
Thu Dec. 12
EPFL, Switzerland
Jailbreaking Leading Safety-Aligned LLMs with Simple Adaptive Attacks
We show that even the most recent safety-aligned LLMs are not robust to simple adaptive jailbreaking attacks. First, we demonstrate how to successfully leverage access to logprobs for jailbreaking: we initially design an adversarial prompt template (sometimes adapted to the target LLM), and then we apply random search on a suffix to maximize a target logprob (e.g., of the token "Sure"), potentially with multiple restarts. In this way, we achieve 100% attack success rate -- according to GPT-4 as a judge -- on Vicuna-13B, Mistral-7B, Phi-3-Mini, Nemotron-4-340B, Llama-2-Chat-7B/13B/70B, Llama-3-Instruct-8B, Gemma-7B, GPT-3.5, GPT-4o, and R2D2 from HarmBench that was adversarially trained against the GCG attack. We also show how to jailbreak all Claude models -- that do not expose logprobs -- via either a transfer or prefilling attack with a 100% success rate. In addition, we show how to use random search on a restricted set of tokens for finding trojan strings in poisoned models -- a task that shares many similarities with jailbreaking -- which is the algorithm that brought us the first place in the SaTML'24 Trojan Detection Competition. The common theme behind these attacks is that adaptivity is crucial: different models are vulnerable to different prompting templates (e.g., R2D2 is very sensitive to in-context learning prompts), some models have unique vulnerabilities based on their APIs (e.g., prefilling for Claude), and in some settings, it is crucial to restrict the token search space based on prior knowledge (e.g., for trojan detection). For reproducibility purposes, we provide the code, logs, and jailbreak artifacts in the JailbreakBench format at https://github.com/tml-epfl/llm-adaptive-attacks.
Fri Dec. 13
NVIDIA
Efficiency in Large Language Models with Post-Training Compression
TBD
Fri Dec. 13
Unoriented point cloud reconstruction using differentiable fields
Neural fields, particularly neural signed distance functions (neural SDFs), have recently become a popular representation for reconstructing point clouds without input normals. While the smoothness of this representation gives many regularization benefits, it also causes a lot of problems for optimization, with the end reconstruction being highly biased by the initialization of the field. In this talk, I will discuss our line of work on improving optimization strategies for neural SDF optimization. I will also talk about our work on a newer representation for this task that is quickly becoming popular, the generalized winding number field. This representation alleviates many of the drawbacks of neural SDFs for this task, and our work shows that many techniques developed for neural field optimization are beneficial here.