UHG
Search
Close this search box.

6 Outstanding Papers Presented at NeurIPS

Out of the 13,321 papers submitted at NeurIPS 2023, the top of the lot have won the outstanding awards this year.

Share

Table of Content
Listen to this story

During the ongoing Neural Information Processing Systems (NeurIPS) annual conference, reviewers and chairpersons are currently evaluating tens of thousands of submissions.

Out of the 13,321 papers submitted by authors and researchers worldwide, the top of the lot have won the outstanding awards this year. Here are the 6 outstanding papers announced by NeurIPS in 2023: 

Outstanding Main Track Papers

Privacy Auditing with One (1) Training Run

Steinke, Nasr, and Jagielski propose an efficient auditing scheme for assessing the privacy of differentially private machine learning (ML) systems in a single training run. They leverage the parallelism of adding or removing multiple training examples independently. They avoid the computational cost of group privacy by analysing the connection between differential privacy and statistical generalisation.

 Their approach works in both black-box and white-box settings, requiring minimal assumptions about the algorithm. They demonstrate the effectiveness of their framework on DP-SGD, achieving meaningful privacy bounds with just one model, while standard methods would need hundreds of models.

Are Emergent Abilities of Large Language Models a Mirage?

Schaeffer, Miranda, and Koyejo challenge the idea that large language models (LLMs) exhibit true emergent abilities. They propose that perceived emergent abilities are often a result of the researcher’s metric choices rather than fundamental changes in model behaviour with scale. They support this with a mathematical model and three analyses:

  1. Confirming predictions on metric effects using InstructGPT/GPT-3
  2. Validating predictions in a meta-analysis on BIG-Bench
  3. Demonstrating how metric choices can create apparent emergent abilities in vision tasks across different networks.

Their findings suggest that alleged emergent abilities may vanish with different metrics, questioning the notion that they are intrinsic to scaled AI models.

Runner-Ups

Scaling Data-Constrained Language Models

In the paper, researchers explored scaling language models in data-limited scenarios, given the potential constraint on internet text data. They conducted extensive experiments, varying data repetition and computed budgets of up to 900 billion tokens and 9 billion parameters. Results showed that with limited data and a fixed computing budget, up to 4 epochs of repeated data had minimal impact on loss. However, further repetition diminished the value of additional compute. 

They proposed a scaling law for compute optimality, considering the declining value of repeated tokens and excess parameters. Additionally, they tested methods to alleviate data scarcity, such as augmenting with code data or removing common filters.

Models and datasets from 400 training runs are freely available on GitHub.

Direct Preference Optimization: Your Language Model is Secretly a Reward Model

Here, researchers introduced Direct Preference Optimization (DPO) as a streamlined alternative to Reinforcement Learning from Human Feedback (RLHF) for controlling large unsupervised language models. Unlike RLHF, DPO avoids the complexity and instability of fitting reward models and fine-tuning. Leveraging a mapping between reward functions and optimal policies, DPO directly optimises a single-stage policy training process, solving a classification problem on human preference data. 

The experiments demonstrate that DPO can effectively align language models with human preferences, outperforming RLHF in sentiment control and improving response quality in summarization and dialogue. Notably, DPO is more straightforward to implement and train.

Datasets and Benchmarks Papers:

ClimSim: A large multi-scale dataset for hybrid physics-ML climate emulation

Machine learning experts have introduced ClimSim, the largest hybrid ML-physics dataset, co-created by climate scientists and ML researchers. With 5.7 billion pairs of input-output vectors, it isolates the impact of high-resolution physics on macro-scale climate states. Global and spanning multiple years, the dataset facilitates emulators compatible with operational climate simulators.

The data and code are released openly to support the development of hybrid ML-physics and high-fidelity climate simulations. 

DecodingTrust: A Comprehensive Assessment of Trustworthiness in GPT Models

With the rise of GPT models, practitioners are considering using them for sensitive applications like healthcare and finance, but research reveals undisclosed vulnerabilities. GPT models, including GPT-4, can produce biassed, toxic outputs and unintentionally leak private information. 

Despite GPT-4’s generally improved trustworthiness, it exhibits vulnerability to jailbreaking systems or misleading user prompts. This study highlights previously unrecognised trustworthiness gaps in GPT models.

The benchmark is publicly available on GitHub.

📣 Want to advertise in AIM? Book here

Related Posts
19th - 23rd Aug 2024
Generative AI Crash Course for Non-Techies
Upcoming Large format Conference
Sep 25-27, 2024 | 📍 Bangalore, India
Download the easiest way to
stay informed

Subscribe to The Belamy: Our Weekly Newsletter

Biggest AI stories, delivered to your inbox every week.
Flagship Events
Rising 2024 | DE&I in Tech Summit
April 4 and 5, 2024 | 📍 Hilton Convention Center, Manyata Tech Park, Bangalore
Data Engineering Summit 2024
May 30 and 31, 2024 | 📍 Bangalore, India
MachineCon USA 2024
26 July 2024 | 583 Park Avenue, New York
MachineCon GCC Summit 2024
June 28 2024 | 📍Bangalore, India
Cypher USA 2024
Nov 21-22 2024 | 📍Santa Clara Convention Center, California, USA
Cypher India 2024
September 25-27, 2024 | 📍Bangalore, India
discord icon
AI Forum for India
Our Discord Community for AI Ecosystem, In collaboration with NVIDIA.