Next week, the Thirty-sixth annual Conference on Neural Information Processing Systems (NeurIPS) will be held in New Orleans, Louisiana from Monday, November 28th, through Friday, December 9th. NeurIPS will include invited talks, demonstrations, oral and poster presentations of accepted papers. Along with the conference is a professional exposition focusing on machine learning in practice, a series of tutorials, and topical workshops that provide a less formal setting for the exchange of ideas (NeurIPS Website). NeurIPS 2022 will begin with an in-person component at the New Orleans Convention Center (Nov. 28 - Dec 3rd), and a virtual component the second week (Dec 5th - 9th).
Conference Sponsorship: Salesforce AI Research is proud to support NeurIPS 2022 as a Diamond level sponsor. Our team of researchers and recruiters will be showcasing demos, discussing career opportunities and chatting with attendees at our booth (#503) all week. Our booth hours are as follows:
Monday, November 28 | 9:00am - 5:00pm, Welcome Reception 6:00pm - 8:00pm
Tuesday, November 29 | 9:00am - 5:00pm
Wednesday, November 30 | 9:00am - 5:00pm
Thursday, December 1 | 9:00am - 1:30pm
LatinX in AI: We’re excited to continue our partnership with the LatinX in AI Community. We will be participating in the LatinX in AI (LXAI) Workshop at NeurIPS on Monday, November 28th.
Networking Event: Salesforce AI Research will host an invite-only Networking and Trivia event on Wednesday, November 30th from 5:00pm - 8:00pm at the Bower Bar in New Orleans.
Salesforce Research is pleased to announce a total of 7 accepted Oral and Poster papers from our team of leading researchers.
Our accepted authors will present their work at NeurIPS throughout the main conference; with specific times, dates, and locations indicated below. We look forward to sharing some of our exciting new research with you!
Salesforce Researchers are shown in bold in the publication descriptions below.
Accepted Oral Paper
Yu Bai, Chi Jin, Song Mei, Ziang Song, Tiancheng Yu
We design the first line of algorithms for minimizing the trigger regret and learning correlated equilibria in Extensive-Form Games against adversarial opponents under bandit feedback. Our algorithms arise from connections to normal-form games, but are efficiently implementable and achieve sharper rates compared with naive such ones.
Thu 1 Dec 4:30 p.m. CST — 6 p.m. CST Hall J #538
Accepted Poster Papers
Hung Le, Wang Yue, Akhilesh Gotmare, Silvio Savarese, Steven Hoi
CodeRL is a groundbreaking new framework for program synthesis through holistic integration of pretrained language models and deep reinforcement learning. By utilizing unit test feedback as part of model training and inference, and integrating with an improved CodeT5 model, CodeRL achieves SoTA results on competition-level programming tasks.
Tue 29 Nov 11:30 a.m. CST — 1 p.m. CST Hall J #138
Devansh Arpit, Huan Wang, Yingbo Zhou, Caiming Xiong
A simple hyper-parameter free strategy of using the simple moving average of model parameters during training and ensemble achieves SOTA on domain generalization benchmarks, and can be explained using the Bias-Variance trade-off.
Tue 29 Nov 4:30 p.m. CST — 6 p.m. CST Hall J #732
Eshaan Nichani, Yu Bai, Jason D. Lee
We show that neural networks trained with gradient descent can probably escape the Neural Tangent Kernel (NTK) regime and achieve better sample efficiency than the NTK for learning certain natural functions, by leveraging properties of the NTK spectrum and newly designed regularizers.
Thu 1 Dec 4:30 p.m. CST — 6 p.m. CST Hall J #921
Runyu Zhang, Qinghua Liu, Huan Wang, Caiming Xiong, Na Li, Yu Bai
We show that a natural optimistic policy optimization algorithm achieves the current best convergence rate for finding equilibria in Markov Games, in both two-player zero-sum and multi-player general-sum settings. We also provide a framework that unifies most existing policy optimization algorithms and their analyses.
Wed 30 Nov 11:30 a.m. CST — 1 p.m. CST Hall J #816
Xuan-Phi Nguyen, Shafiq Joty, Wu Kui, Ai Ti Aw
A four-stage refinement procedure that finetunes a multilingual unsupervised NMT model to significantly outperform the baseline and achieve state of the art in low-resource unsupervised translation tasks.
Wed 30 Nov 4:30 p.m. CST — 6 p.m. CST Hall J #626
Ziang Song, Song Mei, Yu Bai
We design new algorithms for learning correlated equilibria in Extensive-Form Games, achieving the current best convergence rate under full-information feedback and the first sample-efficient learning result under bandit feedback.
Tue 29 Nov 11:30 a.m. CST — 1 p.m. CST Hall J #824
California - Palo Alto, Washington - Seattle (Tableau), Singapore (Linked here)
As a research intern, you will work with a team of research scientists and engineers on a project that ideally leads to a submission to a top-tier conference.
California - Palo Alto
A 12-month research training program intended to kickstart or further one's experience in AI research. Residents will gain valuable hands-on experience in fundamental and applied AI research, working closely with Salesforce researchers. Will kick off in August 2023.
To learn more about these and other research projects, please visit our website at salesforceairesearch.com.