Blog

The Latest and Greatest from Salesforce Research

Meet LAVIS: A One-stop Library for Language-Vision AI Research and Applications

TL;DR: LAVIS (short for LAnguage-VISion) is an open-source deep learning library for language-vision research and applications, offering comprehensive support for a wide range of tasks, datasets, and state-of-the-art models. Featuring a unified interface and modular design, it’s easy to use off-the-shelf and to extend with new capabilities. With

20 Sep 2022 • Dongxu Li #LAVIS

ETSformer: Exponential Smoothing Transformers for Time-Series Forecasting

Authors: Gerald Woo, Chenghao Liu, Donald Rose TL;DR: We developed a new time-series forecasting model called ETSformer that leverages the power of two frameworks. By combining the classical intuition of seasonal-trend decomposition and exponential smoothing with modern transformers – as well as introducing novel exponential smoothing and frequency attention mechanisms

23 Aug 2022 • Gerald Woo #ETSformer

AI for Global Climate Cooperation: Salesforce Research and Mila Announce Climate Change Collaboration and Competition

TL;DR:  Salesforce Research and Mila announce AI for Global Climate Cooperation, a working group collaboration and competition to design negotiation protocols and climate agreements. We plan to coauthor a peer-reviewed scientific paper with top-performing teams; insights will be distilled into a policy brief shared with leading policymakers, informing future

05 Aug 2022 • Stephan Zheng #AI for Global Climate Cooperation

Salesforce Research at ICML 2022

Conference Overview This weekend will kick off the thirty-ninth International Conference on Machine Learning (ICML). This conference specifically aims to bring together professionals who are dedicated to the advancement of Machine Learning (ML) in Artificial Intelligence. Participants at ICML come from many different backgrounds, including academic and industrial researchers, entrepreneurs

17 Jul 2022 • Mia Ferrer

Salesforce Research at NAACL 2022

Conference Overview This weekend marks the start of the Annual Conference of the North American Chapter of the Association for Computational Linguistics (NAACL). NAACL provides a regional focus for members of the Association for Computational Linguistics (ACL) in North America. NAACL organizes annual conferences, promotes cooperation and information exchange among

10 Jul 2022 • Mia Ferrer

Salesforce Research at CVPR 2022

Conference Overview The IEEE / CVF Computer Vision and Pattern Recognition Conference (CVPR) is the annual conference on Computer Vision. CVPR is composed of both the main conference, as well as workshops and other courses, to provide a unique learning experience and networking opportunities in the field of Computer Vision. CVPR

20 Jun 2022 • Mia Ferrer #cvpr2022

TaiChi: Open Source Library for Few-Shot NLP

Authors: Sharvin Shah, Jin Qu, Donald Rose TL;DR - TaiChi is an open source library for few-shot NLP, designed for data scientists and software engineers who want to get some quick results or build proof-of-concept products but don’t have much experience with few-shot learning (FSL). The library abstracts

15 Jun 2022 • Jin Qu #NLP

Turbocharge Multi-Agent Reinforcement Learning with WarpDrive and PyTorch Lightning

TL;DR: WarpDrive is a flexible, lightweight, easy-to-use end-to-end reinforcement learning (RL) framework; enables orders-of-magnitude faster training on a single GPU. PyTorch Lightning enables you to modularize experimental code, and build production-ready workloads fast. Together, they can help significantly accelerate multi-agent RL R&D. Reinforcement Learning: Agents Learn by

20 May 2022 • Sunil Srinivasa

Salesforce Research at ACL 2022

Conference Overview This year marks the 60th annual meeting of the Association for Computational Linguistics Conference (ACL). ACL is the premier international scientific and professional society for people working on computational problems involving human language, a field often referred to as either computational linguistics or natural language processing (NLP). This

19 May 2022 • Mia Ferrer #NLP

Science Advances Publishes AI Economist Research on Improving Tax Policies With Reinforcement Learning

TL;DR: The AI Economist, a reinforcement learning (RL) system, learns dynamic tax policies that optimize equality along with productivity in simulated economies, outperforming alternative tax systems. We have now expanded this research, which is being published in the interdisciplinary scientific journal Science Advances. Humans or AI: Which Can Design

05 May 2022 • Stephan Zheng #AI Economist

Salesforce Research at ICLR 2022

Conference Overview This year marks the Tenth International Conference on Learning Representations (ICLR), one of the premier academic conferences dedicated to advancing research in representation learning - a type of machine learning also referred to as feature learning or deep learning. ICLR features the latest advancements in cutting-edge deep learning

25 Apr 2022 • Mia Ferrer #ICLR

Embracing Ethical AI at NeurIPS 2021

December 21, 2021 The leading AI research conference, NeurIPS 2021, has recently wrapped up, spanning seven very full days, 2,344 accepted papers, eight invited talks, ten tutorials, and nearly 70 workshops. Though there was diverse and innovative thought leadership on display, I found myself drawn to the particular topics

21 Dec 2021 • Anna Bethke #AI-fairness

CodeT5: The Code-aware Encoder-Decoder based Pre-trained Programming Language Models

TL; DR: Introducing CodeT5 --- the first code-aware, encoder-decoder-based pre-trained programming language model, which enables a wide range of code intelligence applications including code understanding and generation tasks. CodeT5 achieves state-of-the-art performance on 14 sub-tasks in the CodeXGLUE code intelligence benchmark. CodeT5 for code-related understanding and generation tasksGiven the goal

03 Sep 2021 • Yue Wang #code-intelligence

Learning without Labels

With data rapidly being generated by millions of people, it's not feasible to label all of it. Learn about the recent advancements in ML for how to train vision models with unlabelled data using self-supervised learning.

21 Jun 2021 • Michael Sollami #deeplearning

Salesforce Research at ICLR 2021

This year marks the 9th annual conference on International Conference on Learning Representations (ICLR) taking place in a fully virtual format from May 4th through May 8th, 2021. ICLR is a premier academic conference in the field of representation learning, generally referred to as deep learning or feature learning. ICLR

26 Apr 2021 • Mia Ferrer #ICLR

When are Neural Networks more powerful than Neural Tangent Kernels?

The empirical success of deep learning has posed significant challenges to machine learning theory: Why can we efficiently train neural networks with gradient descent despite its highly non-convex optimization landscape? Why do over-parametrized networks generalize well? The recently proposed Neural Tangent Kernel (NTK) theory offers a powerful framework for understanding

29 Mar 2021 • Yu Bai #deep learning theory

Applying AI Ethics Research in Practice

February 2020 Summary from FAccT 2020 CRAFT SessionAI Ethics practitioners in industry look to researchers for insights on how to best identify and mitigate harmful bias in their organization’s AI solutions and create more fair or equitable outcomes. However, it can be a challenge to apply those research insights

03 Mar 2021 • Kathy Baxter #ethics

CASTing Your Model: Learning to Localize Improves Self-Supervised Representations

TL; DR: We find that current self-supervised learning approaches suffer from poor visual grounding and receive improper supervisory signal when trained on complex scene images. We introduce CAST to improve visual grounding during pretraining and show that it yields significantly better transferable features. Self-supervised learning and its grounding problemSelf-Supervised Learning

09 Dec 2020 • Ramprasaath R. Selvaraju

Salesforce Research at NeurIPS 2020

This year marks the 34th annual conference on Neural Information Processing Systems (NeurIPS) reimagined for the first time ever in a fully virtual format. NeurIPS is a leading conference in the area of machine learning and neural information processing systems in their biological, technological, mathematical, and theoretical aspects. Neural information

30 Nov 2020 • Denna Mafie

CoMatch: Advancing Semi-supervised Learning with Contrastive Graph Regularization

TL; DR: We propose a new semi-supervised learning method which achieves state-of-the-art performance by learning jointly-evolved class probabilities and image representations.What are the existing semi-supervised learning methods?Semi-supervised learning aims to leverage few labeled data and a large amount of unlabeled data. As a long-standing and widely-studied topic in

23 Nov 2020 • Junnan Li

Salesforce Research at EMNLP 2020

This year marks the 24th annual Empirical Methods in Natural Language Processing (EMNLP) conference reimagined for the first time ever in a fully virtual format. EMNLP is a leading conference in the area of Natural Language Processing covering a broad spectrum of diverse research areas that are concerned with computational

11 Nov 2020 • Denna Mafie #research

A Language Detector for Identifying Machine-Generated Text

In recent years, the natural language processing (NLP) community has seen the development of increasingly powerful language models [1, 2], capable of generating textual output that is indistinguishable from human-written text. This includes our own model called CTRL [3] (Conditional Transformer Language Model) for controllable generation. To prevent misuse or

22 Oct 2020 • Yoav Schlesinger

The First Simulation Card for Ethical AI Simulations

We recently released Foundation, an open-source framework to build economic simulations. Foundation has been designed with flexibility and AI research in mind, and can be modified by anyone. AI simulations offer researchers the power to generate data and evaluate outcomes of virtual economies that capture a part of the real

20 Oct 2020 • Stephan Zheng

Model Cards for AI Model Transparency

At Salesforce, we take seriously our mission to create and deliver AI technology that is responsible, accountable, transparent, empowering, and inclusive. These principles ensure that our AI is safe, ethical, and engenders trust.

29 Sep 2020 • Yoav Schlesinger #ethics

Theory-Inspired Network Architecture Search

TL;DR: We theoretically analyze the differential architecture search (DARTS) for understanding the role and impact of skip connections, which inspires a new method for Neural Architecture Search (NAS) using group-structured sparse gates and path-depth-wise regularization to overcome the limitation of existing NAS methods for AutoML. In our work [1]

25 Sep 2020 • Pan Zhou