Blog Home

Long Sequence Modeling with XGen: A 7B LLM Trained on 8K Input Sequence Length

TLDR We trained a series of 7B LLMs named XGen-7B with standard dense attention on up to 8K sequence length for up to 1.5T tokens. We also fine tune the models on public-domain instructional data. The main take-aways are: * On standard NLP benchmarks, XGen achieves comparable or better results

28 Jun 2023 • #llm
Page 1 of 1
Salesforce.com Careers In the Media Github Contact Privacy Information Terms of Service Legal Responsible Disclosure Trust Cookie Preferences Your Privacy Choices

© Copyright 2023 Salesforce.com, inc. All rights reserved. Rights of ALBERT EINSTEIN are use with permission of The Hebrew University of Jerusalem. Represented exclusively by Greenlight.