@inproceedings{dong2024spreadsheetllm, author = {Dong, Haoyu and Tian, Yuzhang and Zhao, Jianbo and Xiong, Junyu and Zhou, Mengyu and Lin, Yun and Cambronero, José and He, Yeye and Han, Shi and Zhang, Dongmei}, title = {SpreadsheetLLM: Encoding Spreadsheets for Large Language Models}, booktitle = {The 2024 Conference on Empirical Methods in Natural Language Processing (EMNLP '24)}, year = {2024}, month = {November}, abstract = {Spreadsheets are characterized by their extensive two-dimensional grids, flexible layouts, and varied formatting options, which pose significant challenges for large language models (LLMs). In response, we introduce SheetEncoder, pioneering an efficient encoding method designed to unleash and optimize LLMs’ powerful understanding and reasoning capability on spreadsheets. Initially, we propose a vanilla serialization approach that incorporates cell addresses, values, and formats. However, this approach was limited by LLMs' token constraints, making it impractical for most applications. To tackle this challenge, three innovative modules are proposed to compress spreadsheets effectively: structural-anchor-based compression, inverse index translation, and data-format-aware aggregation. It significantly improves performance in spreadsheet table detection task, outperforming the vanilla approach by 25.6\% in GPT4’s in-context learning setting. Moreover, fine-tuned LLM with SheetEncoder has an average compression ratio of 25×, but achieves a state-of-the-art 78.9\% F1 score, surpassing the best existing models by 12.3\%, demonstrating that SheetEncoder greatly boosts LLMs's performance on spreadsheet data.}, url = {http://approjects.co.za/?big=en-us/research/publication/encoding-spreadsheets-for-large-language-models/}, }