Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[Question]: LLMLingua-2 Sample-Wise Dynamic Compression Ratio #174

Open
cornzz opened this issue Aug 2, 2024 · 2 comments
Open

[Question]: LLMLingua-2 Sample-Wise Dynamic Compression Ratio #174

cornzz opened this issue Aug 2, 2024 · 2 comments
Assignees
Labels
question Further information is requested

Comments

@cornzz
Copy link

cornzz commented Aug 2, 2024

Describe the issue

Hi,

I have two questions:

  1. Appendix L of the LLMLingua-2 paper talks about allowing the compressor to adjust the compression rate for different samples, but I cannot find any documentation about this in the git repo and looking at compress_prompt_llmlingua2() it seems like it is not possible? Also, I dont quite understand from the explanation in appendix L, how this dynamic compression is supposed to work, where can I find more details?

  2. What is the use_context_level_filter parameter for?

@cornzz cornzz added the question Further information is requested label Aug 2, 2024
@iofu728 iofu728 self-assigned this Aug 22, 2024
@iofu728
Copy link
Contributor

iofu728 commented Aug 22, 2024

Hi @cornzz, thanks for your interest in LLMLingua.

  1. First, you can find detailed documentation at this link.
  2. In Appendix L, the DCR is actually determined by using the compressor predictor's output as an indicator to allocate the compression ratio. However, this feature hasn't been added to the library yet. [ToDo] @pzs19
  3. The "use_context_level_filter" controls whether to apply coarse-level prompt compression.

@cornzz
Copy link
Author

cornzz commented Aug 22, 2024

@iofu728 thanks a lot for your response!

Regarding 3: is this what is referred to in the last paragraph of section 4.2 in the paper?

our approach can be readily integrated into the coarse-to-fine framework proposed in LLMLingua (Jiang et al., 2023a), allowing for a higher compression ratio of ∼15x for tasks involving multiple demonstrations or documents.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
question Further information is requested
Projects
None yet
Development

No branches or pull requests

3 participants