Skip to content

EAGLE support and examples #10925

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Closed
wants to merge 3 commits into from
Closed

Conversation

jeanyu-habana
Copy link
Contributor

Description

ipex-llm and EAGLE integration plus EAGLE example scripts

1. Why the change?

EAGLE provides significant speedup in addition to ipex-llm optimizations. Please see below:
Llama 7B, temporature=1.0, Intel CPU
speed 27.445331381249126 TPS (optimized with both EAGLE and ipex-llm)
speed 20.132597255230788 TPS (optimized with EAGLE only)
speed 14.549053180428723 TPS (optimized with ipex-llm only)
speed_base 10.275284471199816 TPS (Baseline: not optimized)

Llama 7B, temporature=1.0, Intel GPU
speed 60.68802901159256 TPS eagle + ipex-llm (ratio: 3.74)
speed 41.41260508527679 TPS ipex-llm only (ratio: 2.55)
speed 31.480931699222744 TPS eagle only (ratio: 1.94)
speed_base 16.220403337894584 TPS (Baseline: not optimized)

2. User API changes

N/A

3. Summary of the change

Integrate with EAGLE (https://github.com/SafeAILab/EAGLE) and provide examples

4. How to test?

  • N/A
  • Unit test
  • [ X] Application test
  • Document test
  • ...
    Please follow the setup instructions and example commands in the README.

@jenniew
Copy link
Contributor

jenniew commented May 9, 2024

I suggest you to put examples under ipex-llm/python/llm/example/CPU/PyTorch-Models and ipex-llm/python/llm/example/GPU/PyTorch-Models

@jenniew
Copy link
Contributor

jenniew commented May 14, 2024

@jason-dai For the Eagle example placement, which directory is best to put to? like ipex-llm/python/llm/example/CPU/PyTorch-Models/EAGLE or ipex-llm/python/llm/example/CPU/Speculative-Decoding/EAGLE? If we put eagle example to ipex-llm/python/llm/example/CPU/Speculative-Decoding, we may need to create a new directory, like ipex-llm/python/llm/example/CPU/Speculative-Decoding/ipex-llm, and move our current speculative examples to that directory.

@jason-dai
Copy link
Contributor

@jason-dai For the Eagle example placement, which directory is best to put to? like ipex-llm/python/llm/example/CPU/PyTorch-Models/EAGLE or ipex-llm/python/llm/example/CPU/Speculative-Decoding/EAGLE? If we put eagle example to ipex-llm/python/llm/example/CPU/Speculative-Decoding, we may need to create a new directory, like ipex-llm/python/llm/example/CPU/Speculative-Decoding/ipex-llm, and move our current speculative examples to that directory.

Maybe example/CPU/Speculative-Decoding with Self-Speculation and Eagle folders

@jeanyu-habana
Copy link
Contributor Author

After discussing with @jenniew , we agreed to use a new PR for a cleaner merge : #11104. For this reason, I am closing this PR

@jeanyu-habana jeanyu-habana deleted the eagle branch May 24, 2024 19:12
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

3 participants