Coder Social home page Coder Social logo

Comments (4)

krsh-37 avatar krsh-37 commented on June 18, 2024 3

Hi Team,

Read through the paper as well, great work.

  1. I have a doubt, if have enough space in VRAM to load the model, will these optimizations helps?
  2. How much data % of train data is suggested for DejaVu 'Predictors' finding?
  3. How to obtain predictors from custom-trained models, should we again do inference using DejaVu or any other alternate method?

Thanks

from powerinfer.

YixinSong-e avatar YixinSong-e commented on June 18, 2024

Hi Team,

Read through the paper as well, great work.

  1. I have a doubt, if have enough space in VRAM to load the model, will these optimizations helps?
  2. How much data % of train data is suggested for DejaVu 'Predictors' finding?
  3. How to obtain predictors from custom-trained models, should we again do inference using DejaVu or any other alternate method?

Thanks

Hello, thank you for your interest.

  1. Yes, when we have enough space in VRAM, we will fall back to Deja Vu.But currently, our code has not been optimized for complete offloading, and we will support this feature.
  2. Actually I use 1M data point for predictor training.
  3. For training predictors, we will open source a tool. At present, you can refer to the implementation of predictor training in Dejavu.

from powerinfer.

YixinSong-e avatar YixinSong-e commented on June 18, 2024

I have a fine tuned vicuna 7B model, i tried to convert into PowerInfer with 'LLaMA(ReLU)-2-7B' predictor, but the inference is not right? Is this because of a different predictor used rather than that of fine-tuned model one? How to obtain these weights?

I Todo section i see 'Release core code of PowerInfer, supporting Llama-2, Falcon-40B.' is marked as done.

Can we use PowerInfer for fine-tuned vicuna/ llama models?

Thanks

Prerequisites

Before submitting your question, please ensure the following:

  • I am running the latest version of PowerInfer. Development is rapid, and as of now, there are no tagged versions.
  • I have carefully read and followed the instructions in the README.md.
  • I searched using keywords relevant to my issue to make sure that I am creating a new issue that is not already open (or closed).

Question Details

Please provide a clear and concise description of your question. If applicable, include steps to reproduce the issue or behaviors you've observed.

Additional Context

Please provide any additional information that may be relevant to your question, such as specific system configurations, environment details, or any other context that could be helpful in addressing your inquiry.

Thank you for your interest. First, for now, we just support ReLU-based model. And every model has its own predictor.
For now we do not support fine-tuned vicuna/ llama models because they are not ReLU-based models. By the way, we will release mistral-based model in the future. And we will SFT and DPO finetune this model.

from powerinfer.

jet-yangqs avatar jet-yangqs commented on June 18, 2024

Hi Team,
Read through the paper as well, great work.

  1. I have a doubt, if have enough space in VRAM to load the model, will these optimizations helps?
  2. How much data % of train data is suggested for DejaVu 'Predictors' finding?
  3. How to obtain predictors from custom-trained models, should we again do inference using DejaVu or any other alternate method?

Thanks

Hello, thank you for your interest.

  1. Yes, when we have enough space in VRAM, we will fall back to Deja Vu.But currently, our code has not been optimized for complete offloading, and we will support this feature.
  2. Actually I use 1M data point for predictor training.
  3. For training predictors, we will open source a tool. At present, you can refer to the implementation of predictor training in Dejavu.

Dear Team,

I hope you're doing well. I'm following up on the discussion about the optimization for complete offloading and the fallback to Deja Vu.

Could you kindly provide any updates on the progress of this feature?

Thank you for your time.

from powerinfer.

Related Issues (20)

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. 📊📈🎉

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google ❤️ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.