diff --git a/Dockerfile.hpu b/Dockerfile.hpu index a3b7b11..fdea2c0 100644 --- a/Dockerfile.hpu +++ b/Dockerfile.hpu @@ -42,5 +42,8 @@ COPY requirements.txt /root/whisper/requirements.txt # Install Python packages from requirements.txt RUN pip install --upgrade pip \ - && pip install optimum[habana]==1.13.1 \ - && pip install -r requirements.txt + && pip install optimum-habana==1.14.1 \ + && pip install transformers==4.33.0 \ + && pip install huggingface-hub==0.26.2 \ + && pip install tiktoken==0.8.0 \ + && pip install numba==0.60.0 diff --git a/whisper/__init__.py b/whisper/__init__.py index e66e78e..e8ed005 100644 --- a/whisper/__init__.py +++ b/whisper/__init__.py @@ -165,10 +165,14 @@ def load_model( if device == "hpu": from habana_frameworks.torch.utils.library_loader import load_habana_module - from habana_frameworks.torch.hpu import wrap_in_hpu_graph load_habana_module() if torch.hpu.is_available(): - return wrap_in_hpu_graph(model) + from habana_frameworks.torch.hpu import wrap_in_hpu_graph + model = model.eval().to(device) + + model = wrap_in_hpu_graph(model) + + return model return model.to(device)