Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

GHA: Use 2025.0.0 RC1 #1578

Merged
merged 21 commits into from
Jan 20, 2025
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension


Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
12 changes: 7 additions & 5 deletions .github/workflows/causal_lm_cpp.yml
Original file line number Diff line number Diff line change
Expand Up @@ -16,10 +16,11 @@ concurrency:
cancel-in-progress: true

env:
l_ov_link: https://storage.openvinotoolkit.org/repositories/openvino/packages/nightly/2025.0.0-17800-91ae987c516/l_openvino_toolkit_ubuntu20_2025.0.0.dev20250109_x86_64.tgz
l_u22_ov_link: https://storage.openvinotoolkit.org/repositories/openvino/packages/nightly/2025.0.0-17800-91ae987c516/l_openvino_toolkit_ubuntu22_2025.0.0.dev20250109_x86_64.tgz
m_ov_link: https://storage.openvinotoolkit.org/repositories/openvino/packages/nightly/2025.0.0-17800-91ae987c516/m_openvino_toolkit_macos_12_6_2025.0.0.dev20250109_x86_64.tgz
w_ov_link: https://storage.openvinotoolkit.org/repositories/openvino/packages/nightly/2025.0.0-17800-91ae987c516/w_openvino_toolkit_windows_2025.0.0.dev20250109_x86_64.zip
l_ov_link: https://storage.openvinotoolkit.org/repositories/openvino/packages/pre-release/2025.0.0rc1/l_openvino_toolkit_ubuntu20_2025.0.0.dev20250116_x86_64.tgz
l_u22_ov_link: https://storage.openvinotoolkit.org/repositories/openvino/packages/pre-release/2025.0.0rc1/l_openvino_toolkit_ubuntu22_2025.0.0.dev20250116_x86_64.tgz
m_ov_link: https://storage.openvinotoolkit.org/repositories/openvino/packages/pre-release/2025.0.0rc1/m_openvino_toolkit_macos_12_6_2025.0.0.dev20250116_x86_64.tgz
w_ov_link: https://storage.openvinotoolkit.org/repositories/openvino/packages/pre-release/2025.0.0rc1/w_openvino_toolkit_windows_2025.0.0.dev20250116_x86_64.zip

jobs:
cpp-multinomial-greedy_causal_lm-ubuntu:
runs-on: ubuntu-20.04-8-cores
Expand Down Expand Up @@ -463,6 +464,7 @@ jobs:
env:
PYTHONPATH: "./build/:$PYTHONPATH"
LD_LIBRARY_PATH: "./build/openvino_genai/:$LD_LIBRARY_PATH"

cpp-prompt_lookup_decoding_lm-ubuntu:
runs-on: ubuntu-20.04-16-cores
defaults:
Expand Down Expand Up @@ -520,6 +522,7 @@ jobs:
env:
PYTHONPATH: "./build/:$PYTHONPATH"
LD_LIBRARY_PATH: "./build/openvino_genai/:$LD_LIBRARY_PATH"

cpp-Phi-1_5:
runs-on: ubuntu-20.04-16-cores
defaults:
Expand Down Expand Up @@ -697,7 +700,6 @@ jobs:
diff pred2.txt ref.txt
echo "Chat sample python" passed


visual_language_chat_sample-ubuntu-minicpm_v2_6:
runs-on: ubuntu-22.04-16-cores
steps:
Expand Down
4 changes: 2 additions & 2 deletions .github/workflows/genai-tools.yml
Original file line number Diff line number Diff line change
Expand Up @@ -39,12 +39,12 @@ jobs:
- ${{ github.workspace }}:${{ github.workspace }}

steps:
- uses: openvinotoolkit/openvino/.github/actions/openvino_provider@master
- uses: openvinotoolkit/openvino/.github/actions/openvino_provider@releases/2025/0
id: openvino_download
with:
platform: ubuntu22
commit_packages_to_provide: wheels
revision: ed470e7e40129d6b2bf728bc9527316937a69ef7
revision: latest_available_commit

llm_bench:
name: 'LLM bench tests'
Expand Down
2 changes: 1 addition & 1 deletion .github/workflows/job_vlm_sample_llava.yml
Original file line number Diff line number Diff line change
Expand Up @@ -11,7 +11,7 @@ on:
type: string

env:
l_u22_ov_link: https://storage.openvinotoolkit.org/repositories/openvino/packages/nightly/2025.0.0-17800-91ae987c516/l_openvino_toolkit_ubuntu22_2025.0.0.dev20250109_x86_64.tgz
l_u22_ov_link: https://storage.openvinotoolkit.org/repositories/openvino/packages/pre-release/2025.0.0rc1/l_openvino_toolkit_ubuntu22_2025.0.0.dev20250116_x86_64.tgz

jobs:
visual_language_chat_sample-ubuntu-llava:
Expand Down
6 changes: 3 additions & 3 deletions .github/workflows/lcm_dreamshaper_cpp.yml
Original file line number Diff line number Diff line change
Expand Up @@ -18,8 +18,8 @@ concurrency:

env:
PYTHON_VERSION: '3.9'
LINUX_OV_ARCHIVE_URL: https://storage.openvinotoolkit.org/repositories/openvino/packages/nightly/2025.0.0-17800-91ae987c516/l_openvino_toolkit_ubuntu22_2025.0.0.dev20250109_x86_64.tgz
WINDOWS_OV_ARCHIVE_URL: https://storage.openvinotoolkit.org/repositories/openvino/packages/nightly/2025.0.0-17800-91ae987c516/w_openvino_toolkit_windows_2025.0.0.dev20250109_x86_64.zip
LINUX_OV_ARCHIVE_URL: https://storage.openvinotoolkit.org/repositories/openvino/packages/pre-release/2025.0.0rc1/l_openvino_toolkit_ubuntu22_2025.0.0.dev20250116_x86_64.tgz
WINDOWS_OV_ARCHIVE_URL: https://storage.openvinotoolkit.org/repositories/openvino/packages/pre-release/2025.0.0rc1/w_openvino_toolkit_windows_2025.0.0.dev20250116_x86_64.zip
OV_INSTALL_DIR: ${{ github.workspace }}/ov

jobs:
Expand Down Expand Up @@ -93,7 +93,7 @@ jobs:
${{ env.build_dir }}/samples/cpp/image_generation/inpainting ./models/lcm_dreamshaper_v7 "cyberpunk cityscape like Tokyo New York with tall buildings at dusk golden hour cinematic lighting" ./image.png ./mask_image.png

lcm_dreamshaper_v7_cpp-windows:
runs-on: windows-2019
runs-on: windows-2022
defaults:
run:
shell: pwsh
Expand Down
4 changes: 2 additions & 2 deletions .github/workflows/linux.yml
Original file line number Diff line number Diff line change
Expand Up @@ -47,12 +47,12 @@ jobs:
- ${{ github.workspace }}:${{ github.workspace }}

steps:
- uses: openvinotoolkit/openvino/.github/actions/openvino_provider@master
- uses: openvinotoolkit/openvino/.github/actions/openvino_provider@releases/2025/0
id: openvino_download
with:
platform: ubuntu22
commit_packages_to_provide: wheels
revision: ed470e7e40129d6b2bf728bc9527316937a69ef7
revision: latest_available_commit

- name: Clone docker tag from OpenVINO repo
uses: actions/checkout@11bd71901bbe5b1630ceea73d27597364c9af683 # v4.2.2
Expand Down
20 changes: 10 additions & 10 deletions .github/workflows/mac.yml
Original file line number Diff line number Diff line change
Expand Up @@ -17,8 +17,8 @@ concurrency:

env:
PYTHON_VERSION: '3.10'
OV_BRANCH: 'master'
OV_TARBALL: ''
OV_BRANCH: 'releases/2025/0'
OV_TARBALL: 'https://storage.openvinotoolkit.org/repositories/openvino/packages/pre-release/2025.0.0rc1/m_openvino_toolkit_macos_12_6_2025.0.0.dev20250116_x86_64.tgz'

jobs:
openvino_download:
Expand Down Expand Up @@ -224,17 +224,17 @@ jobs:
- name: Test bindings
run: |
source ${OV_INSTALL_DIR}/setupvars.sh
python -m pip install ./thirdparty/openvino_tokenizers/[transformers] -r ./tests/python_tests/requirements.txt --find-links ${OV_INSTALL_DIR}/wheels
python -m pip install ./thirdparty/openvino_tokenizers/[transformers] -r ./tests/python_tests/requirements.txt --extra-index-url https://storage.openvinotoolkit.org/simple/wheels/pre-release
python -m pytest -v ./tests/python_tests/test_tokenizer.py::test_set_chat_template
env:
PYTHONPATH: "./build/:$PYTHONPATH"

- name: Test bindings (wheel)
run: |
source ${OV_INSTALL_DIR}/setupvars.sh
python -m pip install . --verbose --find-links ${OV_INSTALL_DIR}/wheels
python -m pip install . --verbose --extra-index-url https://storage.openvinotoolkit.org/simple/wheels/pre-release
python -c "from openvino_genai import LLMPipeline"
python -m pip install ./tools/who_what_benchmark --find-links ${OV_INSTALL_DIR}/wheels
python -m pip install ./tools/who_what_benchmark --extra-index-url https://storage.openvinotoolkit.org/simple/wheels/pre-release
python -m pytest -v ./tests/python_tests/ --ignore ./tests/python_tests/test_whisper_pipeline.py --ignore ./tests/python_tests/test_vlm_pipeline.py -k "not test_set_chat_template"

genai_python_lib_whisper:
Expand Down Expand Up @@ -289,17 +289,17 @@ jobs:
- name: Test bindings
run: |
source ${OV_INSTALL_DIR}/setupvars.sh
python -m pip install ./thirdparty/openvino_tokenizers/[transformers] -r ./tests/python_tests/requirements.txt --find-links ${OV_INSTALL_DIR}/wheels
python -m pip install ./thirdparty/openvino_tokenizers/[transformers] -r ./tests/python_tests/requirements.txt --extra-index-url https://storage.openvinotoolkit.org/simple/wheels/pre-release
python -m pytest -v ./tests/python_tests/test_whisper_pipeline.py -k test_smoke
env:
PYTHONPATH: "./build/:$PYTHONPATH"

- name: Test bindings (wheel)
run: |
source ${OV_INSTALL_DIR}/setupvars.sh
python -m pip install . --verbose --find-links ${OV_INSTALL_DIR}/wheels
python -m pip install . --verbose --extra-index-url https://storage.openvinotoolkit.org/simple/wheels/pre-release
python -c "from openvino_genai import LLMPipeline"
python -m pip install ./tools/who_what_benchmark --find-links ${OV_INSTALL_DIR}/wheels
python -m pip install ./tools/who_what_benchmark --extra-index-url https://storage.openvinotoolkit.org/simple/wheels/pre-release
python -m pytest -v ./tests/python_tests/test_whisper_pipeline.py -k "not test_smoke"

genai_package:
Expand Down Expand Up @@ -355,8 +355,8 @@ jobs:
- name: Build and Install dependencies
run: |
source ${OV_INSTALL_DIR}/setupvars.sh
python -m pip install ./thirdparty/openvino_tokenizers/[transformers] --find-links ${OV_INSTALL_DIR}/wheels
python -m pip install -r ./samples/requirements.txt --find-links ${OV_INSTALL_DIR}/wheels
python -m pip install ./thirdparty/openvino_tokenizers/[transformers] --extra-index-url https://storage.openvinotoolkit.org/simple/wheels/pre-release
python -m pip install -r ./samples/requirements.txt --extra-index-url https://storage.openvinotoolkit.org/simple/wheels/pre-release
optimum-cli export openvino --trust-remote-code --model TinyLlama/TinyLlama-1.1B-Chat-v1.0 TinyLlama-1.1B-Chat-v1.0
optimum-cli export openvino --trust-remote-code --model openai/whisper-tiny whisper-tiny

Expand Down
10 changes: 5 additions & 5 deletions .github/workflows/stable_diffusion_1_5_cpp.yml
Original file line number Diff line number Diff line change
Expand Up @@ -40,12 +40,12 @@ jobs:
- ${{ github.workspace }}:${{ github.workspace }}

steps:
- uses: openvinotoolkit/openvino/.github/actions/openvino_provider@master
- uses: openvinotoolkit/openvino/.github/actions/openvino_provider@releases/2025/0
id: openvino_download
with:
platform: ubuntu22
commit_packages_to_provide: wheels
revision: ed470e7e40129d6b2bf728bc9527316937a69ef7
revision: 2025.0.0rc1

openvino_download_windows:
name: Download OpenVINO for Windows
Expand All @@ -66,12 +66,12 @@ jobs:
- ${{ github.workspace }}:${{ github.workspace }}

steps:
- uses: openvinotoolkit/openvino/.github/actions/openvino_provider@master
- uses: openvinotoolkit/openvino/.github/actions/openvino_provider@releases/2025/0
id: openvino_download
with:
platform: windows
commit_packages_to_provide: wheels
revision: ed470e7e40129d6b2bf728bc9527316937a69ef7
revision: 2025.0.0rc1

stable_diffusion_1_5_cpp-linux:
runs-on: ubuntu-22.04-8-cores
Expand Down Expand Up @@ -153,7 +153,7 @@ jobs:

stable_diffusion_1_5_cpp-windows:
needs: [ openvino_download_windows ]
runs-on: windows-2019
runs-on: windows-2022
defaults:
run:
shell: pwsh
Expand Down
2 changes: 1 addition & 1 deletion .github/workflows/windows.yml
Original file line number Diff line number Diff line change
Expand Up @@ -17,7 +17,7 @@ concurrency:

env:
PYTHON_VERSION: '3.11'
OV_BRANCH: 'ed470e7e40129d6b2bf728bc9527316937a69ef7'
OV_BRANCH: 'releases/2025/0'
OV_TARBALL: ''

jobs:
Expand Down
2 changes: 1 addition & 1 deletion samples/cpp/image_generation/README.md
Original file line number Diff line number Diff line change
Expand Up @@ -28,7 +28,7 @@ Users can change the sample code and play with the following generation paramete

The `--upgrade-strategy eager` option is needed to ensure `optimum-intel` is upgraded to the latest version.

It's not required to install [../../export-requirements.txt](../../export requirements.txt) for deployment if the model has already been exported.
It's not required to install [../../export-requirements.txt](../../export-requirements.txt) for deployment if the model has already been exported.

```sh
pip install --upgrade-strategy eager -r ../../requirements.txt
Expand Down
2 changes: 1 addition & 1 deletion samples/cpp/visual_language_chat/README.md
Original file line number Diff line number Diff line change
Expand Up @@ -12,7 +12,7 @@ There are two sample files:

The `--upgrade-strategy eager` option is needed to ensure `optimum-intel` is upgraded to the latest version.

It's not required to install [../../export-requirements.txt](../../export requirements.txt) for deployment if the model has already been exported.
It's not required to install [../../export-requirements.txt](../../export-requirements.txt) for deployment if the model has already been exported.

```sh
pip install --upgrade-strategy eager -r ../../requirements.txt
Expand Down
2 changes: 1 addition & 1 deletion samples/cpp/whisper_speech_recognition/README.md
Original file line number Diff line number Diff line change
Expand Up @@ -6,7 +6,7 @@ This example showcases inference of speech recognition Whisper Models. The appli

The `--upgrade-strategy eager` option is needed to ensure `optimum-intel` is upgraded to the latest version.

It's not required to install [../../export-requirements.txt](../../export requirements.txt) for deployment if the model has already been exported.
It's not required to install [../../export-requirements.txt](../../export-requirements.txt) for deployment if the model has already been exported.

```sh
pip install --upgrade-strategy eager -r ../../requirements.txt
Expand Down
3 changes: 2 additions & 1 deletion src/cpp/src/llm_pipeline_static.cpp
Original file line number Diff line number Diff line change
Expand Up @@ -34,6 +34,7 @@ namespace {
namespace opp = ov::pass::pattern;
class TransposeValueTensors : public ov::pass::MatcherPass {
public:
OPENVINO_MATCHER_PASS_RTTI("TransposeValueTensors");
struct Context {
std::vector<std::shared_ptr<ov::opset13::Parameter>> new_params;
std::vector<std::shared_ptr<ov::opset13::Parameter>> old_params;
Expand Down Expand Up @@ -95,7 +96,7 @@ class TransposeValueTensors : public ov::pass::MatcherPass {

class ScaledDotProductAttentionDecomposition : public ov::pass::MatcherPass {
public:
OPENVINO_RTTI("ScaledDotProductAttentionDecomposition", "0");
OPENVINO_MATCHER_PASS_RTTI("ScaledDotProductAttentionDecomposition");
ScaledDotProductAttentionDecomposition() {
auto pattern_node = ov::pass::pattern::wrap_type<ov::op::v13::ScaledDotProductAttention>();

Expand Down
6 changes: 3 additions & 3 deletions src/cpp/src/lora_adapter.cpp
Original file line number Diff line number Diff line change
Expand Up @@ -473,7 +473,7 @@ struct LoRAWeightStateGetter {
class LoRATransformBase : public ov::pass::MatcherPass {
public:

OPENVINO_RTTI("LoRATransformBase");
OPENVINO_MATCHER_PASS_RTTI("LoRATransformBase");

LoRATransformBase(const LoRAWeightByNodeGetter& lora_weight_getter) {
register_matcher(
Expand Down Expand Up @@ -693,7 +693,7 @@ class LoRAFuseTransform : public LoRATransformBase {

public:

OPENVINO_RTTI("LoRAFuseTransform");
OPENVINO_RTTI("LoRAFuseTransform", "genai", LoRATransformBase);

LoRAFuseTransform(const LoRAWeightByNodeGetter& lora_weight_getter, const std::string& device_for_fusion = "CPU") :
LoRATransformBase(lora_weight_getter),
Expand Down Expand Up @@ -763,7 +763,7 @@ class LoRAFuseTransform : public LoRATransformBase {
class LoRASeparateTransform : public LoRATransformBase {
public:

OPENVINO_RTTI("LoRASeparateTransform");
OPENVINO_RTTI("LoRASeparateTransform", "genai", LoRATransformBase);

LoRASeparateTransform(const LoRAWeightByNodeGetter& lora_getter) : LoRATransformBase(lora_getter) {}

Expand Down
4 changes: 2 additions & 2 deletions src/cpp/src/make_tokenizer_stateful.hpp
Original file line number Diff line number Diff line change
Expand Up @@ -34,7 +34,7 @@ namespace genai {
**/
class MakeCombineSegmentsSatateful : public ov::pass::ModelPass {
public:
OPENVINO_RTTI("MakeCombineSegmentsSatateful", "0");
OPENVINO_MODEL_PASS_RTTI("MakeCombineSegmentsSatateful");
bool run_on_model(const std::shared_ptr<ov::Model>& model) override;
};

Expand Down Expand Up @@ -70,7 +70,7 @@ class MakeCombineSegmentsSatateful : public ov::pass::ModelPass {
**/
class MakeVocabDecoderSatateful : public ov::pass::ModelPass {
public:
OPENVINO_RTTI("MakeVocabDecoderSatateful", "0");
OPENVINO_MODEL_PASS_RTTI("MakeVocabDecoderSatateful");
bool run_on_model(const std::shared_ptr<ov::Model>& model) override;
};

Expand Down
2 changes: 1 addition & 1 deletion src/cpp/src/whisper_pipeline_static.cpp
Original file line number Diff line number Diff line change
Expand Up @@ -347,7 +347,7 @@ void add_attention_mask_input(std::shared_ptr<ov::Model> model) {
using namespace ov::op;
class AttentionMaskInput : public ov::pass::MatcherPass {
public:
OPENVINO_RTTI("AttentionMaskInput");
OPENVINO_MATCHER_PASS_RTTI("AttentionMaskInput");

AttentionMaskInput(std::shared_ptr<ov::Model> model) {
auto range = wrap_type<v4::Range>();
Expand Down
4 changes: 3 additions & 1 deletion src/docs/BUILD.md
Original file line number Diff line number Diff line change
Expand Up @@ -191,9 +191,11 @@ The path to the openvino install directory is referred as <INSTALL_DIR> througho
```
4. Build the wheel in the `dist` directory:
```sh
python -m pip wheel . -w dist/ --extra-index-url https://storage.openvinotoolkit.org/simple/wheels/pre-release
python -m pip wheel . -w dist/ --extra-index-url https://storage.openvinotoolkit.org/simple/wheels/pre-release --extra-index-url https://storage.openvinotoolkit.org/simple/wheels/nightly
```

> **NOTE**: You'd need to build ABI compatible OpenVINO and OpenVINO Tokenizers for Ubuntu instead of downloading them from PyPI. See [OpenVINO™ GenAI Dependencies](../README.md#openvino-genai-dependencies) for the explanation.

### Install OpenVINO GenAI From Source

1. Clone OpenVINO GenAI repository and init submodules:
Expand Down
2 changes: 1 addition & 1 deletion tests/python_tests/requirements.txt
Original file line number Diff line number Diff line change
@@ -1,6 +1,6 @@
--extra-index-url https://download.pytorch.org/whl/cpu
diffusers==0.32.2
optimum-intel @ git+https://github.com/eaidova/optimum-intel@ea/stateful_seq2seq
optimum-intel @ git+https://github.com/huggingface/optimum-intel@main
numpy<2.0.0; platform_system == "Darwin" and platform_machine == "x86_64"
onnx==1.17.0
pytest
Expand Down
2 changes: 1 addition & 1 deletion tests/python_tests/test_continuous_batching.py
Original file line number Diff line number Diff line change
Expand Up @@ -117,7 +117,7 @@ def test_cb_streamer_vs_return_vs_stateful(prompt):
@pytest.mark.parametrize("model_descr", get_chat_models_list())
@pytest.mark.precommit
def test_chat_scenario_vs_stateful(model_descr, generation_config_kwargs: Dict):
model_id, models_path, hf_tokenizer, opt_model, ov_pipe = read_model((model_descr[0], model_descr[1] / '_test_chat'))
model_id, models_path, hf_tokenizer, opt_model, ov_pipe = read_model((model_descr[0], model_descr[1]))
cb_pipe = get_continuous_batching(models_path)

ov_pipe.start_chat()
Expand Down
2 changes: 1 addition & 1 deletion tests/python_tests/test_llm_pipeline.py
Original file line number Diff line number Diff line change
Expand Up @@ -129,7 +129,7 @@ def test_chat_scenario(model_descr, generation_config_kwargs: Dict):
chat_history_hf = []
chat_history_ov = []

model_id, path, tokenizer, opt_model, ov_pipe = read_model((model_descr[0], model_descr[1] / '_test_chat'))
model_id, path, tokenizer, opt_model, ov_pipe = read_model((model_descr[0], model_descr[1]))

ov_generation_config = GenerationConfig(**generation_config_kwargs)
hf_generation_config = convert_to_hf(opt_model.generation_config, ov_generation_config)
Expand Down
5 changes: 4 additions & 1 deletion tests/python_tests/test_sampling.py
Original file line number Diff line number Diff line change
Expand Up @@ -65,12 +65,15 @@ def test_stop_strings(tmp_path, generation_config):
'What is OpenVINO?',
'table is made of',
'The Sun is yellow because',
'你好! 你好嗎?'
'你好! 你好嗎?',
'I have an interview about product speccing with the company Weekend Health. Give me an example of a question they might ask with regards about a new feature'
])
@pytest.mark.parametrize("use_cb", [True, False])
def test_greedy(tmp_path, generation_config, prompt, use_cb):
model_id : str = "katuni4ka/tiny-random-phi3"
if sys.platform.startswith('win') and prompt.startswith('你'):
pytest.skip("For unknown reason this prompt fails on Win")

run_llm_pipeline_with_ref(model_id=model_id,
prompts=[prompt],
generation_config=generation_config,
Expand Down
Loading
Loading