Skip to content

Commit

Permalink
Merge branch 'master' into 2.4
Browse files Browse the repository at this point in the history
  • Loading branch information
sirutBuasai authored Dec 18, 2024
2 parents ef16230 + 4d41082 commit cf51866
Show file tree
Hide file tree
Showing 7 changed files with 185 additions and 19 deletions.
Original file line number Diff line number Diff line change
Expand Up @@ -354,5 +354,36 @@
"title": "CVE-2024-11394 - transformers",
"reason_to_ignore": "N/A"
}
],
"lightgbm": [
{
"description": "LightGBM Remote Code Execution Vulnerability",
"vulnerability_id": "CVE-2024-43598",
"name": "CVE-2024-43598",
"package_name": "lightgbm",
"package_details": {
"file_path": "opt/conda/lib/python3.11/site-packages/lightgbm-4.5.0.dist-info/METADATA",
"name": "lightgbm",
"package_manager": "PYTHONPKG",
"version": "4.5.0",
"release": null
},
"remediation": {
"recommendation": {
"text": "None Provided"
}
},
"cvss_v3_score": 8.1,
"cvss_v30_score": 0,
"cvss_v31_score": 8.1,
"cvss_v2_score": 0,
"cvss_v3_severity": "HIGH",
"source_url": "https://nvd.nist.gov/vuln/detail/CVE-2024-43598",
"source": "NVD",
"severity": "HIGH",
"status": "ACTIVE",
"title": "CVE-2024-43598 - lightgbm",
"reason_to_ignore": "No fix provided"
}
]
}
Original file line number Diff line number Diff line change
Expand Up @@ -354,5 +354,36 @@
"title": "CVE-2024-11394 - transformers",
"reason_to_ignore": "N/A"
}
],
"lightgbm": [
{
"description": "LightGBM Remote Code Execution Vulnerability",
"vulnerability_id": "CVE-2024-43598",
"name": "CVE-2024-43598",
"package_name": "lightgbm",
"package_details": {
"file_path": "opt/conda/lib/python3.11/site-packages/lightgbm-4.5.0.dist-info/METADATA",
"name": "lightgbm",
"package_manager": "PYTHONPKG",
"version": "4.5.0",
"release": null
},
"remediation": {
"recommendation": {
"text": "None Provided"
}
},
"cvss_v3_score": 8.1,
"cvss_v30_score": 0,
"cvss_v31_score": 8.1,
"cvss_v2_score": 0,
"cvss_v3_severity": "HIGH",
"source_url": "https://nvd.nist.gov/vuln/detail/CVE-2024-43598",
"source": "NVD",
"severity": "HIGH",
"status": "ACTIVE",
"title": "CVE-2024-43598 - lightgbm",
"reason_to_ignore": "No fix provided"
}
]
}
Original file line number Diff line number Diff line change
Expand Up @@ -354,5 +354,36 @@
"title": "CVE-2024-11394 - transformers",
"reason_to_ignore": "N/A"
}
],
"lightgbm": [
{
"description": "LightGBM Remote Code Execution Vulnerability",
"vulnerability_id": "CVE-2024-43598",
"name": "CVE-2024-43598",
"package_name": "lightgbm",
"package_details": {
"file_path": "opt/conda/lib/python3.11/site-packages/lightgbm-4.5.0.dist-info/METADATA",
"name": "lightgbm",
"package_manager": "PYTHONPKG",
"version": "4.5.0",
"release": null
},
"remediation": {
"recommendation": {
"text": "None Provided"
}
},
"cvss_v3_score": 8.1,
"cvss_v30_score": 0,
"cvss_v31_score": 8.1,
"cvss_v2_score": 0,
"cvss_v3_severity": "HIGH",
"source_url": "https://nvd.nist.gov/vuln/detail/CVE-2024-43598",
"source": "NVD",
"severity": "HIGH",
"status": "ACTIVE",
"title": "CVE-2024-43598 - lightgbm",
"reason_to_ignore": "No fix provided"
}
]
}
Original file line number Diff line number Diff line change
Expand Up @@ -354,5 +354,36 @@
"title": "CVE-2024-11394 - transformers",
"reason_to_ignore": "N/A"
}
],
"lightgbm": [
{
"description": "LightGBM Remote Code Execution Vulnerability",
"vulnerability_id": "CVE-2024-43598",
"name": "CVE-2024-43598",
"package_name": "lightgbm",
"package_details": {
"file_path": "opt/conda/lib/python3.11/site-packages/lightgbm-4.5.0.dist-info/METADATA",
"name": "lightgbm",
"package_manager": "PYTHONPKG",
"version": "4.5.0",
"release": null
},
"remediation": {
"recommendation": {
"text": "None Provided"
}
},
"cvss_v3_score": 8.1,
"cvss_v30_score": 0,
"cvss_v31_score": 8.1,
"cvss_v2_score": 0,
"cvss_v3_severity": "HIGH",
"source_url": "https://nvd.nist.gov/vuln/detail/CVE-2024-43598",
"source": "NVD",
"severity": "HIGH",
"status": "ACTIVE",
"title": "CVE-2024-43598 - lightgbm",
"reason_to_ignore": "No fix provided"
}
]
}
2 changes: 1 addition & 1 deletion available_images.md
Original file line number Diff line number Diff line change
Expand Up @@ -193,7 +193,7 @@ Starting LMI V10 (0.28.0), we are changing the name from LMI DeepSpeed DLC to LM
|-----------------------------------------------------------------------------------------------------------------------------|-----------|-------------|------------------------|-------------------------------------------------------------------------------------------|
| DJLServing 0.31.0 with LMI Dist 13.0.0, vLLM 0.6.3.post1, HuggingFace Transformers 4.45.2, and HuggingFace Accelerate 1.0.1 | inference | GPU | 3.11 (py311) | 763104351884.dkr.ecr.us-west-2.amazonaws.com/djl-inference:0.31.0-lmi13.0.0-cu124 |
| DJLServing 0.30.0 with LMI Dist 12.0.0, vLLM 0.6.2, HuggingFace Transformers 4.45.2, and HuggingFace Accelerate 1.0.1 | inference | GPU | 3.10 (py310) | 763104351884.dkr.ecr.us-west-2.amazonaws.com/djl-inference:0.30.0-lmi12.0.0-cu124 |
| DJLServing 0.30.0 with TensorRT-LLM 0.12.0, HuggingFace Transformers 4.44.2, and HuggingFace Accelerate 0.33.0 | inference | GPU | 3.10 (py310) | 763104351884.dkr.ecr.us-west-2.amazonaws.com/djl-inference:0.30.0-tensorrtllm0.12.0 |
| DJLServing 0.30.0 with TensorRT-LLM 0.12.0, HuggingFace Transformers 4.44.2, and HuggingFace Accelerate 0.33.0 | inference | GPU | 3.10 (py310) | 763104351884.dkr.ecr.us-west-2.amazonaws.com/djl-inference:0.30.0-tensorrtllm0.12.0-cu125 |
| DJLServing 0.30.0 with Neuron SDK 2.20.1, TransformersNeuronX 0.12.313, and HuggingFace Transformers 4.45.2 | inference | Neuron | 3.10 (py310) | 763104351884.dkr.ecr.us-west-2.amazonaws.com/djl-inference:0.30.0-neuronx-sdk2.20.1 |
| DJLServing 0.29.0 with TensorRT-LLM 0.11.0, HuggingFace Transformers 4.42.4, and HuggingFace Accelerate 0.32.1 | inference | GPU | 3.10 (py310) | 763104351884.dkr.ecr.us-west-2.amazonaws.com/djl-inference:0.29.0-tensorrtllm0.11.0-cu124 |
| DJLServing 0.29.0 with LMI Dist 11.0.0, HuggingFace Transformers 4.43.2, HuggingFace Accelerate 0.32.1 | inference | GPU | 3.10 (py310) | 763104351884.dkr.ecr.us-west-2.amazonaws.com/djl-inference:0.29.0-lmi11.0.0-cu124 |
Expand Down
75 changes: 57 additions & 18 deletions release_images_inference.yml
Original file line number Diff line number Diff line change
Expand Up @@ -57,19 +57,6 @@ release_images:
disable_sm_tag: False
force_release: False
5:
framework: "djl"
version: "0.27.0"
arch_type: "x86"
inference:
device_types: [ "gpu" ]
python_versions: [ "py310" ]
os_version: "ubuntu22.04"
deepspeed_version: "0.12.6"
cuda_version: "cu121"
example: False
disable_sm_tag: True
force_release: False
6:
framework: "huggingface_pytorch"
version: "2.1.0"
hf_transformers: "4.37.0"
Expand All @@ -81,7 +68,7 @@ release_images:
example: False
disable_sm_tag: False
force_release: False
7:
6:
framework: "huggingface_pytorch"
version: "2.1.0"
hf_transformers: "4.37.0"
Expand All @@ -94,7 +81,7 @@ release_images:
example: False
disable_sm_tag: False
force_release: False
8:
7:
framework: "huggingface_pytorch"
version: "1.13.1"
arch_type: "x86"
Expand All @@ -107,7 +94,7 @@ release_images:
example: False
disable_sm_tag: True
force_release: False
9:
8:
framework: "huggingface_pytorch"
version: "1.13.1"
arch_type: "x86"
Expand All @@ -120,7 +107,7 @@ release_images:
example: False
disable_sm_tag: False
force_release: False
10:
9:
framework: "huggingface_pytorch"
version: "2.1.2"
arch_type: "x86"
Expand All @@ -133,7 +120,7 @@ release_images:
example: False
disable_sm_tag: True
force_release: False
11:
10:
framework: "autogluon"
version: "1.1.1"
arch_type: "x86"
Expand All @@ -145,3 +132,55 @@ release_images:
example: False
disable_sm_tag: False
force_release: False
11:
framework: "djl"
version: "0.27.0"
arch_type: "x86"
inference:
device_types: [ "gpu" ]
python_versions: [ "py310" ]
os_version: "ubuntu22.04"
tensorrtllm_version: "0.8.0"
cuda_version: "cu122"
example: False
disable_sm_tag: True
force_release: False
12:
framework: "djl"
version: "0.28.0"
arch_type: "x86"
inference:
device_types: [ "gpu" ]
python_versions: [ "py310" ]
os_version: "ubuntu22.04"
tensorrtllm_version: "0.9.0"
cuda_version: "cu122"
example: False
disable_sm_tag: True
force_release: False
13:
framework: "djl"
version: "0.29.0"
arch_type: "x86"
inference:
device_types: [ "gpu" ]
python_versions: [ "py310" ]
os_version: "ubuntu22.04"
tensorrtllm_version: "0.11.0"
cuda_version: "cu124"
example: False
disable_sm_tag: True
force_release: False
14:
framework: "djl"
version: "0.30.0"
arch_type: "x86"
inference:
device_types: [ "gpu" ]
python_versions: [ "py310" ]
os_version: "ubuntu22.04"
tensorrtllm_version: "0.12.0"
cuda_version: "cu125"
example: False
disable_sm_tag: True
force_release: False
Original file line number Diff line number Diff line change
Expand Up @@ -19,6 +19,7 @@
from sagemaker.predictor import Predictor
from sagemaker.serializers import JSONSerializer
from sagemaker.deserializers import JSONDeserializer
from packaging.version import Version

from ...integration import model_dir, ROLE, pt_model, tf_model
from ...utils import local_mode_utils
Expand Down Expand Up @@ -60,6 +61,8 @@ def _assert_prediction(predictor):
@pytest.mark.model("tiny-distilbert")
@pytest.mark.team("sagemaker-1p-algorithms")
def test_serve_json(docker_image, framework_version, sagemaker_local_session, instance_type):
if "huggingface-pytorch" in docker_image and Version(framework_version) < Version("2.4"):
pytest.skip("Skipping distilbert SM local tests for PT")
with _predictor(
model_dir, docker_image, framework_version, sagemaker_local_session, instance_type
) as predictor:
Expand Down

0 comments on commit cf51866

Please sign in to comment.