From a2435a6e1cf8db209d774973cff09f75dcafb77d Mon Sep 17 00:00:00 2001 From: michaelfeil Date: Sun, 9 Jun 2024 15:56:12 -0700 Subject: [PATCH] commit transformers bump --- .../transformer/embedder/sentence_transformer.py | 2 +- .../infinity_emb/transformer/vision/torch_vision.py | 4 ++-- libs/infinity_emb/poetry.lock | 8 ++++---- libs/infinity_emb/pyproject.toml | 2 +- libs/infinity_emb/tests/end_to_end/test_ct2_sentence.py | 4 +++- 5 files changed, 11 insertions(+), 9 deletions(-) diff --git a/libs/infinity_emb/infinity_emb/transformer/embedder/sentence_transformer.py b/libs/infinity_emb/infinity_emb/transformer/embedder/sentence_transformer.py index a15f6fcd..9c25d0c4 100644 --- a/libs/infinity_emb/infinity_emb/transformer/embedder/sentence_transformer.py +++ b/libs/infinity_emb/infinity_emb/transformer/embedder/sentence_transformer.py @@ -81,7 +81,7 @@ def __init__(self, *, engine_args=EngineArgs): ) if engine_args.compile: - logger.info("using torch.compile()") + logger.info("using torch.compile(dynamic=True)") fm.auto_model = torch.compile(fm.auto_model, dynamic=True) def encode_pre(self, sentences) -> dict[str, "Tensor"]: diff --git a/libs/infinity_emb/infinity_emb/transformer/vision/torch_vision.py b/libs/infinity_emb/infinity_emb/transformer/vision/torch_vision.py index 7748aba2..7a9b909e 100644 --- a/libs/infinity_emb/infinity_emb/transformer/vision/torch_vision.py +++ b/libs/infinity_emb/infinity_emb/transformer/vision/torch_vision.py @@ -40,8 +40,8 @@ def __init__(self, *, engine_args: EngineArgs): self.engine_args = engine_args if engine_args.compile: - self.model.vision_model = torch.compile(self.model.vision_model) - self.model.text_model = torch.compile(self.model.text_model) + self.model.vision_model = torch.compile(self.model.vision_model, dynamic=True) + self.model.text_model = torch.compile(self.model.text_model, dynamic=True) assert hasattr( self.model, "get_text_features" diff --git a/libs/infinity_emb/poetry.lock b/libs/infinity_emb/poetry.lock index 5cb49342..1ba09b08 100644 --- a/libs/infinity_emb/poetry.lock +++ b/libs/infinity_emb/poetry.lock @@ -3843,13 +3843,13 @@ telegram = ["requests"] [[package]] name = "transformers" -version = "4.39.1" +version = "4.39.3" description = "State-of-the-art Machine Learning for JAX, PyTorch and TensorFlow" optional = true python-versions = ">=3.8.0" files = [ - {file = "transformers-4.39.1-py3-none-any.whl", hash = "sha256:df167e08b27ab254044a38bb7c439461cd3916332205416e9b6b1592b517a1a5"}, - {file = "transformers-4.39.1.tar.gz", hash = "sha256:ab9c1e1912843b9976e6cc62b27cd5434284fc0dab465e1b660333acfa81c6bc"}, + {file = "transformers-4.39.3-py3-none-any.whl", hash = "sha256:7838034a12cca3168247f9d2d1dba6724c9de3ae0f73a108258c6b8fc5912601"}, + {file = "transformers-4.39.3.tar.gz", hash = "sha256:2586e5ff4150f122716fc40f5530e92871befc051848fbe82600969c535b762d"}, ] [package.dependencies] @@ -4677,4 +4677,4 @@ vision = ["pillow"] [metadata] lock-version = "2.0" python-versions = ">=3.9,<3.13" -content-hash = "d372a37b5f055c1ad5052394ad0591aae6af764fc751ab9601ec64672bcc17db" +content-hash = "7703978e9387623b1731fb1b1c89919d8858fc4c4ee8c58296771a6fd26c842f" diff --git a/libs/infinity_emb/pyproject.toml b/libs/infinity_emb/pyproject.toml index 6530cd02..13ed441d 100644 --- a/libs/infinity_emb/pyproject.toml +++ b/libs/infinity_emb/pyproject.toml @@ -24,7 +24,7 @@ pydantic = {version = ">=2.4.0,<3", optional=true} # backend torch = {version = "2.3.1", source = "pypi", optional=true} sentence-transformers = {version = "^3.0.1", optional=true} -transformers = {version = ">4.34.0,<4.39.2", optional=true} +transformers = {version = ">4.34.0,<4.39.4", optional=true} ctranslate2 = {version = "^4.0.0", optional=true} optimum = {version = ">=1.16.2", optional=true, extras=["onnxruntime"]} hf_transfer = {version=">=0.1.5"} diff --git a/libs/infinity_emb/tests/end_to_end/test_ct2_sentence.py b/libs/infinity_emb/tests/end_to_end/test_ct2_sentence.py index 9d1f4b0c..6f85676e 100644 --- a/libs/infinity_emb/tests/end_to_end/test_ct2_sentence.py +++ b/libs/infinity_emb/tests/end_to_end/test_ct2_sentence.py @@ -52,7 +52,9 @@ def test_load_model(model_base): s = ["This is a test sentence."] e1 = model_base.encode(s) e2 = CT2SentenceTransformer( - engine_args=EngineArgs(model_name_or_path=MODEL, device="cpu") + engine_args=EngineArgs( + model_name_or_path=MODEL, device="cpu", bettertransformer=False + ) ).encode(s) np.testing.assert_almost_equal(e1, e2, decimal=6)