diff --git a/agency/poetry.lock b/agency/poetry.lock index daa9bc1..7d852de 100644 --- a/agency/poetry.lock +++ b/agency/poetry.lock @@ -3034,13 +3034,13 @@ test = ["opentelemetry-sdk"] [[package]] name = "openinference-instrumentation-llama-index" -version = "1.4.2" +version = "2.0.0" description = "OpenInference LlamaIndex Instrumentation" optional = false python-versions = "<3.13,>=3.8" files = [ - {file = "openinference_instrumentation_llama_index-1.4.2-py3-none-any.whl", hash = "sha256:d7772e0637ae05ab1748f5a72fa09d3e00a28140a61792656ea9135e64ff3ea1"}, - {file = "openinference_instrumentation_llama_index-1.4.2.tar.gz", hash = "sha256:fddfe360a8445bf47aeae8a840543fda6b61a1877f4a0912d921c1d111bcf617"}, + {file = "openinference_instrumentation_llama_index-2.0.0-py3-none-any.whl", hash = "sha256:353040d94ab53c3023cac9aeac62d3c587744de23ac32b303db293191c0dd8ae"}, + {file = "openinference_instrumentation_llama_index-2.0.0.tar.gz", hash = "sha256:300208465fa4e549a71f6426966ae3cfef2a32d3f229141b3fcb2b3276faa90a"}, ] [package.dependencies] @@ -3053,8 +3053,8 @@ typing-extensions = "*" wrapt = "*" [package.extras] -instruments = ["llama-index-core (>=0.10.5)"] -test = ["llama-index (==0.10.36)", "llama-index-core (==0.10.36)", "llama-index-llms-openai", "opentelemetry-sdk", "respx"] +instruments = ["llama-index-core (>=0.10.43)"] +test = ["llama-index (==0.10.43)", "llama-index-core (==0.10.43)", "llama-index-llms-openai", "openinference-instrumentation-openai", "opentelemetry-sdk", "respx"] [[package]] name = "openinference-semantic-conventions" @@ -5873,4 +5873,4 @@ test = ["big-O", "importlib-resources", "jaraco.functools", "jaraco.itertools", [metadata] lock-version = "2.0" python-versions = "~3.12" -content-hash = "9793b3d3e8b8bc89807be86b35149be46dbde4046fd4aed4c88d21d4a083f808" +content-hash = "683741c26fda5fd99877a8b8bb57bd126c865d163a459ed65b6dccf58306638b" diff --git a/agency/pyproject.toml b/agency/pyproject.toml index daf3c65..d0829d1 100644 --- a/agency/pyproject.toml +++ b/agency/pyproject.toml @@ -27,7 +27,7 @@ llama-index-postprocessor-cohere-rerank = "^0.1.5" together = "^0.2.11" llama-index-vector-stores-qdrant = "^0.2.8" openinference-semantic-conventions = "^0.1.5" -openinference-instrumentation-llama-index = "^1.3.0" +openinference-instrumentation-llama-index = "^2.0.0" opentelemetry-exporter-otlp = "^1.23.0" dspy-ai = "^2.4.9" pydantic-settings = "^2.2.1"