You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
D:\work\AI service\venv\Lib\site-packages\langchain\chat_models\__init__.py:32: LangChainDeprecationWarning: Importing chat models from langchain is deprecated. Importing from langchain will no longer be supported as of langchain==0.2.0. Please import from langchain-community instead:
`from langchain_community.chat_models import ChatOpenAI`.
To install langchain-community run `pip install -U langchain-community`.
warnings.warn(
D:\work\AI service\venv\Lib\site-packages\langchain\chat_models\__init__.py:32: LangChainDeprecationWarning: Importing chat models from langchain is deprecated. Importing from langchain will no longer be supported as of langchain==0.2.0. Please import from langchain-community instead:
`from langchain_community.chat_models import ChatOllama`.
To install langchain-community run `pip install -U langchain-community`.
warnings.warn(
Adding: 'The month is October.'
No chunks, creating a new one
Created new chunk (7cb66): Date & Times
Adding: 'The year is 2023.'
D:\work\AI service\venv\Lib\site-packages\langchain_core\_api\deprecation.py:139: LangChainDeprecationWarning: LangChain has introduced a method called `with_structured_output` thatis available on ChatModels capable of tool calling.You can read more about the method here: https://python.langchain.com/docs/modules/model_io/chat/structured_output/Please follow our extraction use case documentation for more guidelineson how to do information extraction with LLMs.https://python.langchain.com/docs/use_cases/extraction/.If you notice other issues, please provide feedback here:https://github.com/langchain-ai/langchain/discussions/18154
warn_deprecated(
Based on the summary, I think this proposition should be joined with the chunk "Date & Times". Therefore, I return the chunk ID: 7cb66.
Traceback (most recent call last):
File "D:\work\AI service\venv\Lib\site-packages\langchain_core\output_parsers\openai_functions.py", line 31, in parse_result
func_call = copy.deepcopy(message.additional_kwargs["function_call"])
~~~~~~~~~~~~~~~~~~~~~~~~~^^^^^^^^^^^^^^^^^
KeyError: 'function_call'
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "D:\work\AI service\test.py", line 29, in <module>
ac.add_propositions(propositions)
File "D:\work\AI service\utils\agentic_chunker.py", line 42, in add_propositions
self.add_proposition(proposition)
File "D:\work\AI service\utils\agentic_chunker.py", line 55, in add_proposition
chunk_id = self._find_relevant_chunk(proposition)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "D:\work\AI service\utils\agentic_chunker.py", line 308, in _find_relevant_chunk
print(extraction_chain.invoke(chunk_found))
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "D:\work\AI service\venv\Lib\site-packages\langchain\chains\base.py", line 166, in invoke
raise e
File "D:\work\AI service\venv\Lib\site-packages\langchain\chains\base.py", line 156, in invoke
self._call(inputs, run_manager=run_manager)
File "D:\work\AI service\venv\Lib\site-packages\langchain\chains\llm.py", line 127, in _call
return self.create_outputs(response)[0]
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "D:\work\AI service\venv\Lib\site-packages\langchain\chains\llm.py", line 281, in create_outputs
result = [
^
File "D:\work\AI service\venv\Lib\site-packages\langchain\chains\llm.py", line 284, in <listcomp>
self.output_key: self.output_parser.parse_result(generation),
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "D:\work\AI service\venv\Lib\site-packages\langchain_core\output_parsers\openai_functions.py", line 219, in parse_result
result = super().parse_result(result)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "D:\work\AI service\venv\Lib\site-packages\langchain_core\output_parsers\openai_functions.py", line 202, in parse_result
_result = super().parse_result(result)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "D:\work\AI service\venv\Lib\site-packages\langchain_core\output_parsers\openai_functions.py", line 33, in parse_result
raise OutputParserException(f"Could not parse function call: {exc}")
langchain_core.exceptions.OutputParserException: Could not parse function call: 'function_call'
I changed the init function into this:
class AgenticChunker:
def __init__(self):
self.chunks = {}
self.id_truncate_limit = 5
# Whether or not to update/refine summaries and titles as you get new information
self.generate_new_metadata_ind = True
self.print_logging = True
self.llm = ChatOllama(model='llama3', base_url='http://localhost:9090')
def add_propositions(self, propositions):
for proposition in propositions:
self.add_proposition(proposition)
def add_proposition(self, proposition):
if self.print_logging:
print (f"\nAdding: '{proposition}'")
...
How to fix it? Does ChatOllama and ChatOpenAI have anything in different?
The text was updated successfully, but these errors were encountered:
Hi, my issue looks like this
I changed the init function into this:
How to fix it? Does ChatOllama and ChatOpenAI have anything in different?
The text was updated successfully, but these errors were encountered: