From e2dc13fb21a6a3bc5f1d436e1bf2f3ae84571858 Mon Sep 17 00:00:00 2001 From: awesomekoder Date: Sat, 31 Jan 2026 20:14:28 -0800 Subject: [PATCH] fix: rename loop variable to avoid shadowing Memory object Fixes #449 The loop variable 'memory' in the memory iteration was shadowing the Memory object from mem0, causing 'AttributeError: dict object has no attribute add' when memory.add() was called later. Renamed to 'mem' to match the pattern already used in other parts of the codebase. Affected files: - ai_travel_agent_memory/travel_agent_memory.py - multi_llm_memory/multi_llm_memory.py --- .../ai_travel_agent_memory/travel_agent_memory.py | 6 +++--- .../multi_llm_memory/multi_llm_memory.py | 6 +++--- 2 files changed, 6 insertions(+), 6 deletions(-) diff --git a/advanced_llm_apps/llm_apps_with_memory_tutorials/ai_travel_agent_memory/travel_agent_memory.py b/advanced_llm_apps/llm_apps_with_memory_tutorials/ai_travel_agent_memory/travel_agent_memory.py index ed57cfa..e0d17c2 100644 --- a/advanced_llm_apps/llm_apps_with_memory_tutorials/ai_travel_agent_memory/travel_agent_memory.py +++ b/advanced_llm_apps/llm_apps_with_memory_tutorials/ai_travel_agent_memory/travel_agent_memory.py @@ -70,9 +70,9 @@ if openai_api_key: relevant_memories = memory.search(query=prompt, user_id=user_id) context = "Relevant past information:\n" if relevant_memories and "results" in relevant_memories: - for memory in relevant_memories["results"]: - if "memory" in memory: - context += f"- {memory['memory']}\n" + for mem in relevant_memories["results"]: + if "memory" in mem: + context += f"- {mem['memory']}\n" # Prepare the full prompt full_prompt = f"{context}\nHuman: {prompt}\nAI:" diff --git a/advanced_llm_apps/llm_apps_with_memory_tutorials/multi_llm_memory/multi_llm_memory.py b/advanced_llm_apps/llm_apps_with_memory_tutorials/multi_llm_memory/multi_llm_memory.py index a251c1e..738812c 100644 --- a/advanced_llm_apps/llm_apps_with_memory_tutorials/multi_llm_memory/multi_llm_memory.py +++ b/advanced_llm_apps/llm_apps_with_memory_tutorials/multi_llm_memory/multi_llm_memory.py @@ -51,9 +51,9 @@ if openai_api_key and anthropic_api_key: relevant_memories = memory.search(query=prompt, user_id=user_id) context = "Relevant past information:\n" if relevant_memories and "results" in relevant_memories: - for memory in relevant_memories["results"]: - if "memory" in memory: - context += f"- {memory['memory']}\n" + for mem in relevant_memories["results"]: + if "memory" in mem: + context += f"- {mem['memory']}\n" full_prompt = f"{context}\nHuman: {prompt}\nAI:"