from smolagents import ( CodeAgent, HfApiModel, TransformersModel, OpenAIServerModel, load_tool, tool, ) import os import datetime import pytz import yaml from tools import ( FinalAnswerTool, VisitWebpageTool, DuckDuckGoSearchTool, # This is also built-in in smolagents ) # Below is an example of a tool that does nothing. Amaze us with your creativity ! @tool def my_custom_tool( arg1: str, arg2: int ) -> str: # it's import to specify the return type # Keep this format for the description / args / args description but feel free to modify the tool """A tool that does nothing yet Args: arg1: the first argument arg2: the second argument """ return "What magic will you build ?" @tool def get_current_time_in_timezone(timezone: str) -> str: """A tool that fetches the current local time in a specified timezone. Args: timezone: A string representing a valid timezone (e.g., 'America/New_York'). """ try: # Create timezone object tz = pytz.timezone(timezone) # Get current time in that timezone local_time = datetime.datetime.now(tz).strftime("%Y-%m-%d %H:%M:%S") return f"The current local time in {timezone} is: {local_time}" except Exception as e: return f"Error fetching time for timezone '{timezone}': {str(e)}" final_answer = FinalAnswerTool() if IS_IN_HF_SPACE := os.getenv("SPACE_ID"): model_id = ( # BUG: Seems we are failed to call this model => bad request "https://wxknx1kg971u7k1n.us-east-1.aws.endpoints.huggingface.cloud" # it is possible that this model may be overloaded if True # BUG: Model not loaded on the server: https://api-inference.huggingface.co/models/Qwen/Qwen2.5-Coder-32B-Instruct/v1/chat/completions. Please retry with a higher timeout (current: 120) # BUG: TooManyRequests: Please log in or use a HF access token else "Qwen/Qwen2.5-Coder-32B-Instruct" # The default value of HfApiModel ) print("Using HfApiModel with model_id:", model_id) model = HfApiModel( max_tokens=2096, temperature=0.5, model_id=model_id, custom_role_conversions=None, ) else: from dotenv import load_dotenv # NOTE: to load the Hugging Face API Key curr_dir = os.path.dirname(os.path.abspath(__file__)) load_dotenv(os.path.join(curr_dir, "../../.env")) if OPENAI_API_KEY := os.getenv("OPENAI_API_KEY"): print("Using OpenAIServerModel with model_id: gpt-3.5-turbo") model = OpenAIServerModel(model_id="gpt-3.5-turbo", api_key=OPENAI_API_KEY) else: # NOTE: this model is not good enough for agent print( "Using TransformersModel with model_id: HuggingFaceTB/SmolLM2-1.7B-Instruct" ) model = TransformersModel( model_id="HuggingFaceTB/SmolLM2-1.7B-Instruct", trust_remote_code=True ) # Import tool from Hub image_generation_tool = load_tool("agents-course/text-to-image", trust_remote_code=True) duckduckgo_tool = DuckDuckGoSearchTool() visit_webpage_tool = VisitWebpageTool() with open("prompts.yaml", "r") as stream: prompt_templates = yaml.safe_load(stream) agent = CodeAgent( model=model, tools=[ final_answer, my_custom_tool, get_current_time_in_timezone, image_generation_tool, duckduckgo_tool, visit_webpage_tool, ], # Add your tools here (don't remove final answer) max_steps=6, verbosity_level=1, grammar=None, planning_interval=None, name=None, description=None, prompt_templates=prompt_templates, ) if __name__ == "__main__": from Gradio_UI import GradioUI GradioUI(agent).launch()