DPT-Agent achieved the best performance across the majority of models, especially on the widely recognized general-purpose SOTA models like GPT-4o. This phenomenon aligns with the conclusions from the experiments in single-agent settings, where larger models can overcome the latency limitations and achieve better performance with the help of DPT-Agent. Such performance improvements are more noticeable in the reasoning model series of GPT o3-mini and DeepSeek-R1. DPT-Agent framework can help reasoning models, which require long periods of thinking, overcome the latency and effectively transition from thinking to action. Additionally, when facing rule-based agents that can only perform a single task, DPT-Agent can maintain a high contribution rate. For some models like Llama3.3-70b, DPT-Agent w/o ToM outperforms the complete DPT-Agent, which may be closely related to the model's ToM capabilities.