Steven Gong
    • Create new note
    • Create a note from template
      • Sharing URL Link copied
      • /edit
      • View mode
        • Edit mode
        • View mode
        • Book mode
        • Slide mode
        Edit mode View mode Book mode Slide mode
      • Customize slides
      • Note Permission
      • Read
        • Only me
        • Signed-in users
        • Everyone
        Only me Signed-in users Everyone
      • Write
        • Only me
        • Signed-in users
        • Everyone
        Only me Signed-in users Everyone
      • Engagement control Commenting, Suggest edit, Emoji Reply
    • Invite by email
      Invitee

      This note has no invitees

    • Publish Note

      Share your work with the world Congratulations! 🎉 Your note is out in the world Publish Note

      Your note will be visible on your profile and discoverable by anyone.
      Your note is now live.
      This note is visible on your profile and discoverable online.
      Everyone on the web can find and read all notes of this public team.
      See published notes
      Unpublish note
      Please check the box to agree to the Community Guidelines.
      View profile
    • Commenting
      Permission
      Disabled Forbidden Owners Signed-in users Everyone
    • Enable
    • Permission
      • Forbidden
      • Owners
      • Signed-in users
      • Everyone
    • Suggest edit
      Permission
      Disabled Forbidden Owners Signed-in users Everyone
    • Enable
    • Permission
      • Forbidden
      • Owners
      • Signed-in users
    • Emoji Reply
    • Enable
    • Versions and GitHub Sync
    • Note settings
    • Note Insights New
    • Engagement control
    • Make a copy
    • Transfer ownership
    • Delete this note
    • Save as template
    • Insert from template
    • Import from
      • Dropbox
      • Google Drive
      • Gist
      • Clipboard
    • Export to
      • Dropbox
      • Google Drive
      • Gist
    • Download
      • Markdown
      • HTML
      • Raw HTML
Menu Note settings Note Insights Versions and GitHub Sync Sharing URL Create Help
Create Create new note Create a note from template
Menu
Options
Engagement control Make a copy Transfer ownership Delete this note
Import from
Dropbox Google Drive Gist Clipboard
Export to
Dropbox Google Drive Gist
Download
Markdown HTML Raw HTML
Back
Sharing URL Link copied
/edit
View mode
  • Edit mode
  • View mode
  • Book mode
  • Slide mode
Edit mode View mode Book mode Slide mode
Customize slides
Note Permission
Read
Only me
  • Only me
  • Signed-in users
  • Everyone
Only me Signed-in users Everyone
Write
Only me
  • Only me
  • Signed-in users
  • Everyone
Only me Signed-in users Everyone
Engagement control Commenting, Suggest edit, Emoji Reply
  • Invite by email
    Invitee

    This note has no invitees

  • Publish Note

    Share your work with the world Congratulations! 🎉 Your note is out in the world Publish Note

    Your note will be visible on your profile and discoverable by anyone.
    Your note is now live.
    This note is visible on your profile and discoverable online.
    Everyone on the web can find and read all notes of this public team.
    See published notes
    Unpublish note
    Please check the box to agree to the Community Guidelines.
    View profile
    Engagement control
    Commenting
    Permission
    Disabled Forbidden Owners Signed-in users Everyone
    Enable
    Permission
    • Forbidden
    • Owners
    • Signed-in users
    • Everyone
    Suggest edit
    Permission
    Disabled Forbidden Owners Signed-in users Everyone
    Enable
    Permission
    • Forbidden
    • Owners
    • Signed-in users
    Emoji Reply
    Enable
    Import from Dropbox Google Drive Gist Clipboard
       Owned this note    Owned this note      
    Published Linked with GitHub
    • Any changes
      Be notified of any changes
    • Mention me
      Be notified of mention me
    • Unsubscribe
    # General response: We thank all reviewers for their insightful comments and acknowledgment of our contributions. We highlight the major contributions of our work as follows: * The creation and development of a new benchmark, entailing a substantial amount of effort (reviwer tSpT). Our benchmark differs from prior benchmark featuring multiple different task objectives in a single episode. Agents need to collaborate to achieve maximum collaboration efficiency. * An infrastructure to explore the potential of LLMs in multi-agent planning (reviewer RwfK ) and its potential of collaborating with human users. In addition, we perform cross-domain experiments to transfer our infrastructure into minecraft to demonstrate its generality. * The introduction of a novel metric, CoS (Collaboration Score), to assess collaboration capabilities (reviewer tSpT, reviewer DUbr ). * We conduct extensive experiments involving more than two agents; we also conduct experiments with human players. (reviewer YZDa). * We demonstrate that our framework can be transferred to a novel domain, like Minecraft. (reviewer YZDa) We’ve revised our manuscript per the reviewers’ suggestions (highlighted in red in the uploaded revision pdf) in both main text and the appendix. Detailed responses to each reviewer’s concerns are carefully addressed point-by-point. Below summarize the major updates we’ve made: * We updated the font size of Figure 4. * We provide more clear definitions for the equations used in the paper. * We updated the manuscript to reflect how we compute \tau (task interval) for different task levels. * We add discussions on the drawbacks of leveraging LLMs as a multi-agent planner. * We add more discussions on incorporating domain knowledge into the infrastructure. * We add missing references and discuss their relevance to our paper. > Gao, Yiming, et al. Towards Effective and Interpretable Human-Agent Collaboration in MOBA Games: A Communication Perspective in ICLR 2023 > Wang, Rose et al. Too many cooks: Bayesian inference for coordinating multi-agent collaboration in CogSci 2020 > Zhang, Hongxin et al. Building cooperative embodied agents modularly with large language models > Jain, Unnat et al. A Cordial Sync: Going Beyond Marginal Policies for Multi-Agent Embodied Tasks in ECCV 2020 * We add an analysis section on the common failure modes. * We include more details on how to transfer MineAgent to real time games. * We add details about prompt engineering process. * We add more details on how to transfer our infrastructure to a different gaming domain. * We append a full interaction log to the end of the appendix. # Response to Reviewer t1AF We sincerely thank you for your time and constructive comments. Below, we provide detailed replies to your comments and hope we can resolve your major concerns. > The font size of Figure 4 is too small. Thanks for the suggestion, we will update the paper accordingly. > The paper seems not provide a clear definition of the terms . Thanks for pointing this out. We apologize for the confusions. Here q refers to the utility/rewards the system will generate if the agent perform the specified action. C refers to the cost of the action. We have updated the paper to make this point more clear. # Response to Reviewer tSpT We sincerely thank you for your time and constructive comments. Below, we provide detailed replies to your comments and hope we can resolve your major concerns. > The proposed LLM coordination system, MineAgent, lacks novelty as its approach of leveraging scratchpad or memory has been extensively explored in prior research. First, we want to emphasize that compare to the generative agents, we are using a centralized coordination planning. In our work, we have a complex combinatorial action space, each agent has a action space of 221 due to the presence of multiple ingredients, multiple tools available in the environment., and with four agents, the action space becomes 2385443281 due to exponential effect of adding more agents. In addition, our work explored sophisticated spatial-temporal reasoning. To efficiently collaborate, agents need to work on multiple different tasks at the same time, manage to avoid space conflict (two or more agents cannot work on the same tools), etc. In addition, compared with concurrent work focus on planning (ex. [1]), we have more agents, more environments, and our environments are extensible and adaptable. Thirdly, we created a new benchmark with extensive experiment. We even perform cross-domain experiments to adapt our settings to Minecraft. [1] JARVIS-1: Open-world Multi-task Agents with Memory-Augmented Multimodal Language Models > The proposed environment features limited state and action spaces and furnishes all the requisite recipes to solve tasks, possibly oversimplifying the challenge. In this situation, heuristic or RL planners could be readily employed. However, the paper does not provide comparisons with these approaches. Thank you for the comments. First of all, we should point out that **the motivation of this paper is not about proposing a set of planning tasks to compare planning proficiency between LLMs and other canonical planning methods (ex. what [1] did)**, but the following (**in our abstract**): ``` However, despite the introduction of numerous gaming frameworks, the community lacks adequate benchmarks that support the implementation of a general multi-agent infrastructure encompassing collaboration between LLMs and human-NPCs. We propose a novel infrastructure-MindAgent—for evaluating planning and coordination capabilities in the context of gaming interaction. ``` That is, we simply would like to explore whether LLMs, which enjoys a powerful and native natural language interface that can interact with humans, can support sohphsticated multi-agent collaboration and human-NPC collaboration tasks, and this results in our benchmark. We've made this very clear -- the subject of study is **collaboration between LLMs and human-NPCs**. That being said, we do agree that showing some results of the canonical planning approach, including RL, could offer a better understanding of the dynamics of the collaboration tasks. Below, we try to address your concern on action space, alternative methods, etc: We respectfully disagree the reviewer's observation regarding the "limited state and action spaces" of our proposed environment. However, we argue that the complexity of our environment is not trivial due to the presence of multiple and duplicated tools and ingredients. In a single-agent scenario, the action space expands to 221, and with four distinct agents, it grows exponentially to approximately 2.39 billion. This immense action space presents a unique challenge specific to multi-agent research. Addressing the enormous action space, we implemented a masked version of Proximal Policy Optimization (PPO) approach. In additon, we provide heavily engineered dense reward. Despite leveraging privileged information (masked action, what action is valid during the execution), the RL strategies show limited success, even in two-agent settings. For level 1, PPO achieve 45.14% success rate and for level 4, PPO achieves 59.8% for \tau=2.5 after training for 30M steps or 500, 000 episodes. In comparison, GPT-4 agent achieves 95% success rate on both settings. Further the trained RL policy cannot adapt to varying numbers of agents and to new tasks, or even a change in the number of ingredients. This limitation underscores the complexity and dynamic nature of our environment. [1] https://arxiv.org/abs/2302.06706 > The proposed environment bypasses low-level control, further oversimplifying the problem. From my understanding, an agent can move to any location with a single action, without the need to consider spatial information. This eliminates the need for spatial reasoning in the LLM planner. Thank you for your insightful feedback. We acknowledge the reviewer's concern regarding the bypassing of low-level control in our environment, which indeed simplifies the problem to an extent. However, our research is intentionally focused on multi-agent high-level planning in gaming scenarios, a domain that presents substantial challenges in its own right [1], also as evidenced by our results where even the best agent achieved only a limited CoS score below 0.7. The complexity in high-level planning is underscored by the need for sophisticated spatial-temporal reasoning, such as resolving conflicts when multiple agents vie for the same location, task planning when there are multiple dependencies. Our work aims to push the boundaries of what Large Language Model-based Multi-agent Systems (MAS) can achieve in this high-level coordination and planning context. By concentrating on this aspect, we contribute valuable insights into the capabilities and limitations of LLMs in complex, multi-agent environments. We have indeed explored the integration of low-level actions in our Minecraft experiment, underscoring our recognition of its importance. However, for the scope of the current work, especially in the Cuisineworld scenario, we consciously chose to focus on high-level planning. This decision aligns with our objective to deeply understand and advance LLM capabilities in coordinating multiple agents before introducing the additional complexities of low-level control. Furthermore, it's pertinent to note that our approach aligns with concurrent research in the field. For instance, in single-agent planning systems like DEPS [2] and JARVIS-1[3], the low-level control is managed through Reinforcement Learning (RL) or Imitation Learning (IL) controllers, while LLMs oversee high-level planning. This also appies to Voyager[4], where the low-level control is scripted with the help of MineFlyer, a library for automating agent actions. This division of labor, separating high-level planning from low-level control, has proven effective and is a recognized methodology in current research community. [1] Korsah et al. A comprehensive taxonomy for multi-robot task allocation. [2] Describe, Explain, Plan and Select: Interactive Planning with Large Language Models Enables Open-World Multi-Task Agents [3] JARVIS-1: Open-world Multi-task Agents with Memory-Augmented Multimodal Language Models [4] Voyager: An Open-Ended Embodied Agent with Large Language Models > The paper claims that the LLM can seamlessly adapt to new planning problems across different domains, but this assertion is questionable considering the dependence on context length. While the simplicity of the current setting allows all environment information to be described within 1K tokens, this approach may prove inadequate for more challenging environments. Additionally, engaging in manual prompt engineering to enhance performance may entail significant monetary costs. We thank the reviewer for pointing out that context length might be a issue for general planning. Indeed, there could be more sophisticated tasks with more complicated and long interaction. But: 1. First of all, MindAgent aims at **benchmarking the multi-agent gaming interaction planning of LLMs and between LLMs and humans**. Per our results suggested, even for this relatively simplified scenarios with shorter interaction history and limited context-length requirement, the collaboration score can still low, meaning that the context-length is not the major concern given the current collaboration proficiency of LLMs. 2. Moreover, even from the perspective of how can LLMs handle the context length, the context length is becoming less of a issue as the current version of GPT-4 can handle 128k context tokens in a single prompt,the approximately equivalent of 300 pages of text as suggested by OpenAI[6], which could allow sufficient interaction history. Moreover, we don’t need to look at the full interaction history as is the case in our work. History trimming is a common practice as in [4,5] and can further mitigate the context length limit. In addition, for tasks with interaction that could be extremely complicated, we can optional use memory-based approach as in [1,2,3]. [1] Retrieval augmented generation for knowledge-intensive nlp tasks [2] Generation-augmented retrieval for open-domain question answering [3] JARVIS-1: Open-world Multi-task Agents with Memory-Augmented Multimodal Language Models [4] Compressing Context to Enhance Inference Efficiency of Large Language Models [5] LLMLingua: Compressing Prompts for Accelerated Inference of Large Language Models [6] https://openai.com/blog/new-models-and-developer-products-announced-at-devday > The description of the environment setting requires further clarification. Does \tau_int,(1) mean that a new task will be added at every timestep? What is the maximum horizon of an episode? We apologize for the confusion. \tau_int,(1) means a new task will be added after a fixed time step. The specific time value of \tau_int(1) depends on the environment complexity. We compute this value based on the minimum number of steps to complete the task in a single agent setting. Here are more details on how the task complexity and subsequently, \tau_int is determined, these can also be found in Appendix Section E.4 In our approach, we initially construct a task graph delineating subgoals, which serves as the foundation for our computations. We then apply breadth-first search in a single-agent context to determine the optimal task sequence. This sequence is a key component for calculating task intervals in multi-agent collaboration scenarios. For each tool requiring activation, we incorporate its activation wait time into the respective task intervals. Additionally, for each new connection in the task graph, we increase the total task interval time by tripling the edge time. We assume each subgoal requires at least, goto, get and put 3 actions. The cumulative task interval is subsequently adjusted by a scaling factor of 0.3 and the variable $\tau$. We pick the value $tau$ ranging from 1.0, 1.5, 2.0, 2.5 and 3.0 to represent different task difficulties. This process enables us to effectively compute task intervals tailored for multi-agent collaborative environments of varied difficulties. We have updated the manuscript to make this more clear. > The paper only highlights the emergent behavior of GPT-4 while neglecting to discuss its potential drawbacks. A more balanced perspective could be achieved by including examples of GPT-4's failure cases. We thank the reviewer for pointing this out. We have updated the paper accordingly. Here we list the common failure mode: 1. GPT-4 heavily rely on feedback, without feedback its performance drops significantly as indicated in our Table 2. 2. GPT-4 is sensitive to prompt inputs as indicated in Table3, without inference knowledge, the performance will drop. We have updated the paper with these new discussions. # Response to Reviewer DUbr > Technical novelty: The technical novelty of this work is lacking because the multi-agent setting is really just adding an LLM call to allocate what different agents should do. Concretely, it is a simple prompting and what is supposed to be an optimization problem is all packed into one LLM call. This doesn't seem to be a very principled way of studying the agent allocation. Thank you for the comments. First of all, we should point out that **the motivation of this paper is not about proposing a set of agent allocation tasks to compare planning proficiency between LLMs and other canonical planning methods (ex. what [1] did)**, but the following (**in our abstract**): ``` However, despite the introduction of numerous gaming frameworks, the community lacks adequate benchmarks that support the implementation of a general multi-agent infrastructure encompassing **collaboration between LLMs and human-NPCs**. We propose a novel infrastructure-MindAgent—for evaluating planning and coordination capabilities in the context of gaming interaction. ``` That is, we simply would like to explore whether LLMs, which enjoys a powerful and native natural language interface that can interact with humans, can support sohphsticated multi-agent collaboration and human-NPC collaboration tasks, and this results in our benchmark. We've made this very clear -- the subject of study is **collaboration between LLMs and human-NPCs**. That being said, we do agree that showing some results of the canonical planning approach, including RL, could offer a better understanding of the dynamics of the collaboration tasks. Below, we provide some results of using a canonical RL algorithm on the proposed benchmarks for LLMs: We implemented a masked version of Proximal Policy Optimization (PPO) approach. In additon, we provide heavily engineered dense reward. Despite leveraging privileged information (masked action, what action is valid during the execution), the RL strategies show limited success, even in two-agent settings. For level 1, PPO achieve 45.14% success rate and for level 4, PPO achieves 59.8% for \tau=2.5 after training for 30M steps or 500, 000 episodes. In comparison, GPT-4 agent achieves 95% success rate on both settings. Further the trained RL policy cannot adapt to varying numbers of agents and to new tasks, or even a change in the number of ingredients. This limitation underscores the complexity and dynamic nature of our environment. [1] https://arxiv.org/abs/2302.06706 > Problem setting: the multi-agent collaboration problem is reduced to a top-down allocation: a distributor distributes tasks for agents to do. But for a collaboration problem to work, there should also be communication between the agents, which this work does not study. We thank the reviewer for pointing out that our method is a top-down allocation. The multi-agent task planning problem [1], is a very difficult problem when they are multiple dependencies. We want to address that is different from previous settings as in [1], where the task is known priori. In our settings, the tasks are not known prior to the agents simulating the experience of a real kitchen, dish orders keep flooding in. This modification will only push the original problem harder. We acknowledge that communication between agents is indeed a critical aspect of collaboration, and our current model does not explicitly study inter-agent communication. This is a limitation of our study, and we believe it represents an exciting avenue for future research. The integration of direct agent-to-agent communication would undoubtedly add a rich layer of complexity and realism to the collaboration model, offering insights into emergent behaviors, negotiation strategies, and decentralized decision-making processes. However, this addition will also add communication cost which might not be ideal for gaming applications where a faster response is desired. Therefore, in this specific work (MindAgent) that focuses on **collaboration between LLMs and human-NPCs** in **gaming scenarios**, we choose to anchor our study on a centralized planning scheme. [1] Korsah et al. A comprehensive taxonomy for multi-robot task allocation. > the title is also quite misleading: it is titled "emergent interaction" however, there is no real interaction between the agents, i.e., communication of agent's individual's ability, limitation, progress etc. It is more of a allocation / coordination problem of a central agent. We appreciate the reviewer's feedback regarding the title of our paper and understand the concerns raised about the term "emergent interaction." To clarify, our use of "interaction" in this context is twofold: - **Interaction with the Environment**: We emphasize the ability of agents to interact dynamically with the environment. This interaction does not require specialized training for specific tasks, showcasing emergent behaviors. - **Interaction with Human Gamers**: Another crucial aspect of interaction we explore is the collaboration and teaming with human players. Our agents are designed to adapt to human strategies and work alongside human gamers efficiently, achieving commendable performance without the need for specialized training to understand human gameplay. > Experimental results: The work evaluate on Minecraft environment, but there has been quite a few LLM-based Minecraft agents, and the work does not offer comparison between this method and existing works as baselines. We appreciate the reviewer’s insight regarding the comparison with existing LLM-based Minecraft agents. It is important to note that our work operates within a distinctly different framework compared to the referenced LLM-based agents in Minecraft. Our focus is on Multi-Agent Systems (MAS), which inherently introduces different challenges and dynamics compared to single-agent environments. This distinction in the setting - MAS versus single-agent systems - is crucial as it changes the nature of the problem and the applicable solutions. Despite the different settings, we have indeed incorporated some techniques from existing single-agent LLM-based works into our approach, for example leveraging feedback, incorporating demonstrations etc, which has been adopted by some seminal and concurrent work on LLM + gaming [1,2]. [1] Describe, Explain, Plan and Select: Interactive Planning with Large Language Models Enables Open-World Multi-Task Agents [2] JARVIS-1: Open-world Multi-task Agents with Memory-Augmented Multimodal Language Models # Response to Reviewer RwfK > Have the authors considered incorporating a human-agent communication module in MindAgent? As suggested by Gao et al. [1], introducing interpretable human-agent communication into collaborative games can effectively improve human-agent collaboration performance and human subjective preferences. Natural language is the best medium for human-agent communication, which is also a natural advantage of LLM-based agents in human-agent collaboration. Thank you for highlighting the potential of incorporating human-agent communication in MindAgent, as suggested by Gao et al. [1]. We agree that introducing interpretable human-agent communication in collaborative games is a promising direction for enhancing human-agent collaboration. Accordingly, we have updated our paper to include citations and discussions related to [1], acknowledging its relevance to our work. We adopted a straightforward method for facilitating communication between the human user and the LLM-based agent. Participants in the experiment were instructed to verbally express their intents or comments. These verbal inputs were then transcribed and integrated into the LLM’s context as “current human instruction: <instruction>”. Implementation and Results: This simple yet effective approach allowed for real-time adjustments in the behavior of the LLM-based agents based on human inputs. For instance, in our demo videos can be found in the supplementary videos, when a user suggested a specific strategy or highlighted a priority task, the agent was able to promptly adapt its actions accordingly. This led to a more dynamic and responsive collaboration between humans and agents, showcasing the potential of natural language communication in enhancing the collaborative experience. In summary, while our initial trials in human-agent communication are relatively basic, they clearly demonstrate the effectiveness of integrating natural language communication in LLM-based collaborative environments. We plan to further develop and refine this aspect of our system in future work, drawing inspiration from the findings of Gao et al. [1] and other seminal works in the field. > Can the design of MindAgent support collaborative games that require more domain-specific knowledge? For example, Multiplayer Online Battle Arena (MOBA) games [1], First-person Shooter (FPS) games [2], and Diplomacy [3] have very complex gameplay and their outcomes heavily depend on the planning and collaboration capabilities of the agents. The authors could discuss how the infrastructure needs to be modified when extending it to other games, which would enhance the generalisability of the infrastructure. We thank the reviewer for suggesting the exploration of domain-specific knowledge in collaborative games like MOBA, FPS, and Diplomacy, and its implications for the design of MindAgent. Our current work in CuisineWorld has indeed laid the groundwork for such adaptations, primarily through the use of In-Context Learning (ICL). In CuisineWorld and Minecraft, we inject domain-specific knowledge (such as recipes) directly into the context, which the Large Language Model System utilizes to inform its decision-making and collaboration strategies. This demonstrates the feasibility of adapting MindAgent to other domains where specific knowledge plays a crucial role. In addition, techniques like Retrieval-Augmented Generation (RAG) and Fine Tuning could be pivotal in further developing MindAgent's capabilities to handle such domain-specific complexities. Thanks for the suggestions, and we have updated the paper accordingly. # Response to Reviewer YZDa > The CuisineWorld, based on the video game Overcooked which is popular for measuring multi-agent cooperation and has many existing environments already[1][2]. The benchmark introduces more cuisines but falls short in diversifying kitchen maps when compared to previous Overcooked environments. We are grateful for your insightful comment regarding the design of our CuisineWorld benchmark in relation to existing environments based on the popular video game Overcooked. Our approach, indeed, differs from the environments presented in references [1] and [2]. While these existing environments offer a variety of kitchen maps, our benchmark is specifically tailored to explore high-level multi-agent planning and coordination using Large Language Model-based Multi-agent Systems (MAS). To achieve this, instead of focusing on diversifying kitchen maps, which often necessitates low-level control of agent motion, we have introduced several unique elements to enhance the complexity and diversity of the tasks: - **Enhanced Kitchen Layouts**: We have diversified the layout of kitchens by incorporating a wider range of **tools** (both in types and quantities) and **ingredients**. This approach differs from [1] and [2], where the emphasis is not on the variety of kitchen tools and ingredients. For example, in [1], there are only two types of ingredients and two types of tools. In [2], there are two types of ingredients, and two types of tools. In comparision, there are 27 unique ingredients, and 10 tools in CuisineWorld. - **Complex Task Design**: Our benchmark includes a broader spectrum of recipes, varying significantly in difficulty levels. This variation is not just in terms of the number of tools and ingredients required but also in the intermediate steps involved in each recipe. We invite you to refer to Figure 23 for a detailed illustration. This aspect of task complexity, particularly in the context of high-level planning, is not extensively explored in [1] and [2]. In [1,2] there are very limited number of dishes, 1 and 3 respectively. However, in CuisineWorld, there are 33 unique dishes. - **Multi-Dish Episodes and Collaborative Strategy Assessment**: We require agents to complete multiple dishes within a single episode, with the types of dishes varying to challenge and assess the collaborative strategies of the agents. Our level design ensures that there are shared intermediate steps among the types of dishes in a single episode. The system is tasked with multiple different goals at the same time. This approach allows us to use metrics like 'Collaborative Score' (CoS) to evaluate how agents collaborate to achieve higher dish throughput. This dynamic aspect of collaboration, especially in the context of dish expiration and shared tasks, offers a new dimension to the study of multi-agent cooperation, which is distinct from the environments in [1] and [2]. In [1] the goal is to finish as many dishes as possible in a limited amount of time. In [2], the goal is to finish one dish in the least amount of time. Both of them do not consider the density of the tasks (interval between dish orders coming to the kitchen) and its effect on coordination. As we mentioned earlier, this concpet (changing density of the tasks and measuring collaboration proficiency upon it) is at heart of CuisineWorld and our CoS metric, and it has demonstrated its effectiveness of benchmarking **collaboration between LLMs and human-NPCs** (as indicated in the abstract). > Table 12 made a good summary of related benchmarks though some related embodied multi-agent cooperation benchmarks with most of the features in the table are missing, such as [3][4]. Thanks for pointing this out. We updated the paper and added the missing references. > While CuisineWorld boasts an appealing visual design, it does not directly align with the paper's experiments. The paper primarily relies on text-based states and interactions (Figure 26, even during human experiments, where humans are restricted to a textual interface with lots of text description of the game state) We appreciate your observation regarding the use of text-based states and interactions in our experiments, as contrasted with the visually appealing design of CuisineWorld. You are correct in noting that our experiments primarily utilize text-based interfaces for thinking and coordination processes, which are then projected into the virtual world of the CuisineWorld game. This approach is indeed in line with several seminal works in the field, such as Voyager[1] and DEPS[2], where Large Language Model based Autonomous Systems operate primarily through text-based interactions due to the current limitations in powerful vision-language models that can process visual data directly. However, it's important to clarify that this text-based approach does not restrict human players to a purely textual interface. In fact, as demonstrated in our Minecraft experiment, human players can engage with the game in a more immersive Virtual Reality (VR) interface, enjoying natural gameplay while interacting with agent counterparts. This aspect of our research showcases the versatility of our infrastructure in accommodating different modes of human-agent interaction, catering to both research needs and user experience. The video demo can be found in the submitted supplementary materials. [1] Voyager: An Open-Ended Embodied Agent with Large Language Models [2] Describe, Explain, Plan and Select: Interactive Planning with Large Language Models Enables Open-World Multi-Task Agents > In section D.2, the statement "In this game, the actions of human players can be obtained from an inverse dynamic model by checking pre-conditions and post-effects" is confusing to me. How was Figure 22 obtained? It seems to be in real-time with human players using the keyboard to move, how's the time step defined here and the "goto" action for the LLM agent implemented? Thanks for pointing this out. This due to the differnce between the two versions of CuisineWorld, text-based game and the real-time VR game that human can play in VR. Unlike the step-by-step nature of the text-based version of CuisineWorld, the real-time VR game operates continuously. To align the LLM’s processing with this dynamic environment, we implemented a system where the LLM checks user actions at regular intervals during the game loop, referred to as "time steps." These time steps are defined as 0.1 seconds. Then we can ensure LLM can respond to user actions in a timely manner, matching the pace of the real-time game. In the real-time version, human players control their agents directly through keyboard inputs, resulting in low-level actions like moving or picking up items. However, the LLM-agent operates on a higher, more temporally-extended level of atomic actions. To bridge this gap, when the LLM checks user actions, it doesn't just read the keyboard inputs. Instead, it assesses changes in the high-level game state, such as the agent’s location, and the status of tools and ingredients. This assessment allows the LLM to infer the temporally extended actions that align with its text-based decision-making process. Implementing this required a simple inverse dynamics model through checking state changes to translate low-level actions into high-level atomic actions, facilitating a seamless transition from the text game to the real-time virtual game. Regarding the specific query about the 'GoTo' action, we utilized the A* pathfinding algorithm, integrated into the Unreal Engine, to facilitate this movement. We hope these updates and clarifications provide a better understanding of how we adapted the CuisineWorld game to a real-time virtual environment, and how we managed the differences in user and LLM-agent interactions between the text-based and real-time VR versions. > The benchmark predominantly employs a "new auto-metric collaboration score CoS for assessing the collaboration efficiency". However, this metric is defined as the average task success rate with different time intervals for each dish that appears highly tailored to the specific environment and lacks a clear connection to "cooperation efficiency". We thank you for pointing this out. Below, we start with some intuition on why CoS makes sense in CuisineWorld, then we will clarify how can this generalize to other domains as well: Rationale of CoS: - In the kitchen scenario as demonstrated in CuisineWorld, hypothetically, when the dish order come very rarely (with a large interval), no matter if there is any collaboration, high success rate can easily attained as there is sufficient time. - However, as we reduce the interval, more and more dish order are flooding in. If the agents (and humans) are able to collaborate well, the productivity will be high, or namely, they can still manage to maintain a decent success rate. On the contrary, teams with poor collaborate will likely suffer from a substantial drop on success rate as the interval gets smaller. The same collapse will ultimately happen to a good collaborative team too when the interval gets too small, but good collaboration can always sustain longer. - Therefore, it makes sense to use the averaged success rate across different intervals as an indicator of the collaboration proficiency. More importantly, such metric asks for a game setting where the dish order will keep coming, in a changing interval, which also aligns with the original Overcooked! game experiences. How to generalize to other games: - We've demonstrated that CoS is able to generalize to Minecraft (see table 5). We simply modify the interval of the crafting tasks and therefore measure the collaboration efficiency there. - In more general multi-agent gaming scenarios, as long as the goal of the game itself includes completing some tasks (which we believe could be a generally applicable point), CoS can be applied -- the productivity of a good team should be able to handle both when there is very few tasks or many more and therefore, CoS could measure their collaboration proficiency. [1] On the Utility of Learning about Humans for Human-AI Coordination [2] Too many cooks: Bayesian inference for coordinating multi-agent collaboration > The absence of a train/test split is concerning, as prompt engineering can substantially impact performance. A thorough understanding of how the prompt is tuned for different tasks is crucial. We appreciate the reviewer's concern regarding the absence of a traditional train/test split and the potential impact of prompt engineering on performance. Our approach with the CuisineWorld dataset is indeed unconventional, and we align with recent research efforts in probing datasets, such as the study highlighted in [1]. Our primary objective with CuisineWorld is to establish a grand challenge for LLM-based multi-agent collaboration and emergent gaming interaction. In this context, the entire dataset serves as a probing tool, intended to explore the capabilities and limitations of LLMs in complex, dynamic environments. We consciously chose not to enforce a train/test split to avoid limiting the scope of exploration and experimentation within the dataset. That being said, we do not encourage prompt engineering efforts that are specific for tweaking our benchmark but we believe openly inviting all avenues of approach on tackling this challenge can spark more interesting ideas and findings. [1] Winoground: Probing Vision and Language Models for Visio-Linguistic Compositionality > From C.3.4, level 3 has only two similar recipes which differ only with the words "salmon" and "tuna", only one demonstration may decrease the task difficulty significantly for the LMs, raising doubts about the benchmark formation. We appreciate your concern about the perceived task simplicity in Level 3 of CuisineWorld, where the recipes differ only in terms of "salmon" and "tuna." However, we respectfully disagree with the notion that a single demonstration significantly decreases task difficulty for LLMs. Our benchmark's design is not centered around creating complex recipes for LLMs to master, but rather on introducing and evaluating the complexity of collaboration through carefully designed recipe combinations. In Level 3, the challenge for LLMs is to effectively coordinate agents around shared tasks and bottleneck resources. A strategic collaboration example could involve assigning one agent to prepare cooked rice—a common component in all dishes—while others handle different kitchen stations, such as mixers or chopping boards. This requires the LLMs to identify and manage shared resources and task bottlenecks, and to establish an efficient “pipeline” for task execution, which significantly enhances throughput. Such strategic planning and coordination cannot be trivially inferred from a single demonstration. Furthermore, it is important to note that the demonstrations provided to the LLMs are based on Level 0 tasks, which are fundamentally different from those in Level 3. Therefore, the LLMs are essentially working on novel tasks that require adaptive and strategic thinking beyond replicating demonstrated actions. This aspects of our design underscores the benchmark's focus on assessing high-level planning and collaboration, rather than recipe complexity. > In section 5.1, The paper claims "more agents will lead to higher collaboration efficiencies. Thus, indicating that the LLM dispatcher can coordinate additional agents to execute tasks more efficiently". However, from Table 1, 4 agents perform worse than 3 agents in most scenarios, and even 2 agents achieve the highest score in levels 2 and 4, which contradicts the claim. It may provide more insights if the paper can provide more analysis on these contradictory results instead of only ablating on level_3, which seems to be the only level that "looks normal" (a.k.a 2 agents < 3 agents < 4 agents with somewhat clear gaps) Thank you for highlighting the discrepancies in our results concerning the relationship between the number of agents and collaboration efficiency. We have revised our claim to more accurately reflect that while generally increasing the number of agents can lead to higher efficiency, there are scenarios where adding more agents may actually hinder performance. Our revised claim now acknowledges that the relationship between the number of agents and task efficiency is not always linear and can be influenced by various factors, including task complexity and the nature of the agents' collaboration. We have added detailed discussions around this revised claim to explore these nuances and provide a more comprehensive understanding of when and why more agents may or may not lead to better performance. In our initial study, we chose Level 3 for our ablation study because it showed a clear pattern where increasing the number of agents improved performance. We've updated our paper to explain that this choice was based on the observation that the LLM performed well on this level, offering a stable platform to assess the impact of varying agent numbers. When LLMs do not perform well on certain levels, additional unknown factors might confound the results, making it challenging to isolate the effects of ablated factors on performance. In response to your suggestion, we have conducted a preliminary qualitative analysis and found the following failure mode: 1. Inability to prioritize the task order. Sometimes LLM will skip the task that is on the top of the task queue and leads to task expiration. Therefore, the success rate will drop. 2. Unable to understand occupy() state instruction. In our CuisineWorld, agents need to wait for a \textbf{varying} timesteps before the cooking is done. The wait time depends on the specific tool. When agents attempt to take out ingredients immediately after activating the tool, actions will fail. Instead of keeping waiting, the agents might work on other task components which will slow down the overall progress. 3. Unable to allocate agents to the correct subgoals when there are multiple concurrent dishes. This is especially the case when there are 4 agents. > Table 4, "For a fair comparison, all tests employed identical prompt inputs" using "identical" prompt for different LM families may not be "fair", especially if the prompt is "tuned" specifically for one model. More details on the prompt engineering process may help clarify these concerns. Thank you for your comment regarding the use of identical prompts across different Language Model (LM) families in Table 4. We understand your concern about the fairness of this approach, particularly if a prompt is more optimized for one model over others. In response, we have updated our paper with more detailed information about our prompt engineering process. Our initial approach began with a simple and generic prompt, designed to be as neutral as possible. This baseline prompt was tested across all models, including GPT-4, Claude, and other LLMs in our study. We observed that only GPT-4 and Claude were able to generate reasonable planning responses with this initial prompt. Consequently, we refined our prompt engineering efforts to better suit these two models, aiming to optimize their performance and response quality. For the other LLMs, we also attempted individual prompt engineering. However, these efforts did not yield significant improvements in their responses compared to the initial trial. After several rounds of testing and refinement, we concluded that further prompt customization for these models did not result in notably better outcomes. As a result, we decided to use a consistent prompt across all LLMs for the final comparison. This decision was made to maintain uniformity in testing conditions, despite the prompts being more closely tuned to GPT-4 and Claude. We believe this approach offers a reasonable balance between fairness and practicality in evaluating the capabilities of different LLMs under similar conditions. We hope this explanation clarifies our methodology and addresses your concerns regarding prompt engineering and fairness in our comparison. We are open to further discussion on this matter and welcome any additional thoughts or suggestions you may have. > Table 3 presents a perplexing scenario where four agents using a two-agent demo outperform four agents using a four-agent demo, albeit marginally. This result could benefit from a clearer explanation. Thank you for pointing out the intriguing result in Table 3 where four agents using a two-agent demonstration outperform those using a four-agent demonstration, albeit by a marginal degree. We have revised our analysis of these results to provide a clearer understanding. The primary focus of the results presented in Table 3 is that GPT-4 can leverage demonstration of a different number of agents to perform task planning and allocations for a team size, which remains valid and is not contradicted by the observed anomaly. However, we understand your concern that the comparison between GPT-4 leverags two-agent and four-agent demo could be a bit counterintuitive. Therefore, we visualize the result, here is what we found: - This might due to the fact that the task structure is friendly to a two agent or four agent team, which means with four agents, LLM can divide the agnets into two teams of two agents, with each team operating independently. This observation also highlight the model’s capability to perform task allocations and task plannings for more agents than if has previously seen. > Novel game adaptation of Minecraft seems very promising, but the details are extremely limited. It would be more convincing if there were more details and formal experiments on it. For example, how is the "adaptation" conducted? What's the additional human effort required (such as prompt engineering and rounds of re-playing for prompt tuning)? We appreciate your interest in our novel adaptation of Minecraft and acknowledge the need for more detailed information. Our approach to adapting Minecraft for our experiments was indeed multifaceted, encompassing both adjustments to the Language Model (LLM) and developments on the game development side. 1. **LLM Modifications**: On the LLM side, we focused on updating the background knowledge of the model. This included: - **Action Space Explanation**: We provided the LLM with detailed information about the possible actions and interactions within the Minecraft environment. This step was crucial to align the model's understanding with the game's mechanics. - **Recipe and Tool Definitions**: We also included comprehensive definitions of recipes, tools, and ingredients specific to Minecraft. This enriched the LLM's contextual knowledge, enabling it to make more informed decisions within the game. We believe these modifications are reasonable and necessary, as without these knowledge, it’s very difficult for model to operate in an unknown environments. 2. **Game Development Adjustments**: On the game development side, we dedicated efforts to: - **Text-to-Game Interaction Translation**: We developed code to translate text-based interactions and commands from the LLM into actionable inputs within the virtual game environment. This translation layer was key to bridging the gap between the LLM's text-based outputs and the game's interactive elements. We intentionally avoided extensive prompt engineering for adaptation to the new game domain. Except for the modified background knowledge mentioned above, our approach did not involve iterative prompt tuning processes. We believed that keeping prompt engineering to a minimum would provide a more authentic and generalizable understanding of the LLM's capabilities in adapting to new environments. We have updated the paper to include more details and we thanks the reviewer for pointing this out. > The method employed in the paper primarily relies on intensive prompt engineering, without introducing novel designs for multi-agent cooperation. It's well known that LLMs can perform "in-context learning", providing demonstrations and reasoning steps can help improve performance, there's nothing new to take away from the method and experiments. As mentioned above, the "emergent ability" of MindAgent is not well supported by the results. Providing only screenshots of part of the prompts as in Figures 6 and 7 is not enough. More details on the full prompt and game episode may help clarify these concerns. Thank you raising this. Here, we will first try to address your concern on "in-context learning" and take-away messages, then move to the emergent ability and prompting details. In-context learning and take-away message: - As we've illustrated in Fig. 6 and 7 in the supplementary, the demonstration we provide to the LLM **is always on the same simple task** across all level of games -- no matter how complex the dishes and collaboration regime within the current level can be, the demonstration, is always on a simple task with one dish "porkMeatCake", which is very similar to "salmonMeatCake", a dish that is categorized as "level_0" or "very simple" in Figure 8. - Compared to the previous in-context learning setting in LLMs, VLMs, where they simply learn from examples of the same task with similar complexity, the ability of **generalizing** from a demonstration on such simple task to more complex dishes/levels, with very simple prompt including game rules and recipies only, is rarely explored, especially in LLM planning and multi-agent collaboration. Our experiments conduct a comprehensive exploration on just that. We believe this can be one take-home message from our result. - Obviously, there are many more take-home messages as well, including how many agents the current LLMs are able to robustly coordinate under what complexity of tasks (from our main results), to which extent can LLMs generalize from a demonstration of fewer agents (from results in table 3), how does our proposed prompting techniques affect the collaboration (from ablations in table 2), how does different LLMs perform (from results in table 4), etc. Prompting details: - Figure 7 \& 8 show the **complete** prompt we use, there is indeed nothing more -- just game rules, recipes of the current level, and one-shot demonstration from a "very simple" portMeatCake task (yes, we use the same task as in-context demonstration for all levels). This is the **full prompt**. We're sorry for not making this clear. In any case, please let us know if there are further details on the prompt we need to clarify. Emergent ability: - We thank your for your suggestion. To visulize the emergent collaboration from our designed prompt (only rules, recipes, and one-shot demonstrations on a much simpler task), we've listed some gaming interaction in appendix H for cooking burgers. Considering the length of the full game log, we only show the interaction until the first dish is fulfilled. The full log can be found [here](https://drive.google.com/file/d/1pLViwnC_ngL4cArDnjhLqVchZKZ-6sp6/view?usp=sharing). # Re-response: We thank your reviewers for the favorable and constructive comments again. We address your main concers below and describe the every changes that made to the manuscript in our revision. > The CuisineWorld, based on the video game Overcooked which is popular for measuring multi-agent cooperation and has many existing environments already[1][2]. The benchmark introduces more cuisines but falls short in diversifying kitchen maps when compared to previous Overcooked environments. We addressed this comment in red-ink of Appendix Sec. E.1 in the new revision paper. > In section D.2, the statement "In this game, the actions of human players can be obtained from an inverse dynamic model by checking pre-conditions and post-effects" is confusing to me. How was Figure 22 obtained? It seems to be in real-time with human players using the keyboard to move, how's the time step defined here and the "goto" action for the LLM agent implemented? We addressed your comments in red-ink of Appendix Sec. E.2. (we add one more section in appendix of revision, so the Sec. E.2 map to the previous Sec. D.2 as your comments.) > The benchmark predominantly employs a "new auto-metric collaboration score CoS for assessing the collaboration efficiency". However, this metric is defined as the average task success rate with different time intervals for each dish that appears highly tailored to the specific environment and lacks a clear connection to "cooperation efficiency". We addressed this comments in red-ink of Appendix Sec. E.6 in our revision. > In section 5.1, The paper claims "more agents will lead to higher collaboration efficiencies. Thus, indicating that the LLM dispatcher can coordinate additional agents to execute tasks more efficiently". However, from Table 1, 4 agents perform worse than 3 agents in most scenarios, and even 2 agents achieve the highest score in levels 2 and 4, which contradicts the claim. It may provide more insights if the paper can provide more analysis on these contradictory results instead of only ablating on level_3, which seems to be the only level that "looks normal" (a.k.a 2 agents < 3 agents < 4 agents with somewhat clear gaps) We addressed your comments in red-ink of Section 5.1 findings. > Table 4, "For a fair comparison, all tests employed identical prompt inputs" using "identical" prompt for different LM families may not be "fair", especially if the prompt is "tuned" specifically for one model. More details on the prompt engineering process may help clarify these concerns. We describe the detailed prompt engineering process in red-ink Section B of our revision. > Table 3 presents a perplexing scenario where four agents using a two-agent demo outperform four agents using a four-agent demo, albeit marginally. This result could benefit from a clearer explanation. We addressed your comments in red-ink of Section 6.2 of our revision. > Novel game adaptation of Minecraft seems very promising, but the details are extremely limited. It would be more convincing if there were more details and formal experiments on it. For example, how is the "adaptation" conducted? What's the additional human effort required (such as prompt engineering and rounds of re-playing for prompt tuning)? We addressed your comments in red-ink of Appendix Sec. F.1 for our revision. > The method employed in the paper primarily relies on intensive prompt engineering, without introducing novel designs for multi-agent cooperation. It's well known that LLMs can perform "in-context learning", providing demonstrations and reasoning steps can help improve performance, there's nothing new to take away from the method and experiments. As mentioned above, the "emergent ability" of MindAgent is not well supported by the results. Providing only screenshots of part of the prompts as in Figures 6 and 7 is not enough. More details on the full prompt and game episode may help clarify these concerns. We addressed this issue in Appendix Section H and Section B. We add parts of the log as "The full interaction history for finishing one dish in level 9", which showed in Sec. H. and we provide the full game log [here](https://drive.google.com/file/d/1pLViwnC_ngL4cArDnjhLqVchZKZ-6sp6/view?usp=sharing) for your reference. Due the page limitaion, we can't attach the full game log in our revised paper. Hope for your understanding. > Table 12 made a good summary of related benchmarks though some related embodied multi-agent cooperation benchmarks with most of the features in the table are missing, such as [3][4]. We compare our work vs [3][4] in the Table 13 of Appendix Section B1 with red-ink, and we cited the papers and discussed the differences with our work in line 3, line4, line 5, line 11 and line 12 in the section 2 as well as our revised manuscript . > Thanks for the detailed responses! I highly appreciate it and some of my concerns are addressed. However, my main concern remains: the claims made in this paper are not well supported by the experiments and results provided, and there are not many new efforts in the revision to improve this. Therefore I'm keeping my score for this version of the paper. Thanks for the thoughtful comments, we also conducted additional reinforcement learning experiments. Results and comparsion with our MindAgent results are in Table 12 for supporting our claims. We would like to thank you for all your constructive feedback again. We addressed each of your comments in our revised manuscript. # Message to AC Dear Area Chairs, Senior Area Chairs and Program Chairs, Frist and foremost, we would like to express our sincere gratitude for providing the opportunity to address our concerns with you. We are writing regarding our submission, "MindAgent: Emergent Gaming Interaction" (Submission 3098). We greatly appreciate the insightful feedback from the reviewers, which have significantly contributed to enhancing the quality of our work. However, there is a lack of faithful feedback from reviewers based on our revisions. We have significantly updated our manuscripts and add additional experiment results to address each reviewer's concerns. We would like to highlight how we have addressed the key concerns raised by the reviewers, demonstrating our commitment to engaging with their feedback constructively: **Active Engagement with Reviewer Concerns:** We have thoroughly addressed each point raised by the reviewers in our responses, ensuring that we engage with their concerns directly and comprehensively. **Novelty and Complexity of MindAgent:** In response to concerns about the novelty of MindAgent, we have clarified its unique approach to **centralized coordination planning** and the rational for using centralized coordination planning in **gaming domain** and its handling of **complex spatial-temporal reasoning** challenges in multi-agent environments. **Environmental Complexity and Action Spaces:** We have provided additional information and comparative data to address concerns about the environment’s complexity and the large action space, underscoring the sophistication of the challenges our system addresses. **High-Level Planning Focus:** In response to queries about bypassing low-level control, we have emphasized our research's focus on high-level planning, clarifying the rationale behind our approach and its relevance to the research objectives. **Added Additional Details** We have enriched the manuscript with additional details, particularly in areas where reviewers sought further clarification. These enhancements include more in-depth explanations of our methodology, expanded results sections, and a more comprehensive discussion of the implications and limitations of our work. In addition, as per reviewer request, we added partial interaction log into the manuscript and we also added full interaction log to the comment section. We have meticulously responded to each specific question and feedback provided by the reviewers, incorporating these responses into our revised manuscript to ensure that all concerns are adequately addressed. We believe that our efforts to address their feedback have significantly strengthened our submission. We are hopeful that these amendments and clarifications meet the expectations of the review committee. We are open to further suggestions and guidance to improve our work; **however, there is a lack of interactions from reviewers**. **In particular**, we wish to draw attention to the lack of response from **Reviewer tSpT** and **Reviewer DUbr** to our comments and the revisions made in our manuscript. A large section of revisions were specifically aimed at addressing their misunderstandings and concerns. Their feedback, or lack thereof, has been a critical part of our revision process, and we hoped for a more interactive dialogue to ensure all issues were adequately addressed. Regarding **Reviewer YZDa**, while we appreciate the acknowledgment that some concerns have been addressed, the lack of detailed feedback on remaining issues or specific areas of improvement has been challenging. A more detailed response would greatly assist us in further refining our manuscript to meet the review committee's standards. We believe that addressing these points of engagement, or lack thereof, with the reviewers is crucial for a fair and constructive review process. We have done our utmost to address all concerns raised in the initial reviews and hope that our efforts are evident in the revised manuscript and the comments section. We truly value the chance to showcase our research at ICLR and wish that our insights can aid in enhancing the review procedure for upcoming submissions. We appreciate your attention to this matter. We eagerly await your response. Sincerely, Authors

    Import from clipboard

    Paste your markdown or webpage here...

    Advanced permission required

    Your current role can only read. Ask the system administrator to acquire write and comment permission.

    This team is disabled

    Sorry, this team is disabled. You can't edit this note.

    This note is locked

    Sorry, only owner can edit this note.

    Reach the limit

    Sorry, you've reached the max length this note can be.
    Please reduce the content or divide it to more notes, thank you!

    Import from Gist

    Import from Snippet

    or

    Export to Snippet

    Are you sure?

    Do you really want to delete this note?
    All users will lose their connection.

    Create a note from template

    Create a note from template

    Oops...
    This template has been removed or transferred.
    Upgrade
    All
    • All
    • Team
    No template.

    Create a template

    Upgrade

    Delete template

    Do you really want to delete this template?
    Turn this template into a regular note and keep its content, versions, and comments.

    This page need refresh

    You have an incompatible client version.
    Refresh to update.
    New version available!
    See releases notes here
    Refresh to enjoy new features.
    Your user state has changed.
    Refresh to load new user state.

    Sign in

    Forgot password

    or

    By clicking below, you agree to our terms of service.

    Sign in via Facebook Sign in via Twitter Sign in via GitHub Sign in via Dropbox Sign in with Wallet
    Wallet ( )
    Connect another wallet

    New to HackMD? Sign up

    Help

    • English
    • 中文
    • Français
    • Deutsch
    • 日本語
    • Español
    • Català
    • Ελληνικά
    • Português
    • italiano
    • Türkçe
    • Русский
    • Nederlands
    • hrvatski jezik
    • język polski
    • Українська
    • हिन्दी
    • svenska
    • Esperanto
    • dansk

    Documents

    Help & Tutorial

    How to use Book mode

    Slide Example

    API Docs

    Edit in VSCode

    Install browser extension

    Contacts

    Feedback

    Discord

    Send us email

    Resources

    Releases

    Pricing

    Blog

    Policy

    Terms

    Privacy

    Cheatsheet

    Syntax Example Reference
    # Header Header 基本排版
    - Unordered List
    • Unordered List
    1. Ordered List
    1. Ordered List
    - [ ] Todo List
    • Todo List
    > Blockquote
    Blockquote
    **Bold font** Bold font
    *Italics font* Italics font
    ~~Strikethrough~~ Strikethrough
    19^th^ 19th
    H~2~O H2O
    ++Inserted text++ Inserted text
    ==Marked text== Marked text
    [link text](https:// "title") Link
    ![image alt](https:// "title") Image
    `Code` Code 在筆記中貼入程式碼
    ```javascript
    var i = 0;
    ```
    var i = 0;
    :smile: :smile: Emoji list
    {%youtube youtube_id %} Externals
    $L^aT_eX$ LaTeX
    :::info
    This is a alert area.
    :::

    This is a alert area.

    Versions and GitHub Sync
    Get Full History Access

    • Edit version name
    • Delete

    revision author avatar     named on  

    More Less

    Note content is identical to the latest version.
    Compare
      Choose a version
      No search result
      Version not found
    Sign in to link this note to GitHub
    Learn more
    This note is not linked with GitHub
     

    Feedback

    Submission failed, please try again

    Thanks for your support.

    On a scale of 0-10, how likely is it that you would recommend HackMD to your friends, family or business associates?

    Please give us some advice and help us improve HackMD.

     

    Thanks for your feedback

    Remove version name

    Do you want to remove this version name and description?

    Transfer ownership

    Transfer to
      Warning: is a public team. If you transfer note to this team, everyone on the web can find and read this note.

        Link with GitHub

        Please authorize HackMD on GitHub
        • Please sign in to GitHub and install the HackMD app on your GitHub repo.
        • HackMD links with GitHub through a GitHub App. You can choose which repo to install our App.
        Learn more  Sign in to GitHub

        Push the note to GitHub Push to GitHub Pull a file from GitHub

          Authorize again
         

        Choose which file to push to

        Select repo
        Refresh Authorize more repos
        Select branch
        Select file
        Select branch
        Choose version(s) to push
        • Save a new version and push
        • Choose from existing versions
        Include title and tags
        Available push count

        Pull from GitHub

         
        File from GitHub
        File from HackMD

        GitHub Link Settings

        File linked

        Linked by
        File path
        Last synced branch
        Available push count

        Danger Zone

        Unlink
        You will no longer receive notification when GitHub file changes after unlink.

        Syncing

        Push failed

        Push successfully