Our mission is to help computational modelers at all levels engage in the establishment and adoption of community standards and good practices for developing and sharing computational models. Model authors can freely publish their model source code in the Computational Model Library alongside narrative documentation, open science metadata, and other emerging open science norms that facilitate software citation, reproducibility, interoperability, and reuse. Model authors can also request peer review of their computational models to receive a DOI.
All users of models published in the library must cite model authors when they use and benefit from their code.
Please check out our model publishing tutorial and contact us if you have any questions or concerns about publishing your model(s) in the Computational Model Library.
We also maintain a curated database of over 7500 publications of agent-based and individual based models with additional detailed metadata on availability of code and bibliometric information on the landscape of ABM/IBM publications that we welcome you to explore.
Displaying 4 of 4 results mental models clear search
We present a socio-epistemic model of science inspired by the existing literature on opinion dynamics. In this model, we embed the agents (or scientists) into social networks - e.g., we link those who work in the same institutions. And we place them into a regular lattice - each representing a unique mental model. Thus, the global environment describes networks of concepts connected based on their similarity. For instance, we may interpret the neighbor lattices as two equivalent models, except one does not include a causal path between two variables.
Agents interact with one another and move across the epistemic lattices. In other words, we allow the agents to explore or travel across the mental models. However, we constrain their movements based on absorptive capacity and cognitive coherence. Namely, in each round, an agent picks a focal point - e.g., one of their colleagues - and will move towards it. But the agents’ ability to move and speed depends on how far apart they are from the focal point - and if their new position is cognitive/logic consistent.
Therefore, we propose an analytical model that examines the connection between agents’ accumulated knowledge, social learning, and the span of attitudes towards mental models in an artificial society. While we rely on the example from the General Theory of Relativity renaissance, our goal is to observe what determines the creation and diffusion of mental models. We offer quantitative and inductive research, which collects data from an artificial environment to elaborate generalized theories about the evolution of science.
The model simulates seven agents engaging in collective action and inter-network social learning. The objective of the model is to demonstrate how mental models of agents can co-evolve through a complex relationship among factors influencing decision-making, such as access to knowledge and personal- and group-level constraints.
Decision-makers often have to act before critical times to avoid the collapse of ecosystems using knowledge \textcolor{red}{that can be incomplete or biased}. Adaptive management may help managers tackle such issues. However, because the knowledge infrastructure required for adaptive management may be mobilized in several ways, we study the quality and the quantity of knowledge provided by this knowledge infrastructure. In order to analyze the influence of mobilized knowledge, we study how the following typology of knowledge and its use may impact the safe operating space of exploited ecosystems: 1) knowledge of the past based on a time series distorted by measurement errors; 2) knowledge of the current systems’ dynamics based on the representativeness of the decision-makers’ mental models of the exploited ecosystem; 3) knowledge of future events based on decision-makers’ likelihood estimates of extreme events based on modeling infrastructure (models and experts to interpret them) they have at their disposal. We consider different adaptive management strategies of a general regulated exploited ecosystem model and we characterize the robustness of these strategies to biased knowledge. Our results show that even with significant mobilized knowledge and optimal strategies, imperfect knowledge may still shrink the safe operating space of the system leading to the collapse of the system. However, and perhaps more interestingly, we also show that in some cases imperfect knowledge may unexpectedly increase the safe operating space by suggesting cautious strategies.
The code enables to calculate the safe operating spaces of different managers in the case of biased and unbiased knowledge.
How does the world population adapt its policies on energy when it is confronted with a climate change? This model combines a climate-economy model with adaptive agents.