Our mission is to help computational modelers develop, document, and share their computational models in accordance with community standards and good open science and software engineering practices. Model authors can publish their model source code in the Computational Model Library with narrative documentation as well as metadata that supports open science and emerging norms that facilitate software citation, computational reproducibility / frictionless reuse, and interoperability. Model authors can also request private peer review of their computational models. Models that pass peer review receive a DOI once published.
All users of models published in the library must cite model authors when they use and benefit from their code.
Please check out our model publishing tutorial and feel free to contact us if you have any questions or concerns about publishing your model(s) in the Computational Model Library.
We also maintain a curated database of over 7500 publications of agent-based and individual based models with detailed metadata on availability of code and bibliometric information on the landscape of ABM/IBM publications that we welcome you to explore.
Displaying 10 of 803 results for "Momme Von Sydow" clear search
The wisdom of the crowd refers to the phenomenon in which a group of individuals, each making independent decisions, can collectively arrive at highly accurate solutions—often more accurate than any individual within the group. This principle relies heavily on independence: if individual opinions are unbiased and uncorrelated, their errors tend to cancel out when averaged, reducing overall bias. However, in real-world social networks, individuals are often influenced by their neighbors, introducing correlations between decisions. Such social influence can amplify biases, disrupting the benefits of independent voting. This trade-off between independence and interdependence has striking parallels to ensemble learning methods in machine learning. Bagging (bootstrap aggregating) improves classification performance by combining independently trained weak learners, reducing bias. Boosting, on the other hand, explicitly introduces sequential dependence among learners, where each learner focuses on correcting the errors of its predecessors. This process can reinforce biases present in the data even if it reduces variance. Here, we introduce a new meta-algorithm, casting, which captures this biological and computational trade-off. Casting forms partially connected groups (“castes”) of weak learners that are internally linked through boosting, while the castes themselves remain independent and are aggregated using bagging. This creates a continuum between full independence (i.e., bagging) and full dependence (i.e., boosting). This method allows for the testing of model capabilities across values of the hyperparameter which controls connectedness. We specifically investigate classification tasks, but the method can be used for regression tasks as well. Ultimately, casting can provide insights for how real systems contend with classification problems.
In the model agents make decisions to contribute of not to the public good of a group, and cooperators may punish, at a cost, defectors. The model is based on group selection, and is used to understan
A system that receives from an agent-based social simulation the agent’s emotional data, their emotional-related data such as motivations and beliefs, as well as their location, and visualizes of all this information in a two dimensional map of the geographic region the agents inhabit as well as on graphs along the time dimension.
This is the final version of the model. To simulate the normative dynamics we used the EmIL (EMergence In the Loop) Framework which was kindly provided by Ulf Lotzmann. http://cfpm.org/EMIL-D5.1.pdf
Continuing on from the Adder model, this adaptation explores how rationality, learning and uncertainty influence the exploration of complex landscapes representing technological evolution.
An ABM to simulate the behaviour of households within a village and observe the emerging properties of the system in terms of food security. The model quantifies food availability, access, utilisation and stability.
We propose an agent-based model where a fixed finite population of tagged agents play iteratively the Nash demand game in a regular lattice. The model extends the bargaining model by Axtell, Epstein and Young.
A simple model that aims to demonstrate the influence of agri-environmental payments on land-use patterns in a virtual landscape. The landscape consists of grassland (which can be managed extensively or intensively) and a river. Agri-environmental payments are provided for extensive management of grassland. Additionally, there are boni for (a) extensive grassland in proximity of the river; and (b) clusters (“agglomerations”) of extensive grassland. The farmers, who own randomly distributed grassland patches, make decisions either on the basis of simple income maximization or they maximize only up to an income threshold beyond which they seize making changes in management. The resulting landscape pattern is evaluated by means of three simple models for (a) agricultural yield, (b) habitat/biodiversity and (c) water quality. The latter two correspond to the two boni. The model has been developed within a small project called Aligning Agent-Based Modelling with Multi-Objective Land-Use Allocation (ALABAMA).
This ABM re-implements and extends the simulation model of peer review described in Squazzoni & Gandelli (Squazzoni & Gandelli, 2013 - doi:10.18564/jasss.2128) (hereafter: ‘SG’). The SG model was originally developed for NetLogo and is also available in CoMSES at this link.
The purpose of the original SG model was to explore how different author and reviewer strategies would impact the outcome of a journal peer review system on an array of dimensions including peer review efficacy, efficiency and equality. In SG, reviewer evaluation consists of a continuous variable in the range [0,1], and this evaluation scale is the same for all reviewers. Our present extension to the SG model allows to explore the consequences of two more realistic assumptions on reviewer evaluation: (1) that the evaluation scale is discrete (e.g. like in a Likert scale); (2) that there may be differences among their interpretation of the grades of the evaluation scale (i.e. that the grade language is heterogeneous).
Ecosystems are among the most complex structures studied. They comprise elements that seem both stable and contingent. The stability of these systems depends on interactions among their evolutionary history, including the accidents of organisms moving through the landscape and microhabitats of the earth, and the biotic and abiotic conditions in which they occur. When ecosystems are stable, how is that achieved? Here we look at ecosystem stability through a computer simulation model that suggests that it may depend on what constrains the system and how those constraints are structured. Specifically, if the constraints found in an ecological community form a closed loop, that allows particular kinds of feedback may give structure to the ecosystem processes for a period of time. In this simulation model, we look at how evolutionary forces act in such a way these closed constraint loops may form. This may explain some kinds of ecosystem stability. This work will also be valuable to ecological theorists in understanding general ideas of stability in such systems.
Displaying 10 of 803 results for "Momme Von Sydow" clear search