Computational Model Library

Displaying 10 of 464 results for "Tim M Daw" clear search

This code simulates the WiFi user tracking system described in: Thron et al., “Design and Simulation of Sensor Networks for Tracking Wifi Users in Outdoor Urban Environments”. Testbenches used to create the figures in the paper are included.

WeDiG Sim

Reza Shamsaee | Published Monday, May 14, 2012 | Last modified Saturday, April 27, 2013

WeDiG Sim- Weighted Directed Graph Simulator - is an open source application that serves to simulate complex systems. WeDiG Sim reflects the behaviors of those complex systems that put stress on scale-free, weightedness, and directedness. It has been implemented based on “WeDiG model” that is newly presented in this domain. The WeDiG model can be seen as a generalized version of “Barabási-Albert (BA) model”. WeDiG not only deals with weighed directed systems, but also it can handle the […]

9 Maturity levels in Empirical Validation - An innovation diffusion example

Martin Rixin | Published Wednesday, October 19, 2011 | Last modified Saturday, April 27, 2013

Several taxonomies for empirical validation have been published. Our model integrates different methods to calibrate an innovation diffusion model, ranging from simple randomized input validation to complex calibration with the use of microdata.

Online Collaboration, Competing for Attention

M Manning | Published Wednesday, July 19, 2017 | Last modified Thursday, January 24, 2019

This is a model of a community of online communities. Using mechanisms such as win-stay, lose-shift, and preferential attachment the model can reproduce similar patterns to those of the Stack Exchange network.

The Li-BIM model aims at simulating the behavior of occupants in a building. It is structured around the numerical modeling of the building (IFC format) and a BDI cognitive architecture. The model has been implemented under the GAMA platform.

Organisms, Individuals and Organizations face the dilemma of exploration vs. exploitation
Identifying the optimal trade-off between the two is a challenge
Too much exploration (e.g. gaining new knowledge) can be detrimental to day-to-day survival and too much exploitation (applying existing knowledge) could be detrimental to long term survival esp. if conditions change over time

The purpose of the model is to investigate how the amount of resources acquired (wealth/success) is related to persistence with the strategy of local exploration under different resource distributions, availability of resources over time and cost of relocation

The largely dominant meritocratic paradigm of highly competitive Western cultures is rooted on the belief that success is due mainly, if not exclusively, to personal qualities such as talent, intelligence, skills, smartness, efforts, willfulness, hard work or risk taking. Sometimes, we are willing to admit that a certain degree of luck could also play a role in achieving significant material success. But, as a matter of fact, it is rather common to underestimate the importance of external forces in individual successful stories. It is very well known that intelligence (or, more in general, talent and personal qualities) exhibits a Gaussian distribution among the population, whereas the distribution of wealth - often considered a proxy of success - follows typically a power law (Pareto law), with a large majority of poor people and a very small number of billionaires. Such a discrepancy between a Normal distribution of inputs, with a typical scale (the average talent or intelligence), and the scale invariant distribution of outputs, suggests that some hidden ingredient is at work behind the scenes. In a recent paper, with the help of this very simple agent-based model realized with NetLogo, we suggest that such an ingredient is just randomness. In particular, we show that, if it is true that some degree of talent is necessary to be successful in life, almost never the most talented people reach the highest peaks of success, being overtaken by mediocre but sensibly luckier individuals. As to our knowledge, this counterintuitive result - although implicitly suggested between the lines in a vast literature - is quantified here for the first time. It sheds new light on the effectiveness of assessing merit on the basis of the reached level of success and underlines the risks of distributing excessive honors or resources to people who, at the end of the day, could have been simply luckier than others. With the help of this model, several policy hypotheses are also addressed and compared to show the most efficient strategies for public funding of research in order to improve meritocracy, diversity and innovation.

Peer reviewed Emergent Firms Model

J M Applegate | Published Friday, July 13, 2018

The Emergent Firm (EF) model is based on the premise that firms arise out of individuals choosing to work together to advantage themselves of the benefits of returns-to-scale and coordination. The Emergent Firm (EF) model is a new implementation and extension of Rob Axtell’s Endogenous Dynamics of Multi-Agent Firms model. Like the Axtell model, the EF model describes how economies, composed of firms, form and evolve out of the utility maximizing activity on the part of individual agents. The EF model includes a cash-in-advance constraint on agents changing employment, as well as a universal credit-creating lender to explore how costs and access to capital affect the emergent economy and its macroeconomic characteristics such as firm size distributions, wealth, debt, wages and productivity.

Peer reviewed General Housing Model

J M Applegate | Published Thursday, May 07, 2020

The General Housing Model demonstrates a basic housing market with bank lending, renters, owners and landlords. This model was developed as a base to which students contributed additional functions during Arizona State University’s 2020 Winter School: Agent-Based Modeling of Social-Ecological Systems.

Peer reviewed Virus Transmission with Super-spreaders

J M Applegate | Published Saturday, September 11, 2021

A curious aspect of the Covid-19 pandemic is the clustering of outbreaks. Evidence suggests that 80\% of people who contract the virus are infected by only 19% of infected individuals, and that the majority of infected individuals faile to infect another person. Thus, the dispersion of a contagion, $k$, may be of more use in understanding the spread of Covid-19 than the reproduction number, R0.

The Virus Transmission with Super-spreaders model, written in NetLogo, is an adaptation of the canonical Virus Transmission on a Network model and allows the exploration of various mitigation protocols such as testing and quarantines with both homogenous transmission and heterogenous transmission.

The model consists of a population of individuals arranged in a network, where both population and network degree are tunable. At the start of the simulation, a subset of the population is initially infected. As the model runs, infected individuals will infect neighboring susceptible individuals according to either homogenous or heterogenous transmission, where heterogenous transmission models super-spreaders. In this case, k is described as the percentage of super-spreaders in the population and the differing transmission rates for super-spreaders and non super-spreaders. Infected individuals either recover, at which point they become resistant to infection, or die. Testing regimes cause discovered infected individuals to quarantine for a period of time.

Displaying 10 of 464 results for "Tim M Daw" clear search

This website uses cookies and Google Analytics to help us track user engagement and improve our site. If you'd like to know more information about what data we collect and why, please see our data privacy policy. If you continue to use this site, you consent to our use of cookies.
Accept