A few things the Symbol Research team are reading. Complex Insight is curated by Phillip Trotter (www.linkedin.com/in/phillip-trotter) from Symbol Research
Browse the GTC conference catalog of sessions, talks, workshops, and more.
Phillip Trotter's insight:
The good folks at NVIDIA have made the recent 324 GTC conference sessions freely available. With topics like Omniverse, Bots, Digital Twins, Autonomous Vehicles and extensive applications of AI , visualization and simulation from experts around the world. They are worth watching and investigating. Enjoy (and thanks NVIDIA!!)
Two simulations are executed with two different sets of parameters. It highlights the movement of goods over the Seine axis territory and also, how distinct ...
Phillip Trotter's insight:
Agent based simulastion of logistics is becoming increasingly commmon - nice example.
A couple months back, Mark Buchannan wrote an article in which he argued that ABMs might be a productive way of trying to understand the economy. In fact, he went a bit further – he said that ABMs...
Phillip Trotter's insight:
Been a while since we published anything on economics but with over 20 years experience building agent based models - we are suckers for articles that encourage people to discover them as thinking tools. Good explanation on how agent based models fit into economic forecasting.
The modern world is complex beyond human understanding and control. The science of complex systems aims to find new ways of thinking about the many interconnected networks of interaction that defy traditional ...
Social systems are among the most complex known. This poses particular problems for those who wish to understand them. The complexity often makes analytic approaches infeasible and natural language approaches inadequate for relating intricate cause and effect. However, individual- and agent-based computational approaches hold out the possibility of new and deeper understanding of such systems.
Simulating Social Complexity examines all aspects of using agent- or individual-based simulation. This approach represents systems as individual elements having each their own set of differing states and internal processes. The interactions between elements in the simulation represent interactions in the target systems. What makes these elements "social" is that they are usefully interpretable as interacting elements of an observed society. In this, the focus is on human society, but can be extended to include social animals or artificial agents where such work enhances our understanding of human society.
This handbook is intended to help in the process of maturation of this new field. It brings together, through the collaborative effort of many leading researchers, summaries of the best thinking and practice in this area and constitutes a reference point for standards against which future methodological advances are judged.
My preferred job title is 'theorist', but that is often too ambiguous in casual and non-academic conversation, so I often settle for 'computer scientist'. Unfortunately, it seems that the overwhelming majority of people equate ...
Phillip Trotter's insight:
Artem Kaznatcheev, a researcher in theoretical computer science - i.e. the ideas that underpin computing - has a wonderful write up of Stephanie Forrest's Stannislaw Ulam lecture at the SFI on using inspiration from Biology to address challenges in Software industry. The Ulam lecture is available in video - but its a few hours long - through seriously worth watching and covers modern developments in genetic programming and other approaches. If you need an abbrieviated write up of the key ideas underpinning the Professor Forrest's lecture - then Artem's write up is an awesomely succinct. Worth reading (and the lectures linked in his article - are worth watching!)
ECAL 2013, the twelfth European Conference on Artificial Life, presents the current state of the art of a mature and autonomous discipline collocated at the intersection of a theoretical perspective (the scientific explanations of different levels of life organizations, e.g., molecules, compartments, cells, tissues, organs, organisms, societies, collective and social phenomena) and advanced technological applications (bio-inspired algorithms and techniques to building-up concrete solutions such as in robotics, data analysis, search engines, gaming).
Advances in Artificial Life, ECAL 2013
Proceedings of the Twelfth European Conference on the Synthesis and Simulation of Living Systems
Edited by Pietro Liò, Orazio Miglino, Giuseppe Nicosia, Stefano Nolfi and Mario Pavone
I have a big soft spot for artificial life research - partly because i was a young researcher shortly after Chris Langton coined the term and a lot of my early hacking was around games of life, vants and cellular automata but also because over the years I have found many of the techniques discussed in ALIFE circles applicable to other fields such as machine learning, control architectures, and emergent simulation etc so this is definitely one for the reading list.
Pietro Liò is Reader in Computational Biology at the University of Cambridge and a member of the Artificial Intelligence group of the University's Computer Laboratory. He researches on Predictive models in Personalized medicine and Multiscale modelling of molecules-cell-tissue-organ interactions.
Orazio Miglino is a full Professor of Psychology at University of Naples Federico II where he leads the Natural and Artificial Cognition Lab. He is also an associate researcher at the Institute of Cognitive Sciences and Technologies of Italian National Research Council (ISTC-CNR) in Rome.
Giuseppe Nicosia is an Associate Professor in Computational Systems and Synthetic Biology in the Dept. of Mathematics and Computer Science of the University of Catania, Italy. His research activities focus on the design of biological systems, neuroinformatics, system design, design automation, optimization, solar cells, circuit and semiconductor design.
Stefano Nolfi is Research Director at the Italian National Research Council (CNR), director of the Laboratory of Autonomous Robots and Artificial Life of the Institute of Cognitive Sciences and Technologies. His research activities focus on the evolution and development of behavioural and cognitive skills in natural and artificial embodied agents (robots).
Mario Pavone is an Assistant Professor in computer science at the Department of Mathematics and Computer Science of the University of Catania. He is co-founder of TaoScience Research center, and he is also a member of the EURO association (The Association of European Operational Research Societies)
Commercial aviation is feasible thanks to the complex socio-technical air transportation system, which involves interactions between human operators, technical systems, and procedures. In view of the expected growth in commercial aviation, significant changes in this socio-technical system are in development both in the USA and Europe. Such a complex socio-technical system may generate various types of emergent behavior, which may range from simple emergence, through weak emergence, up to strong emergence. The purpose of this paper is to demonstrate that agent-based modeling and simulation allows identifying changed and novel rare emergent behavior in this complex socio-technical system. (...)
Agent-based modeling and simulation of emergent behavior in air transportation Bouarfa S, Blom HA, Curran R, Everdij MH Complex Adaptive Systems Modeling 2013, 1:15 (15 August 2013)
GE Healthcare is pushing a system called Corvix for doing agent-based simulations on complex problems. In India, the technology simulated a population of 80 million people in order to determine the best places to build medical facilities.
Around the world, the health care system is rife with inefficiencies, and General Electric thinks it can help solve the problem using data. Only it’s not talking about bureaucrats looking at reports: GE has built an artificial intelligence system called Corvix that uses historical data to predict the future, including everything from how diseases will spread to the cities where hospitals will be needed the most.
It might sound futuristic, but the techniques behind Corvix have actually been around for a while. The platform uses agent-based modeling to build, essentially, a reasonable facsimile of some sort of complex system and then simulate its evolution over time. The “agents” represent the atomic units of those systems, such as individual people in the case of human populations or perhaps cells in the case of a biological simulation. They act according to a set of rules in any given situation, which is how the models are able to keep the simulations progressing.
However, thanks to the advent of big data, GE Healthcare Chief Economist Mitch Higashi thinks the time is right for a platform like Corvix to provide some real value to real-world decisions. There’s enough raw computing power, machine intelligence and data-modeling expertise to start doing fast, accurate simulations over very large and complicated datasets. Also, advances in user-interface design have made these types of models more consumable: GE’s Corvix uses a game-like UI “that any 10-year-old can figure out how to use in 10 minutes,” Higashi said.
The first live run for Corvix happened in the state of Andhra Pradesh in India, where the system simulated a population of 80 million people in order to figure out where to build hospitals and medical training centers over the coming years. The GE team used two census datasets and one health care survey in order to build what Higashi calls “a reasonable representation of 80 million people,” as well as a map of India’s existing hospital and energy grid. Health care analysts studying the problem of where to build can drag a new hospital over an area on the map and see how the situation plays out, Higashi explained.
The original plan, said Chaitanya Sarawate, GE’s head of health economics and reimbursement for India, was for the Public Health Foundation of India to invest $2 billion building training institutions in different cities over the next five years. Corvix suggested some possible changes in location of those institutions, including placing two institutions in the country’s most-populous state, Uttar Pradesh, instead of just one as was originally planned. The advice is part of a report from the foundation to India’s Minstry of Health, which will make the ultimate decision.
Developing countries such as India are great places to use this type of technology, Higashi explained, because they are doing greenfield investing in areas such as health infrastructure and a lot of good can happen if they get it right off the bat. The problem, Sarawate noted, is that they often lack detailed data that can help governments make objective comparisons — that’s the kind of stuff a company like GE, in this case, can track down and try to feed into a model that takes into account its relative importance.
In a general sense, a Markov Network Brain (MNB) implements a probabilistic finite state machine, and as such is a Hidden Markov Model (HMM). MNBs act as controllers and decision makers for agents ...
Phillip Trotter's insight:
If like me, you are old enough to remember the animals to animats proceedings from the early 1990's which detailed early researchon agent based modeling, reinforcement learning algorithms and autonomous robots using neural networks, genetic algorithms and other probabilistic finite state machines as control architectures this will be of interest. If you are not -try and find a copy and read up - since a lot of current research is based on early ideas presented in those proceedings. The Adamilab have produced a stable implementation and platform for hidden markov model based controllers for agent based models and robotics. Code is available on Github and the Markov Network Brains article gives a good overview of why its of interest and underlying reasoning behind the implementation for anyone working on agent based simulation and autonomous robot and sensor platforms.
Businesses are fighting hard to withstand the effects of natural and man-made disasters and the significant disruptions both can have on supply chains, and while attempts to model the risks has run into some obstacles, a solution may be near.
Phillip Trotter's insight:
We have been researching large scale agent based simulation models for serveral years. One of our aims has been to be able to simulate systems models to understand pivot points and risk mitigation. For the last 18 months our internal research has been focused on working out how to do this in the context of catastrophe and lifescience models. It was rather delightful to see our thoughts and research echoed by Joseph Calandro, Managing Director of PricewaterhouseCooper’s insurance practice. If you want a better understanding of where systems simulations and agent based models are going to make a major impact - Joe's article on contingent business risk is a great place to start. Very much worth reading.
Good post on modeling human responses in agent based simulations. In addition to papers suggested here - there is a fair amount of game development AI that also falls into related studies that should be added to this list. Click on image or title to learn more.
For one thing, complex systems do not easily lend themselves to analysis, the process of taking apart a system and examining its components individually. If taken apart, many complex systems lose precisely the character that makes them complex. The essence of these systems, then, seems to lie not in the nature of their components but in how the components interact—across different hierarchies, in synergistic and antagonistic manners. The agents within these systems are heterogeneous (think participants in a market economy or molecules within a cell), and their behavior is influenced by the type and quantity of other agents nearby. Such systems defy description with the traditional tool of theory builders: mathematics. Instead, they must be modeled by taking into account the rules of interaction, the natures of the agents, and the way the agents, rules, and ultimately whole systems came about. In his Signals and Boundaries: Building Blocks for Complex Adaptive Systems, John Holland proposes that computational modeling is the appropriate tool not only for describing but, fundamentally, for understanding such systems. In particular, he argues that this modeling approach is in no way inferior to a mathematical one. Rather, he advocates that the computational modeling of signal-boundary systems (which I will describe in more detail below) goes where mathematics cannot go while being no less rigorous, no less exact.
Patients do not access physicians at random but rather via naturally emerging networks of patient flows between them. As retirements, mass quarantines and absence due to sickness during pandemics, or other shocks thin out these networks, the system might be pushed closer to a tipping point where...
Phillip Trotter's insight:
A very interesting ABM based approach for a stress-testing framework that could enable health authorities to rapidly identify bottlenecks in access to care and assess a healthcare networks resilience in the face of emergent situations. Well worth reading and investigating further.
Accessing information efficiently is vital for animals to make the optimal decisions, and it is particularly important when they are facing predators. Yet until now, very few quantitative conclusions have been drawn about the information dynamics in the interaction between animals due to the lack of appropriate theoretic measures. Here, we employ transfer entropy (TE), a new information-theoretic and model-free measure, to explore the information dynamics in the interaction between a predator and a prey fish. We conduct experiments in which a predator and a prey fish are confined in separate parts of an arena, but can communicate with each other visually and tactilely. TE is calculated on the pair’s coarse-grained state of the trajectories. We find that the prey’s TE is generally significantly bigger than the predator’s during trials, which indicates that the dominant information is transmitted from predator to prey. We then demonstrate that the direction of information flow is irrelevant to the parameters used in the coarse-grained procedures. We further calculate the prey’s TE at different distances between it and the predator. The resulted figure shows that there is a high plateau in the mid-range of the distance and that drops quickly at both the near and the far ends. This result reflects that there is a sensitive space zone where the prey is highly vigilant of the predator’s position.
Information Dynamics in the Interaction between a Prey and a Predator Fish Feng Hu, Li-Juan Nie and Shi-Jian Fu
Our empirical analysis demonstrates that in the chosen network data sets, nodes which had a high Closeness Centrality also had a high Eccentricity Centrality. Likewise high Degree Centrality also correlated closely with a high Eigenvector Centrality. Whereas Betweenness Centrality varied according to network topology and did not demonstrate any noticeable pattern. In terms of identification of key nodes, we discovered that as compared with other centrality measures, Eigenvector and Eccentricity Centralities were better able to identify important nodes.
The main aim of the Symposium is to facilitate the meeting of people working on different topics in different fields (mainly Economics, Finance and Computer Science) in order to encourage a structured multi-disciplinary approach to social sciences. Presentations and keynote sessions center around multi-agent modelling, from the viewpoint of both applications and computer-based tools. The event is also open to methodological surveys.
The event will be hosted by Social Simulation 2014, the 10th Conference of the European Social Simulation Association at the Universitat Autonoma de Barcelona, Barcelona, Spain. September 1-5th, 2014.
Understanding how to simulate economic and social systems will be critical in future planning and analysis tools- The Social Simulation 2014 conference will be a key event.
OptionsCity Launches Freeway Analytics Trading-Simulation Platform www.waterstechnology.com Electronic options trading technology vendor OptionsCity Software has launched Freeway Analytics, a visualization engine and an exchange simulator based on...
Phillip Trotter's insight:
As more financial simulation trading platforms ermerge it will be interesting to see how economics researchers incorporate, contrast, benchmark them to validate trading platforms and validate economic modeling of trading versus other modeling approaches.
BBC, Ford and Microsoft use its service and it is essential for drivers all across the world. So who exactly is INRIX?
Phillip Trotter's insight:
As part of our research in agent based modeling we have spent some time experimenting with traffic modeling in different platforms such as Mason, Gamma and MatSim as well as our own in house tools. This research led us to look at data sources and toolsets around the same time that Waze was starting to get tech blog press in the United States. Outside of California - we didnt hear much about Waze - the company we kept running into was INRIX. As we spoke with researchers and transport planning departments, traffic forecastor and traffic and transport industry experts, INRIX inevitably got mentioned. In fact if you have listened to traffic forecast in Europe - its odds on that you have heard someone using INRIX data. Yet many people have not heard of the company or the types of products they offer. INRIX are a real traffic big data success story and as a company they have an awesome singular mission - 'solve traffic worldwide'. If you are undertaking city planning, traffic and transport research you need to know who they are, the products they offer and how they may be able to help you. If your not - then just check out their traffic apps they will help you get from A to B safely and efficiently! Either way to get you started this is a great introductory article on what the company does.
The International Advanced Studies in Demography (IDEM) program is currently accepting applications for the upcomingIDEM winter semester 2013/14 at the Max Planck Institute for Demographic Research (MPIDR) in Rostock, Germany. The IDEM winter program includes an advanced course on Agent-based Modeling and Simulation, seehttp://tinyurl.com/idem-abm Other IDEM winter courses cover the Demography of Aging across the Tree of Life, Spatial Demography, and Measuring and Modeling Population Processes. Courses are scheduled from early November 2013 to the end of January 2014. Candidates may apply for one or several courses. There is no tuition fee. However, students are expected to pay their own transportation and living costs. A limited number of scholarships may be awarded to outstanding candidates or to those applicants who demonstrate financial need.The application deadline is 18 September 2013.
The result is now clear to just about everyone on the planet. The smartest guys in the room were no match for terabytes of data and smart algorithms. There is no one “theory of the case” anymore, but thousands of them, being run constantly. The point isn’t to be right, but to become less wrong over time.
Eric Bonabeau's team at Icosystems have helped pioneer commercial application of agent based modelling applications. It's nice to see them get some recognition in the Forbes article. However the article itself would have been much stronger if it had opened its scope and mentioned other teams such as Argonne National Labs (Repast), George Mason University (Mason), Uri Wilensky and the CCL team at NorthWestern (NetLogo). Technical University of Berlin/Federal Institute Zurich and Senozon (Matsim) amongst many others. All these teams have built platforms that are slowly but surely moving agent based simulation from academic theory to important real world applications and adoption. Nevertheless - good article and nice to see this being covered by the wider press.
When the Thomas Edison was asked about success amidst failure, he said that “If I find 10,000 ways something won’t work, I haven’t failed. I am not discouraged, because every wrong attempt discarded is another step forward.”
If Thomas Edison was alive today he would say that innovation is 1% inspiration and 99% simulation (Photo credit: Wikipedia)
With that kind of dedication, it’s no wonder that Edison was awarded over 1000 patents, including the light bulb, the phonograph and the motion picture camera, making him one of the most prolific inventors in history.
It also becomes clear why he regarded success as “1% inspiration and 99% perspiration.” Failing 10,000 times is a physical and mental undertaking that far exceeds most people’s endurance. Today, however, a new breed of innovators are outsourcing failure to computer simulations and it’s changing everything we thought we knew about business strategy.
This paper provides a logical framework for complexity economics. Complexity economics builds from the proposition that the economy is not necessarily in equilibrium: economic agents (firms, consumers, investors) constantly change their actions and strategies in response to the outcome they mutually create. This further changes the outcome, which requires them to adjust afresh. Agents thus live in a world where their beliefs and strategies are constantly being “tested” for survival within an outcome or “ecology” these beliefs and strategies together create. Economics has largely avoided this nonequilibrium view in the past, but if we allow it, we see patterns or phenomena not visible to equilibrium analysis. These emerge probabilistically, last for some time and dissipate, and they correspond to complex structures in other fields. We also see the economy not as something given and existing but forming from a constantly developing set of technological innovations, institutions, and arrangements that draw forth further innovations, institutions and arrangements.(...)
Complexity Economics: A Different Framework for Economic Thought W. Brian Arthur SFI WP 13-04-012
Brian Arthur was an early pioneer of applying concepts of complex systems to economic systems and its good to see an update publication that builds on his earlier work and other work in this area. Certainly worth reading.
If you've read Waldrop's account of the development of the complexity paradigm at the Sante Fe Institute (Waldrop, M, (1992) Complexity: The Emerging Science at the Edge of Chaos, Simon & Schuster, New York), the name Brian Arthur will be familiar.
Recently, melanoma has become the most malignant and commonly occurring skin cancer. Melanoma is not only the major source (75%) of deaths related to skin cancer, but also it is hard to be treated by the conventional drugs.
Phillip Trotter's insight:
This is the first time a 3D multi-scale agent-based cancer model was employed to describe he communication between the melanoma, the vasculature around the tumor and drug reaction. The research gives insights into the underpinning mechanisms of tumor growth and potential treatments to prevent this. Most importantly the findings will help provide a foundation to develop predictable in silico cancer models which echo in vitro findings incorporating realistic biological and physical data and features. Worth reading.
The school is intended for postdocs, lecturers and predocs with a background in computer science (artificial intelligence) or computational linguistics (corpus linguistics or natural language processing) and a strong interest in music and the origins of language. There will be background lectures that introduce concepts from biology, anthropology, psychology, music theory and linguistics that are helpful to understand the nature of creativity, the role and intimate relations between language and music, and the mechanisms underlying cultural evolution. It contains technical lectures on the fundamental computational components required for language processing and technical ateliers to learn how to set up evolutionary linguistics experiments. Participants have the opportunity to present their latest research in a poster session. The school also features artistic ateliers in which participants create new creative works and engage in performance.
Music and the Origins of Language International Summer School on Agent-based Computational Models of Creativity, 15 – 20 September 2013 in Cortona (Italy)
the GisAgents blog is always a great read. Nice link to new algorithm for population synthesis that fuses remote-sensing data with partial and sparse demographic surveys. The algorithm addresses non-binding constraints and complex sampling designs by translating population synthesis into a computationally efficient procedure for constrained network growth. Click on image or title to learn more.
To get content containing either thought or leadership enter:
To get content containing both thought and leadership enter:
To get content containing the expression thought leadership enter:
You can enter several keywords and you can refine them whenever you want. Our suggestion engine uses more signals but entering a few keywords here will rapidly give you great content to curate.
The good folks at NVIDIA have made the recent 324 GTC conference sessions freely available. With topics like Omniverse, Bots, Digital Twins, Autonomous Vehicles and extensive applications of AI , visualization and simulation from experts around the world. They are worth watching and investigating. Enjoy (and thanks NVIDIA!!)