Quantcast
  • Register
PhysicsOverflow is a next-generation academic platform for physicists and astronomers, including a community peer review system and a postgraduate-level discussion forum analogous to MathOverflow.

Welcome to PhysicsOverflow! PhysicsOverflow is an open platform for community peer review and graduate-level Physics discussion.

Please help promote PhysicsOverflow ads elsewhere if you like it.

News

New printer friendly PO pages!

Migration to Bielefeld University was successful!

Please vote for this year's PhysicsOverflow ads!

Please do help out in categorising submissions. Submit a paper to PhysicsOverflow!

... see more

Tools for paper authors

Submit paper
Claim Paper Authorship

Tools for SE users

Search User
Reclaim SE Account
Request Account Merger
Nativise imported posts
Claim post (deleted users)
Import SE post

Users whose questions have been imported from Physics Stack Exchange, Theoretical Physics Stack Exchange, or any other Stack Exchange site are kindly requested to reclaim their account and not to register as a new user.

Public \(\beta\) tools

Report a bug with a feature
Request a new functionality
404 page design
Send feedback

Attributions

(propose a free ad)

Site Statistics

146 submissions , 123 unreviewed
3,961 questions , 1,408 unanswered
4,889 answers , 20,762 comments
1,470 users with positive rep
507 active unimported users
More ...

Is Stephen Wolfram's NKS, an attempt to explain the universe with cellular automata, in conflict with Bell's Theorem?

+ 9 like - 0 dislike
153 views

Stephen Wolfram's A New Kind of Science (NKS) hit the bookstores in 2002 with maximum hype. His thesis is that the laws of physics can be generated by various cellular automata--simple programs producing complexity. Occasionally (meaning rarely) I look at the NKS blog and look for any new applications. I see nothing I consider meaningful. Is anyone aware of any advances in any physics theory resulting from NKS? While CA are both interesting and fun (John Conway, Game of Life), as a theory of everything, I see problems. The generator rules are deterministic, and they are local in that each cell state depends on its immediate neighbors. So NKS is a local deterministic model of reality. Bell has shown that this cannot be. Can anyone conversant with CA comment?

This post imported from StackExchange Physics at 2014-05-14 18:28 (UCT), posted by SE-user Gordon
asked Jan 30, 2011 in Theoretical Physics by Gordon (400 points) [ no revision ]
Gerard 't Hooft has been looking at cellular automata inspired models for fundamental physics. You might find some of his recent (and readable) articles at arxiv.org/find/quant-ph/1/au:+Hooft_G

This post imported from StackExchange Physics at 2014-05-14 18:28 (UCT), posted by SE-user Siva

5 Answers

+ 8 like - 0 dislike

Wolfram's early work on cellular automata (CAs) has been useful in some didactical ways. The 1D CAs defined by Wolfram can be seen as minimalistic models For systems with many degrees of freedom and a thermodynamic limit. Insofar these CAs are based on a mixing discrete local dynamics, deterministic chaos results.

Apart from these didactical achievements, Wolfram's work on CAs has not resulted in anything tangible. This statement can be extended to a much broader group of CAs, and even holds for lattice gas automata (LGAs), dedicated CAs for hydrodynamic simulations. LGAs have never delivered on their initial promise of providing a method to simulate turbulence. A derivative system (Lattice Boltzmann - not a CA - has some applications in flow simulation).

It is against this background that NKS was released with much fanfare. Not surprisingly, reception by the scientific community has been negative. The book contains no new results (the result that the 'rule 110 CA' is Turing complete was proven years earlier by Wolfram's research assistant Matthew Cook), and has had zero impact on other fields of physics. I recently saw a pile of NKS copies for sale for less than $ 10 in my local Half Price Books store.

This post imported from StackExchange Physics at 2014-05-14 18:28 (UCT), posted by SE-user Johannes
answered Jan 30, 2011 by Johannes (280 points) [ no revision ]
Somehow I ended up with two of them sent to me by Amazon. They make good ballast for my boat.

This post imported from StackExchange Physics at 2014-05-14 18:28 (UCT), posted by SE-user Gordon
-1: 1D CA's do not result in deterministic chaos when they are computing, like 110, they result in complex structures that evolve. The lattice Boltzmann model you give is essentially a CA with random update rules, and it is used in hydrodynamic simulations. The book has a few "new" results (but these are mostly incorrect). It is most important as a summary of Wolfram's thinking.

This post imported from StackExchange Physics at 2014-05-14 18:28 (UCT), posted by SE-user Ron Maimon
Ron, if you are making a blanket statement that 1D automata can not lead to deterministic chaos, I wonder how you define the latter?

This post imported from StackExchange Physics at 2014-05-14 18:28 (UCT), posted by SE-user Johannes
@Johannes: (sorry for the downvote, I suppose this is not sufficiently self-explanatory). The definition I use for chaos is not Wolfram's, it is that the automaton radomizes. This means that if you take a finite size snapshot in a window of finite extent, you can compute the statistical distribution in that window to arbitrary accuracy without running the automaton at all, just with a fixed length computation which only depends on the accuracy, not on how long the automaton is run.

This post imported from StackExchange Physics at 2014-05-14 18:28 (UCT), posted by SE-user Ron Maimon
Ok, that confirms my suspicion. You might want to read some about chaos, Luyapunov coefficients and the like.

This post imported from StackExchange Physics at 2014-05-14 18:28 (UCT), posted by SE-user Johannes
@Johannes: I don't need to read anything--- I know what Lyapunov exponents are. There are chaotic automata, like 25, where the stuff randomizes, and computing automata, like 110, where the stuff is alive. The two are different. The 25 automata have nonlocal information flow analogous to Lyapunov exponents, while 110 is not analogous to any simple dynamical system. It's a full computer, it has no analogs other than other full computers.

This post imported from StackExchange Physics at 2014-05-14 18:28 (UCT), posted by SE-user Ron Maimon
@Gordon: I can't upvote you, but if I could, you would get +10 for your sense of humor which goes against mainstream Ayatollahs/inquisitional/communists/fascists/pro/con point of views. (hope, I have not hurt any side nor feelings).

This post imported from StackExchange Physics at 2014-05-14 18:28 (UCT), posted by SE-user Shaktyai
+ 8 like - 0 dislike

While NKS came out with much hype, and with a lot of skepticism from scientists, the scientific ideas there are not completely trivial. I just think they are not foundational for the science of physics (at least not as we know it so far), rather they are foundational for the science of biology.

The main discovery made by Wolfram (although with an important confusion which I will explain below, and with an extremely significant precurser in Coway's game of life) is that a simple 1 dimensional cellular automata whose rules are chosen at random will have a finite not so small probability of being a full computer (in Wolfram's system, 2 of the 128 possibilities). The proof that the system he found, rule 110 in his terminology, is actually a full computer only came two decades later, thanks to the pioneering work of Cook (working under Wolfram). But it justifies his focus on the system as central to science, since before, it was often implicitly assumed that to get a certain amount of complexity, you had to put in complexity by hand. This result is also present in Conway's system, but Wolfram's work is somewhat complementary, because the information flows in 1d systems make it more difficult to imagine a full computer emerging. The fact that it does anyway (although, as Cook's construction shows, with horrible running times, because of the difficulty of shuttling imformation long distances) is surprising and notable.

This is not so important for physics, because any attempt to model physics with cellular automata will have to be grossly nonlocal in order to avoid Bell's theorem. This is not so implausible today, given gravitational holography, but Wolfram suggested that there would be a direct correspondence between local elementary particle paths and automata structures, and these ideas are flat out impossible, and were ruled out before he proposed them, by Bells theorem. This means that the chapter of his book dealing with physics is completely wrong, and may be ignored.

But this work is important in a completely different way, it is the foundation of biology!

(EDIT: Chaitin's new book makes some brief comments about NKS which echo the main biological points below. I am not cribbing Chaitin, his book postdates this.)

Biology and Religion

The most puzzling aspect of the world we find ourselves in is that we are surrounded by complex computing devices not of our own design! Namely ourselves, other people, animals, plants, and bacteria. How did these computational structures get built, when we have to work pretty hard to make a computer? It seems that there is a puzzle here.

The puzzle has, in the past, been resolved by assuming some sort of magic put life on Earth, a supernatural agency. This idea is clearly at odds with the laws of nature as we understand them today, but it is important to keep in mind the superstitious answer, because elements of it are salvageable.

The superstitious answer is that God came down into the primordial soup, and mixed up the molecules to make life. The notion of God is not clearly defined in religious texts, where rigor is not the top priority. But I will try to give a positivistic definition below. I find that using this positivistic definition, which does not mention anything supernatural, I can translate the thoughts of religious people and make complete sense of what they are saying, where otherwise it just sounds like the ranting of delusional people suffering from severe brain-damage.

In order to discuss biology sensibly, I believe one must understand this religious point of view thoroughly, in a logical positivistic way, because it is important in biology to the same extent that it is completely unimportant in physics.

In a complex system, such as human social structures, we tend to observe patters which cannot be attributed solely to the actions of individual people. For example, the protestant reformation seem to have happened all at once, within the span of a few decades in the early 16th century, where Church reformers were active and working for centuries before, with very little success. What made it happen? It wasn't just Luther and Calvin, it was also a network of businessmen and bankers, and disenchanted Catholics. The discovery of America was important in some way, as was the expulsion of Jews from England. To my mind, the most important was the 14th century edict forbidding usury by Catholics, which prevented the formation of banking. But it clearly wasn't one cause, nor was it the work of one person working alone.

When we see such complex phenomena, it is reasonable to attribute them to the working of a larger intelligence than the intelligence of any individual, and this is the intelligence of the collective. Just as a person is a collection of neurons, not any one of which is responsible for her intelligence, the society is a collection of individuals, no one of which is responsible for everything the society does or thinks collectively. The collective pattern is in many ways smarter than the individual--- it contains collective memories, in traditions and conventions, which inform individual action in complex ways.

The notion of of god (lower case g, like Zeus, or Mars) in ancient cultures is the name given to the entities formed from collective human actions. They are nebulous, but important, because the decision to go to war cannot be attributed to any one person, but to an entity, the god of war, formed from many individuals working together with the aim of forming a coherent collective which will lead the society to make that phase-transition of behavior which is going to war. Identifying a notion of a god, and explcitly setting people working for this god, makes them aware of the fact that they are working as parts of a machine, not solely as individual actors. Further, it can inspire them to act without direct orders from a King, or a priest, just through their own introspection, so as to best achieve the goal.

The notion of god was refined somewhere in India or Iran into the notion of God (upper case G), from which the Brahma cults and Abrahamic religion, and Zoroastianism emerged. This notion suggests that the conflict between gods is similar to the conflict between individuals, the gods also make collectives, and some win and some lose. In the end, there is a notion of a supreme God, the God which is the limit of the collective of whatever gods survive, defined as infinitely high up the god heirarchy, and demanding ethical actions.

This limiting conception of God was considered so important by the ancient thinkers, that they let all thier other ideas die away in the midieval collapse, choosing to preserve only this through the middle ages.

But in addition to the practical notions of guiding behavior in collectives, the ancients also attributed all sorts of supernatural feats to God, including creating the universe and hand-designing life. These ideas about God are out of place with the conception as a meta-property of a complex system, and are completely contradicted by modern scientific discoveries. They are superfluous to religion, and detrimental to it, because they make people expect miracles and divine intervention in ways that violate the laws of nature, and such things just never happen.

The notion of God, as far as I have been able to make sense of it, is essentially a limiting computational conception--- it is the limit as time goes to infinity of the behavior of a complex system where the computational entities combine and grow in power into ever larger units. The idea of the limit suggests that there will be a coherence between the units at all levels, so that in the infinite time limit, for example, all societies will agree on the ethical course of action in a given circumstance, and will agree on how to organize their economies, and structure their interpersonal relations. These predictions are surprising, considering the divergence in human behavior, and yet, history suggests that such a convergence is slowly happening.

This computational decidability in the evolving limit has a direct counterpart in the idea that as mathematical systems become more complex, by reflection, they decide all arithmetical theorems. This is not a theorem, but an observation. It is noted that as we go up the tower of set theoretic reflection principles, more and more arithmetic theorems are resolved, and there is no in-principle limitation that suggests that the theorems will not all be decided by strong enough reflection. This is Paul Cohen's "Article of faith" in mathematical practice, and I will accept it without reservation.

Further, the article of faith tells you that we already have a name for the mathematical idea of God, it can be identified with the concept of the Church Kleene ordinal, the limit of all countable computable ordinals. Any computable formal system is only able to approach this ordinal gradually, and this ordinal is infinitely rich. If you have a description of this ordinal, you have a reflection principle which should be powerful enough to decide all theorems of arithmetic, to decide what consequences of any axiomatic system will be.

Because this ordinal has all the theological attributes religious folks attribute to God, in relation to pure mathematics, I consider it a sort of heresy to assume that there are larger ordinals. In particular, any notion of the first uncountable ordinal, or inaccessible ordinals, are only meaningful once they are placed in a given axiomatic system, and then they should collapse in the appropriate countable model to be less than the Church Kleene ordinal. This is not technically precise, but it gets the main idea across (it is easy to collapse the ordinals to be countable, but it is not so easy to rearrange the scheme to make them less than Church Kleene, but this is because within any deductive system which is of a set-theoretic nature, you can give a name to the Church Kleene ordinal, and define this ordinal plus 1, etc. These technical considerations are not so significant for the philosophical ideas)

So the interpretation I will take for religious doctrine is that God is to be identified with the Church Kleene ordinal, no higher ordinal is to be interpreted as actually higher, and gods will be identified with human collectives acting together to form a unit greater than the individuals. The monotheistic law of complex systems will state that all gods converge to the ideal represented by God over time, as they battle it out in a Darwinian struggle.

Automata and Darwin's experiment

When you have a cellular automata capable of universal computation, there is a strange phenomenon--- sub-parts of it are always in competition with itself. To explain this, one needs to look at Darwin's experiment, detailed in the Origin of Species.

Most of the Origin is theoretical, but Darwin did do one important experiment. He took a square plot of land, and carefully removed all visible living things from the soil. He uprooted all the plants, sifted to remove insects, and left the plot alone to see how it would be recolonized.

What he observed is that the plant species that recolonized the plot were first of the fast-growing unstable variety, that a whole bunch of weeds and bugs spread over the new area. Then, over time, other more hardy species slowly took over from the weeds, until, many months later, the plot was indistinguishable from the remaining land in the lot.

The purpose of the experiment was to see whether there is an actual struggle for resources in nature. Darwin hypothesized that if nature is in constant struggle, different elements, which are more successful but slowly replicating, will only win out after a time over elements which are less hardy, but whose strategy is quick colonization of new territory. His observations were consistent with the idea that the living things in any area are continually struggling for primacy, and that the limitation is the finite resources in any given plot of land.

This idea can be tested in computing cellular automata. By zeroing out a square patch in a 2d cellular automata which looks stable, one can see whether the remaining data colonizes the space in a way that is uniform, or in a gradually transforming way. I did this experiment using an 8-bit cellular automata (256 values) with random rules, and I found that in many cases, those cases which are complex, the colonization is in stages, much as in Darwin's plot of land. The stages are short-lived, perhaps reflecting the limited computation possible in a small region with 8-bit values. It would be interesting to repeat the experiment using arbitrarily large integers on each cell, which can be thought of as representing a complex polymer, which catalyzes transformations on its neighbors

But the inhomogenous colonization suggests that once you have a computing cellular automata, there is a constant competition between parts of the automata, which make collective computations, for resources. In other words, that Darwin's struggle is begun.

To make this idea more precise, consider dividing a CA in two, by placing a wall between the left and right half, and not allowing the halves to interact. If the CA is truly computational and complex, the two halves will not come to a statistical equilibrium, but will have complex structures on either side which acquire new characteristics at random over time, as its subparts evolve.

If you now remove the wall, it is unlikely that the left half will have compatible characteristics with the right half. They will not be able to mix. So in this case, the two halves must battle for domination, and whichever half wins will impose its characteristics on the other half fillling the whole space with cells which are compatible with its characteristics. These characteristics include typical CA "animals" or structures which are qualitatively similar in their relations, particular configurations which are only stable in the environment of other structures around them. It is difficult to extract these characteristics from a running simulation, because you don't know a-priori what to look for, but I am confident that it can be done.

This type of thing implies that there is continuous competition in a CA which appears the moment it is first seeded, and continues as long as it is operating. In this environment, Darwinian selection and evolution are possible even without any explicit self-replicating structure. Any self-replication is of very high-level qualitative traits, not of low level bit structures.

Replication and evolution

This point of view is different from the most usual point of view regarding evolution (which is not the one originally proposed by Darwin). The usual point of view is the modern-synthesis evolution, which suggests that evolution proceeds by copying bit-strings in molecules, with errors, and that the result is that optimized bit strings are eventually selected.

This point of view is extremely poor in modelling actual biological evolution. First, nothing you are familiar with actually replicated itself. People have sex, bacteria share genes, and crossing over is complicated on non-genetic sequences, it is only a simple shuffling on genes.

Further, mutations seem to be produced by shadowy internal mechanism directed by complex RNA networks in egg cells and in testicles. They are not random copying errors. To assume that the biological world is produced by a process of copying with error, coupled with selection is as silly as the following parable suggests:

Many years ago, there was only one book. It was a cookbook, with detailed instructions on how to make macaroni and cheese. The book was copied by scribes, who made an error here, an omitted passage there, and these books then competed for attention. Some recipes were improved by the errors, others became unreadable. Eventually, the books grew in length, with new passages produced by accidental duplicated copying, until today, behold! The library of Congress!

This story is ridiculous. But it is this ridiculous story that is currently sold as dogma in the biological sciences.

It is my view that any realistic theory of evolution must be closer to Darwin than to the modern synthesis. It must take into account that the process of mutation is authorly, it proceeeds by complex RNA editing of DNA sequences. It must take into account the idea that sexual selection is primary, so that mate-selection is the dominant driving force of evolution in sexual species. It must also take into account the idea that the competition begins well before replication, and requires nothing more than a computing CA.

There is support for this position from computer experiments on self-replicating evolution. In order to test natural selection, little chunks of code were allowed to replicated and self-modify in the 1970s-1980s, to see what the end result would be. The end result was that the programs modified themselves until they found the shortest fastest self-replicator, which then filled up the computer memory.

At the time, this was considered a positive sign, the programs had evolved. But the obvious stasis in the final state leads me to see this as death of a complex system. There is no further progress possible from the end state, without an external agent to kick things around. The result is not a complex system, but a system trapped in a stable equilbrium of parasitic fast replication. Far from being a model of life, it is a model of a self-replicating cancer killing all evolution.

CA properties: Wolfram's Annoying Error

Wolfram classified Cellular automata into four types

  1. homogenous end state
  2. Simple periodic structures, perhaps separated, with different periods
  3. Self-similar ("Chaotic") structures
  4. Complex structures

Type 1 are automata that die. These just have a single stable endpoint that you always reach. Type 2 have infinitely many endpoints, but they are as simple to describe as a classical integrable motion--- you just have cycles of certain types, and to specify the endpoint, you give a list of all the cycles, and where you are in the cycle, and this specifies the result of running the CA from a given initial condition. These first two types of automata obviously will not reproduce a general purpose computer.

Type 3 are those automata that lead to self-similar fractal structures, like the Sierpinsky gasket. These are more complex, so that the end-state requires an actual computation to specify, and wolfram identifies these with classical chaotic motions. I think this identification is wrong, but this is what it is.

The Type 4 are the complex automata, where you have to actually run them in full to figure out what they do. I don't like the final category, so I will now give my personal classification.

  1. homogenous end state
  2. simple periodic end states, perhaps separated with different periods
  3. Self-similar or statistically self-similar fractal structures
  4. random automata, chaotic stable endpoint, stat mech.
  5. Complex automata, biology.

The class 3 is expanded slightly, and class 4 is divided in two. There are random automata, which act to produce a randomized collection of values which wander ergodically through the allowed value space, and class 5, those automata which produce true complex behavior, with a way to map a computer into them with a map of map of reasonable complexity, which can actually be described by a finite procedure.

Because Wolfram doesn't distin

answered Jan 19, 2012 by Ron Maimon (7,535 points) [ no revision ]
holy tldr, Batman!

This post imported from StackExchange Physics at 2014-05-14 18:28 (UCT), posted by SE-user user2963
Don't worry about it, it's only the secret and meaning of life. It might go beyond the scope of the question a bit, but I'm travelling tomorrow, so I thought it might be good to share, ya know, in case the plane crashes.

This post imported from StackExchange Physics at 2014-05-14 18:28 (UCT), posted by SE-user Ron Maimon
Actually I did read a good chunk and I'll probably read the rest at some point - interesting stuff as usual but it seems like a lot is only very tangentially related to the question :)

This post imported from StackExchange Physics at 2014-05-14 18:28 (UCT), posted by SE-user user2963
@zephyr: I expected plenty of downvotes for that, but I wanted to explain this stuff, and this was the closest thing to a question about NKS.

This post imported from StackExchange Physics at 2014-05-14 18:28 (UCT), posted by SE-user Ron Maimon
@ronmaimon very interesting, read once and will have to read again and again some other day! Thanks +1!

This post imported from StackExchange Physics at 2014-05-14 18:28 (UCT), posted by SE-user FrankH
+ 6 like - 0 dislike

Shortly after NKS came out, I wrote a review in which I tried to explain why the answer to your excellent question is yes. A deterministic model like Wolfram's can't possibly reproduce the Bell inequality violations, for fundamental reasons, without violating Wolfram's own rule of "causal invariance" (which basically means that the evolution of a CA shouldn't depend on the order in which updates are applied to spatially-distant regions). Even with some "long-range threads" in the cellular automaton (which Wolfram explicitly allows, after noticing the Bell issue), you still can't get causal invariance, unless the actual states of the automaton are probabilistic or quantum. A closely-related observation was later dubbed the "Free-Will Theorem" by John Conway and Simon Kochen.

This post imported from StackExchange Physics at 2014-05-14 18:28 (UCT), posted by SE-user Scott Aaronson
answered Aug 21, 2012 by ScottAaronson (795 points) [ no revision ]
+ 5 like - 0 dislike

Most of these automata models are deterministic in the same sense as pseudorandom number generators are. For example in the lattice gas models the deterministic rules end up generating noise and large scale fluctuations in accord to the Navier-Stokes equations (including turbulence, although this is computationally impractical because of the large lattice dimensions required for reducing the lattice viscosity). The lattice gas game turned in the late eighties from noisy discrete particle automata to smooth distribution based lattice Boltzmann mesoscopic scale continuous-value automata (see Guy R. McNamara and Gianluigi Zanetti, Use of the Boltzmann Equation to Simulate Lattice-Gas Automata, Phys. Rev. Lett. 61, 2332–2335 (1988) ), so that's where you find most relevant advances these days.

This post imported from StackExchange Physics at 2014-05-14 18:28 (UCT), posted by SE-user Janne808
answered Jan 30, 2011 by Janne808 (50 points) [ no revision ]
+ 0 like - 0 dislike

Murray Gell-Mann has an interesting take on Bell's theorem which pertains directly to Stephen Wolfram's thesis on modeling physical laws with cellular automata in his tome: 'A New Kind of Science', an analysis which took him over 20 years to complete.

According to Murray, elegant models of physics involve fundamental laws in addition to the results of random chance outcomes of a number of things which are non-deterministic in a quantum mechanical sense (he is referring to physical constants). Indeed, it is hard to imagine Wolfram's cellular automata on any scale determining the fundamentals of a theory like quantum chromodynamics, which has been fine tuned and/or renormalized at every step to assure that the theory works as closely as possible to the way nature does. It is dubious, to say the least, that cellular automata would be able to reproduce even a portion of this iterative process in a manner that would output anything other than utterly useless simulations with no relationship to what happens in the natural world.

One thing that Stephen predicted in NKS that does seem to be happening in a big way is the idea that science is increasingly dependent on big computing in order to get results that advance our knowledge of the universe. The LHC in Geneva is a case in point.

This post imported from StackExchange Physics at 2014-05-14 18:28 (UCT), posted by SE-user user41670
answered Mar 6, 2014 by user41670 (-20 points) [ no revision ]

Your answer

Please use answers only to (at least partly) answer questions. To comment, discuss, or ask for clarification, leave a comment instead.
To mask links under text, please type your text, highlight it, and click the "link" button. You can then enter your link URL.
Please consult the FAQ for as to how to format your post.
This is the answer box; if you want to write a comment instead, please use the 'add comment' button.
Live preview (may slow down editor)   Preview
Your name to display (optional):
Privacy: Your email address will only be used for sending these notifications.
Anti-spam verification:
If you are a human please identify the position of the character covered by the symbol $\varnothing$ in the following word:
p$\varnothing$ysicsOverflow
Then drag the red bullet below over the corresponding character of our banner. When you drop it there, the bullet changes to green (on slow internet connections after a few seconds).
To avoid this verification in future, please log in or register.




user contributions licensed under cc by-sa 3.0 with attribution required

Your rights
...