He will give a speech in the Royal Gallery in the UK Parliament, attended by various MPs, Lords and Ladies, on October 23.
On October 17, the Newsletter of the UK Royal Economic Society contains this comment by me on Brexit. See my earlier weblog entries on Brexit.
Alex read the comment in the Newsleter, and called me that he wanted to speak to me.
I spoke Alex yesterday over dinner of wildbraad in a fine restaurant in Leiden. He confessed to me that he seriously considers calling in sick and cancelling the state visit to the UK. He is experiencing increasingly negative feelings about the trip, and he had read my explanation that negative numbers reflect the process of cancelling.
Stroking his belly on the good meal, he was looking for consistency.
“Like for example,” he said, “on the 24th we will have a visit to two ships, to mark the importance of seafare for our two nations. But Max and I will travel to the UK by air. So why don’t we travel by ship ? When people would expose me as being a total hypocrite then I can’t blame them.”
Brexit is wrenching him. He hates seeing the UK leaving the EU, or even the messy way how they are doing so. He doesn’t like the sight of a Queen of a proud independent nation making fun of him as a EU-shackled fake-king. Perhaps when he wakes up in Buckingham Palace he might find his shoelaces tied together. He feels much tempted to follow the example of William III (1650-1702) of sailing from Scheveningen to Britain with an armada of 250 ships and 35,000 troops and take the island, pound some sense into the natives, and have another Glorious Revolution.
“They call themselves Anglo-Saxons but to a high degree they are Vikings, just like the Russians. Vikings need an occasional bang to their heads to reset them to common sense again,” he explained. He also reminded me that he himself had a Romanov princess as an ancestor, see his claim on taking Russia too.
Thus I asked whether he was actually inviting me to bang his head now too, given his own Viking genes. He nodded. We were in Leiden because he had wanted to reread his MA thesis in history, that is kept there in the vaults of Leiden University. He handed me his thesis and I duly banged it to his head. To our regret it hardly had an impact. Alex’s head was too hard or the sheet of paper too thin.
“But I can’t conquer Britain because Max doesn’t want to live there. The Brits have something against Argentines. You know: the Falklands.”
He continued: “My comments on Brexit already have caused a stir. Prime Minister Mark Rutte now worries that my speech in the Royal Gallery will be boycotted by British parliamentarians. Oh my, oh my, a boycott is such an effective method to make someone look silly, especially when your shoelaces are tied together.”
“So this is why you wanted to talk to me,” I inferred. “You want me to smooth out some ruffled feathers so that you will still have an audience that will be interested in what you would have to say, and so that they know in advance that they will not be offended or lectured to.”
“Exactly,” he nodded. “Prime Minister Mark Rutte and his loony Foreign Secretary Stef Blok suggested that I talk to you, and ask you to draft a speech for me, and to put it on your website so that people in the UK have some advance information that it will be a highly relevant speech for the UK. So, please, make me look good.”
“If you pay for the meal,” I granted.
This got me a groan and pained look. He was in such a tight spot that he actually footed the bill. Thank you, Brexiteers.
Thus, below is my draft for the speech.
– – – D R A F T – – –
Dear friends, come together here in the Royal Gallery on the beautiful day of October 23 2018.
I am the Head of State of Holland. I am speaking for the Dutch government. The following does not contain my personal view, but it contains the view by the Dutch government, including me. This goes to show that one can still have a useful function even when there are some constraints. Also the UK can have a useful function even when there would be constraints by the EU.
[ rumble ]
You all know that I also went to school in Wales and that I developed a taste for Monty Python. Let me start with some jokes, so that you can all loosen up and relax before we become serious again.
[ sounds of horror ]
The first joke is about a boss who hires a new employee. At lunch on the first day, the employee complains at the boss: “There is something wrong with the wheelbarrow. It goes “screetch … screech … screech”.” [ spoken with long pauses ] The boss replies: “Well, you can go home now and don’t return. It should go “screetch screetch screetch”.” [ spoken fast ]
[ some grins and laughter ]
This joke makes me think about the speed of ships. For very long in human history, seafaring was the most efficient way of transport. Civilisation developed on the shores of rivers and seas. Many traces of this development have been lost because of the rise of the sea level since the last Ice Age. The British Isles were known in the Bronze Age for their tin to make bronze, and transport then was also by sea. The British Isles have a long history of settlers and invaders from the sea, like the Romans, Angles, Saxons, Vikings, and William the Conqueror in 1066, to name a few. In 1688, William III of Orange boated from Holland with 250 carriers and 35,000 troops to cause the Glorious Revolution of William and Mary. It can be said that the British Isles are a center of European civilisation just like the shores of the Baltic, the North Sea, the Atlantic and the Mediterranean. The Dutch government is amazed that some people on the British Isles do not know about this history, and refer to “The Continent” as if Britain would be somewhat distant from Europe itself. Well, let me invite you to keep ships in mind and to see Britain embedded in Europe. Nowadays, planes can be faster, which is why I traveled here not by boat. Yet ships still cannot be beaten for bulk transport, and we should keep this in mind, also for our mutual borders, and also for the Political Economy for the Caribbean.
[ pause ]
The second joke is from the days of the USSR, actually only 30 years ago, still in our lifetimes. The joke may appeal to Jeremy Corbyn who is scheduled to meet with me today unless he would snake out to avoid me.
The joke concerns some factory for trucks and work appliances in the old USSR. At the end of the working day workers leaving the factory are checked on smuggling or stealing stuff. Each day there is also a worker, Vladimir, who pushes a wheelbarrow with scrap metal through the exit gate. Screetch, screetch, screetch. The guards are suspicious, check him and his load thoroughly, find only worthless stuff, and let him pass. One day the man turns 65, retires, and leaves the factory without his wheelbarrow. “Vladimir,” says the gatekeeper, “we will see each other for the last time today. Would you please tell me whether you have been smuggling something or not?”
[ pause ]
Vladimir confesses with a smile: “Every day for forty years, I have taken a wheelbarrow from the factory.”
[ some grins and laughter ]
The joke illustrates the importance of observation. For mathematics or philosophy you might make an abstract assumption and forget about reality, but science requires careful observation on reality.
This joke brings us to a key difference between Holland and the United Kingdom. We are both democratic nations, but have different concepts of democracy, which shows that the term “democracy” is a tricky term that we must use with care. The Dutch House of Commons is elected with equal proportional representation (EPR) and the UK House of Commons is elected with district representation (DR). Also the words “election” and “representation” are the same but mean entirely different things here.
The Dutch system provides for proper election and representation. The UK system of Single Member Districts with First Past The Post provides for contests rather than proper election. It discards votes and neglects the condition of representation of the interest of the electorate. A majority of voters in the UK have “taxation without representation”. The electoral system in the UK is in direct violation with article 21 of the Universal Declaration of Human Rights that was adopted by the United Nations in 1948.
The Dutch government has consulted with scientists in Holland and has come to the conclusion that there is something amiss with the academic field of so-called “political science on electoral systems”. Scholars in this field have neatly checked the content of the wheelbarrow but have failed to see the very wheelbarrow itself. They claim that the field would be a science but these scholars lack a training in empirical science. The field still is in the humanities in which indoctrination into tradition is more important than empirical observation and testing.
Though this field of study also fails in Holland, the effect of failure in Holland is limited, since we have equal proportional representation (EPR) and not district representation (PR).
For the UK, the effect is much larger. We can only understand Brexit by diagnosing that the UK voters have been disinformed for decades. The UK has a proto-democracy and no proper democracy yet. For this reason, I currently come to you carrying gifts, namely the news about this new discovery that “political science on electoral systems” is no real science, but it is pseudo-science, comparable to astrology, alchemy or homeopathy.
How this blind spot developed is an issue for history. At this moment it is important to bring this message to you. The Dutch government regards this news as a game changer for the present discussion in the UK on Brexit. Namely, the UK can now ask its empirical scientists, from physics to psychometrics, to help clarify the situation of systematic disinformation that has been created by this pseudo-science of “political science on electoral systems”.
The Dutch government suggests that the UK switches to equal proportional representation (EPR), say adopt the Dutch system of open lists (in which you may always vote for a regional candidate though people don’t tend to do so), has proper elections, and then let the new House of Commons discuss the relation with the European Union again. It is not unlikely that the EU would allow the UK the time for such a fundamental reconsideration on both its democracy and Brexit.
It remains to be seen whether the UK would want to switch from DR to EPR, but the first step would be to provide the public with proper information. If the UK would stick with DR and Brexit, then at least we will know that such a choice has been made by an informed electorate, while presently we can only diagnose that the UK electorate lacks this information. Obviously, logic requires that you first switch to EPR for the proper procedure to remain with DR.
[ mix of moans and laughter ]
Another topic.
Let me quote historian Helene Von Bismarck about the Single European Act of 1986:
“Thatcher’s dedication to the idea of a completed Single Market was such that she made considerable concessions during the negotiations to bring it about. (…) The Single European Act had enormous ramifications for the way the European Community was to be run. No only did it provide for the completion of the Single Market, it also simplified the decision-making in the European Council by extending the principle of qualified majority voting to a large number of issues that until then had been decided unanimously.”
When the people in the UK have the feeling that they are losing power to the European Union, then they should rather look at Margaret Thatcher, who accepted the abolition of veto power. She also rushed the agreement through Parliament instead of allowing for deliberation.
It is well-known that the Dutch people are also hesitant about the evolution of the European Union. Rather than seeing the UK leave the EU, Holland would like to have the UK as an ally, for example on raising the requirements on the qualified vote.
Another topic.
English has become the lingua franca of the world. Billions of people in distant lands learn English to partake in the development of our planet.
When people migrate to Europe, then many have learned English, and then many may have a preference to migrate to the UK since this is where English is spoken on a daily base. We have a mixed situation of course. Parts of Africa use French and those migrants may prefer to go to France. In Eastern Europe and Turkey, German tends to be important too. Nevertheless, we can understand that the UK might have additional migration because of the plain fact that English has become the lingua franca.
To my understanding there is no scientific research on this aspect yet, and it has not been a factor in current policy making yet. The Dutch government urges the UK to reconsider Brexit and to remain in the EU and to help develop sensible policies that also include such considerations.
Let us also be sensible about alternative developments. If the UK would leave the EU, as it now plans to do, then it would still be a magnet to those who have learned English as the lingua franca. Since ships can also reach the UK directly – and I have highlighted the ease of shipping to you – then the horrible scenes that we now see in the Mediterrranean are merely shifted to similar scenes around the British Isles, with a great cost for the UK to guard its coasts and to transport illegal immigrants back to their lands of origin. We only need to be aware of the forecasts about the rise in the world population and the issue of climate change.
It is the firm belief of the Dutch government that democracy requires deliberation, and that we all in the same boat here.
Other points.
[ … additional points by Mark Rutte and Stef Blok ….]
Thus, my friends in this Royal Gallery, let me round up this speech on behalf of the Dutch government. Our nations are closely linked by geography and history. Let us continue with increased awareness of this elementary fact. Thank you for your attention.
– – – E N D D R A F T – – –
Acknowledgement: Jokes taken from here.
]]>The UK has been fundamentally disinformed about its electoral system with district representation and the use of referenda like the Brexit Referendum of 2016. The UK is locked in tradition and fuzzy thinking in the humanities. The situation may be explained by the historical path that the UK has taken, but this history hasn’t included a proper application of science to the notion of democracy.
Compare the current chaos w.r.t. Brexit to the chaos with the financial crisis of 2008. On the latter, the UK Queen asked famously:
“Why did nobody notice it?”
There is a longer list of economists who issued warnings in time, with Hyman Minsky at the top and me somewhere too. The next question rather is why such warnings weren’t taken seriously in the policy making process. My diagnosis since 1990 is that there is a failure of the separation of powers, the Trias Politica, with still too much room for politicians to manipulate the information. The remedy is to create an Economic Supreme Court (ESC) that will guard the quality of information for policy. The House of Commons would still determine policy but it would get less room to disinform the public. The current UK Office for Budget Responsibility (OBR) is a far cry from what is actually needed.
With this analogy established, consider Brexit again. Might the Queen not repeat the question ? Now however there doesn’t seem to be a list of early warnings that were overlooked. Now we have a “political science” that has gotten lost in abstraction. Here, the remedy is to ask proper scientists from physics to biology to psychometrics to econometrics to look at democracy and to help “political science / studies on electoral systems” become a proper science too. My suggestion is to team up empirical scientists from the Royal Society with members of the PSA and the British Academy, and to encourage a buddy-system to start delving into this. The place to start is my paper “One woman, one vote. Though not in the USA, UK and France” at MPRA 2018, and a presentation 1270381 at Zenodo.org on the distance between votes and seats.
Many people think that the Brexit Referendum of 2016 allowed voters to express their decision, with 52% Leave and 48% Remain. However, not all voters expressed their decision but many were only guessing. A YouGov poll at the time of the GIGO 2017 showed that 17% of voters still listed Remain between different options of Leave. Voters were forced to make a strategic choice about what they feared most what might happen. See my deconstruction of this mess in the October 2017 Newsletter of the Royal Economic Society (RES).
Now there are calls for a second referendum. This call wants to resolve the current chaos by creating more chaos, and potentially a “stab in the back” myth that the 2016 supposed decision isn’t listened to. The lesson from the current chaos should rather be that referenda are generally dumb and dangerous, even in the form of the neverendum. The real problem lies in the UK system of district representation that structurally fails to reflect the views and interests of voters. The deeper problem is that the House of Commons and the electorate are disinformed by an academic field that still is comparable to astrology, alchemy or homeopathy. There is a grand scale of disinformation by famous UK scholars like Iain McLean, John Curtice, younger Alan Renwick, and (other) members of PSA.
My suggestion is that the UK switches to equal proportional representation (EPR), say adopt the Dutch system of open lists (in which you may always vote for a regional candidate though people don’t tend to do so), has proper elections, and then let the new House of Commons discuss the relation with the EU again. It is not unlikely that the EU would allow the UK the time for such a fundamental reconsideration on both its democracy and Brexit. UK political parties may need to split up to offer voters the relevant spectrum of views, though one must allow for election alliances (especially the former Dutch method of list combinations). To some readers this suggestion may remind of earlier discussions about district or proportional representation (DR vs EPR). However, there now is the key new insight about the disinformation by the “political science / studies on electoral systems”, that causes the need to re-evaluate what has been claimed in the past by the academic ivory towers, and also by the disinforming UK Electoral Reform Society (ERS). It remains to be seen whether the UK would want to switch from DR to EPR, but the first step would be to provide the public with proper information.
PS. An eye-opener can be that “political science on electoral systems” relies upon common language instead of developed definitions. Physics also borrowed common words like “force” and “mass”, yet it provided precise definitions, and gravity in Holland has the same meaning as gravity in the UK. The “political science on electoral systems” uses the same word “election” but an “election” in Holland with EPR is entirely different from an “election” in the UK with DR. In reality there is a difference between a contest (DR) or a bundling of votes to support a representative (EPR). We find that the UK is locked into confusion by its vocabulary. An analogy is the following. Consider the medieval trial by combat or the “judgement of God”, that persisted into the phenomenon of dueling to settle conflicts. A duel was once seriously seen as befitting of the words “judgement” and “trial”. Eventually civilisation gave the application of law with procedures in court. Using the same words “judgement” and “trial” for both a duel and a court decision confuses what is really involved, though the outward appearance may look the same, that only one party passes the gate. The UK suffers the same kind of confusion about the “General Election for the House of Commons” when this actually is no proper election of interest representatives but concerns contests for getting district winners. The system of DR is proto-democratic and no proper democracy that uses EPR.
Picture: Wikimedia Queen in the UK, Duel in France, Judges in The Hague.
]]>This is my farewell letter to the board of NKWP.
]]>I started looking into Brexit after the surprise referendum outcome in 2016. This memo sums up my findings over the last two years. The following identifies where the educated community in the UK is in need of re-educating themselves.
Earlier in 1990-1994 I already concluded that Montesquieu’s model of the separation of powers of the Trias Politica fails in a key aspect since its conception in 1748. Democracies need the fourth power of an Economic Supreme Court, see (2014). It is necessary to mention this earlier conclusion that predates Brexit, but let us now continue with findings following Brexit.
To start with: What does the UK electorate really want w.r.t. Brexit or Bremain ? Both the Referendum of 2016 and the General Election of 2017 do not provide adequate information. One would think that it is rather damning for a claimed democracy when its procedures do not result into adequate clarity on such a fundamental issue.
The 2016 Referendum Question concerned the legal issue of Leave or Remain but was disinformative about the ways of Leaving or Remaining. The political parties that are elected into the House of Commons are split on both direction and ways as well. The overall situation can only be described as chaotic. One might try to characterise this more positively as that a population with divided views generated a House of Commons with divided views, which would be democracy itself, but this neglects that there is no information about what those divided views actually are. The true process is “garbage in, garbage out” and this doesn’t fit the definition of democracy.
The very Brexit or Bremain Referendum Question fails the criteria for a decent statistical enquiry. I am surprised that the Royal Statistical Society (RSS) did not protest. The question of Leave or Remain is a binary legal issue but the true issue are the policy options. It took some time to analyse this, but with the help of Anthony Wells of YouGov.com I managed to dissect this, see (2017abc). Some 17 per cent of voters ranked Remain between different versions of Leave, which implies a grand game of guessing what to vote for, and which means that the Referendum failed on its purpose of expression of preferences. The UK Electoral Commission missed this but it does not care about this and is happy to take the legal position. They claim to provide proper information to the general public, but what they regard as “information” is regarded by statistical science as disinformation (but the RSS is silent on this). One is reminded of Byzantium instead of claimed modernity.
The main question is why the UK had the referendum in the first place. In Holland since 1917 there is system of equal proportional representation (EPR) for the House of Commons so that referenda are not required. The UK has a system of district representation (DR) that lacks such proportionality, and that invites the confusion that referenda might be used to find out what the electorate really thinks. The latter is a confusion indeed, since it neglects the important role of bargaining, see (2017c).
This diagnosis set me on the course of investigating why the USA, UK and France have DR and not EPR. My original thought was that a party that won an election would have no reason to change a system that caused its election. This would explain why the USA, UK and France were stuck with DR and did not switch to EPR. Last year I discovered that the true cause is different. My finding for the UK is that there is an amazing blindness in the UK intellectual community. The report in (2018a) causes a chill down the spine. It appears that “political science on electoral systems” is no science yet, but still solidly within the Humanities, and alike astrology, alchemy and homeopathy. The eye-opener is that these academics use the same word “election” for both DR and EPR while they actually have entirely different meanings. In reality only EPR has proper elections fitting of proper democracy. The DR system is a proto-democracy that relies on contests. Political “science” is blind to what this means not only for proper scientific analysis but also for communication with the general public. Voters are disinformed on a grand scale, both in the textbooks in government classes and in public discussion e.g. at “election” nights.
Compare physics that also borrowed words from colloquial English, like “force” and “mass”. Yet in physics these words have recieved precise meaning. In physics, gravity in Holland has the same meaning as gravity in the UK. Political “science” uses colloquial terms like “election” and “democracy” but those meanings are not fixed. An “election” in Holland with EPR is entirely different from an “election” in the UK with DR. Political “science” thus uses terms that confuse both the academics and the public. When historians describe how the West developed into democracy, they occlude the fact that the USA, UK and France are still in a proto-democratic phase.
A first complication is: There appears to be a special role for the UK Electoral Reform Society (ERS) founded in 1884 and originally known as the Proportional Representation Society. Here we find an independent and disinterested group that criticises DR and that claims to further the UK on the historical path towards EPR. However, it appears that ERS wants a transferable vote, while their claim that transferability generates proportionality is simply false. Such distortion contributed to the debacle of the 2011 Referendum on the “alternative vote”, which is a counterproductive construct to start with. When one presents the ERS with this criticism then the reply appears to be disingenuous. Instead of a clear adoption of EPR, either in the Dutch version or like the UK elections for the EU Parliament, with their wealth of experience by actual application, one can only conclude that the ERS is addicted to this notion of a transferable vote, and they want this model at any cost. Psychology might explain how such zealotism may arise but it remains far removed from proper information for the general public.
A second complication is: There appears to exist a confusion w.r.t. the interpretation of Arrow’s Impossibility Theorem on democracy. In this, there is a major role for mathematicians who mainly look at models and who neglect empirical science. This leads too far for this memo, and an overview is given in (2018e).
A third complication is: There is the interference by a grand coalition of statistics and political science (with some ambiguity whether quotation marks should be used) in creating a black hole on democracy and its measurement, see (2018bcd). Political science never managed to find a good measure for the difference between vote shares and seat shares. My proposal is to use the “sine-diagonal inequality / disproportionality” (SDID) measure, that does for democracy what the Richter scale does for earthquakes. Political science has shown less understanding of statistics, or perhaps failed in finding such a measure because statistical science did not develop this theory or did not understand what the political scientists were looking for. This hole has been plugged now, see (2018bcd). Nevertheless, this diagnosis calls for a reorganisation of university courses in statistics and political science.
The enclosed graph highlights the “perfect storm” of blindness of the intellectual community that lurks behind Brexit. The figure is documented in (2018d). The main idea is that statistics and other sciences like physics, biology, psychometrics and econometrics could help “political science on electoral systems” to become a proper science. Then science can provide adequate information to the general public.
A conclusion is: The UK electoral system has “winner take all” district representation (DR) that does not provide for equal proportional representation (EPR) of what voters want. Again the word “representation” means something else for proto-democratic DR versus democratic EPR. My suggestion is that the UK switches to EPR, say adopt the Dutch system of open lists, has new elections, and let the new House discuss Brexit or Bregret again. Bregret is defined by that the House adopted Brexit before and thus might reconsider. It is not unlikely that the EU would allow the UK the time for such a fundamental reconsideration on both electoral system and Brexit.
It remains to be seen whether the UK electorate would want to stick to the current system of DR or rather switch to EPR. The first step is to provide the UK electorate with adequate information. For this, the UK intellectual community must get its act together on what this information would be. A suggestion is to check the analysis that I have provided here.
Colignatus (2014), “An economic supreme court”, RES Newsletter issue 167, October, pp. 20-21
Colignatus (2017a), “Voting theory and the Brexit referendum question”, RES Newsletter, Issue 177, April, pp. 14-16
Colignatus (2017b), “Great Britain’s June 2017 preferences on Brexit options”, RES Newsletter, Issue 177, October, http://www.res.org.uk/view/art2Oct17Features.html
Colignatus (2017c), “Dealing with Denial: Cause and Cure of Brexit”, https://boycottholland.wordpress.com/2017/12/01/dealing-with-denial-cause-and-cure-of-brexit/
Colignatus (2018a), “One woman, one vote. Though not in the USA, UK and France”, https://mpra.ub.uni-muenchen.de/84482/
Colignatus (2018b), “Comparing votes and seats with cosine, sine and sign, with attention for the slope and enhanced sensitivity to inequality / disproportionality”, https://mpra.ub.uni-muenchen.de/84469/
Colignatus, (2018c), “An overview of the elementary statistics of correlation, R-Squared, cosine, sine, Xur, Yur, and regression through the origin, with application to votes and seats for parliament ”, https://doi.org/10.5281/zenodo.1227328
Colignatus, (2018d), “An overview of the elementary statistics of correlation, R-Squared, cosine, sine, Xur, Yur, and regression through the origin, with application to votes and seats for parliament (sheets)”, Presentation at the annual meeting of Dutch and Flemish political science, Leiden June 7-8, https://zenodo.org/record/1270381
Colignatus, (2018e), “The solution to Arrow’s difficulty in social choice (sheets)”, Second presentation at the annual meeting of Dutch and Flemish political science, Leiden June 7-8, https://zenodo.org/record/1269392
The CCSS refer to the Trends in International Mathematics and Science Study (TIMSS) (wikipedia).
TIMSS is made by the International Association for the Evaluation of Educational Achievement (IEA) (wikipedia). It so happens that IEA has its headquarters in Amsterdam but the link to Holland is only historical.
I am wondering whether CCSS and TIMSS adequately deal with the redesign of mathematics education.
There are conditions under which TIMSS is invalid.
There are conditions under which TIMSS is incomplete.
See my letter to IEA (makers of TIMSS) and NGA Center and CCSSO (users of TIMSS, makers of CCSS).
]]>Economic theory needs a stronger defence against unwise application of mathematics. Mathematicians are trained for abstract thought and not for empirical science. Their contribution can wreak havoc, for example in education with real life pupils and students, in finance by neglecting real world risks that contribute to a world crisis, or in voting theory where they don’t understand democracy.
Nowadays, though, I am also wary of students from the Humanities who rely upon legal views (their version of mathematics) instead of empirical understanding.
For the following, distinguish single seat elections (president, prime minister) and multiple seats elections (parliament). There is also a key distinction between Equal Proportional Representation (EPR) with proper elections and District Representation (DR) that has contests rather than proper elections.
(1) Montesquieu’s Trias Politica of the separation of powers is failing, and we need the separation of a fourth power, an Economic Supreme Court, based upon science, with a position in the constitution at the same level as the Executive, Legislative and Judiciary. The current setup allows too much room for politicians to manipulate the information for policy making. This need for separation can also be proven logically in a model using stylised facts, see the book DRGTPE. A short discussion on the 2007+ European crisis is here.
(2) Kenneth Arrow in his Impossibility Theorem has a correct deduction (there is an impossibility) but a wrong interpretation. He confuses voting and deciding. For this debunking of Arrow’s Theorem, see Chapter 9.2 of Voting Theory for Democracy (p239-251). Sheets of a presentation in June 2018 are here.
(3) A voting method that many might find interesting is the Borda Fixed Point method. See the counterfactual example of selecting a Prime Minister for Holland.
(4) Political science on electoral systems is no science yet but still locked in the Humanities, and comparable to astrology, alchemy and homeopathy. People in the USA, UK and France still have taxation without representation.
(4a) The key paper is One woman, one vote. Though not in the USA, UK and France.
(4b) A supportive paper develops the SDID distance measure for votes and seats.
(4c) This paper reviews the role of statistics for the latter measure. Sheets of a presentation in June 2018 are here.
(4d) An earlier comparison of Holland and the UK in 2010 (update 2015) contains a major stepping stone, but is not as critical as (4a). This analysis resulted in a short paper for Mathematics Teaching 222 (May 2011) at the time of the UK referendum on Alternative Vote.
(5) There are some supplementary findings, that I do not regard as major, but as roads that you might need to walk in order to discover that they do not lead far.
(5a) There are Two conditions for the application of Lorenz curve and Gini coefficient to voting and allocated seats. The Lorenz curve is a neat way to graphically show the disproportionality and inequality of votes and seats. The Gini is its associated measure. However, above measure SDID is to be preferred, since it is symmetric and doesn’t require sorting, has a relation to the R-squared and the Weber-Fechner law.
(5b) We can compare votes and seats but also use a policy distance. A crucial question is who determines the distance between policies ? When we have a distance, how do we process it ? I am not convinced by the method, but a discussion is here.
(5c) The Aitchison geometry might present a challenge to SDID. This paper provides an evaluation and finds this geometry less relevant for votes and seats. Votes and seats satisfy only two of seven criteria for application of the Aitchison distance.
(5d) This paper tries to understand the approach by Nicolaus Tideman and compares it with the distinction between voting and deciding.
(5e) Mathematician Markus Schulze was asked to review VTFD but did not check his draft review with me, which caused needless confusion, see here and here. PM. Schulze now has this 2017 paper, but doesn’t refer to Borda Fixed Point, perhaps thinking that he understands it, but he apparently is not open to the diagnosis that his “review” is no proper review.
For the above, it is pleasant that a distinction can be made between key results and findings about dead ends. I listed my debunking of Arrow’s Theorem as a key result, but it also identifies this theorem as a dead end. Thus, it is also a matter of perspective. When you are at the dead end, and turn around, the whole road is open again.
PM. Earlier weblog entries on democracy are here.
]]>I regard mathematics by computer as a two-sided coin, that involves both human thought (supported by tools) and what technically happens within a computer. The computer language (software) is the interface between the human mind and the hardware with the flow of electrons, photons or whatever (I am no physicist). We might hold that thought is more fundamental, but this is of little consequence, since we still need consistency that 1+1 = 2 in math also is 1+1 = 2 in the computer, and properly interfaced by the language that would have 1+1 = 2 too. The clearest expression of mathematics by computer is in “computer algebra” languages, that understand what this revolution for mankind is about, and which were developed for the explicit support of doing mathematics by computer.
The makers of Mathematica (WRI) might be conceptually moving to regarding computation itself as a more fundamental notion than mathematics or the recognition and handling of patterns. Perhaps in their view there would be no such two-sided coin. The brain might be just computation, the computer would obviously be computation, and the language is only a translator of such computations. The idea that we are mainly interested in the structured products of the brain could be less relevant.
Stephen Wolfram by origin is a physicist and the name “Mathematica” comes from Newton’s book and not from “mathematics” itself, though Newton made that reference. Stephen Wolfram obviously has a long involvement with cellular automata, culminating in his New Kind of Science. Wolfram (2013) distinguishes Mathematica as a computer program from the language that the program uses and is partially written in. Eventually he settled for the term “Wolfram language” for the computer language that he and WRI use, like “English” is the language used by the people in England (codified by their committees on the use of the English language).
My inclination however was to regard “Mathematica” primarily as the name of the language that happened to be evaluated by the program of the same name. I compared Mathematica to Algol and Fortran. I found Wolfram’s Addison-Wesley book title in 1991 & 1998 “Mathematica. A system for doing mathematics by computers” as quite apt. Obviously the system consists of the language and the software that runs it, but the latter might be provided by other providers too, like Fortran has different compilers. Every programmer knows that the devil is in the details, and that a language documentation on paper might not give the full details of actually running the software. Thus when there are not more software providers then it is only accurate to state the the present definition of the language is given precisely by the one program that runs it. This is only practical and not fundamental. In this situation there is no conflict in thinking of “Mathematica as the language of Mathematica“. Thus in my view there is no need to find a new name for the language. I thought that I was using a language but apparently in Wolfram’s recent view the emphasis was on the computer program. I didn’t read Wolfram’s blog in 2013 and otherwise might have given this feedback.
Wolfram (2017) and (2018) uses the terms “computational essay” and “computational thinking” while the latter is used such that he apparently intends this to mean something like (my interpretation): programming in the Wolfram Language, using internet resources, e.g. the cloud and not necessarily the stand-alone version of Mathematica or now also Wolfram Desktop. My impression is that Wolfram indeed emphasizes computation, and that he perhaps also wants to get rid of a popular confusion of the name “Mathematica” with mathematics only. Apparently he doesn’t want to get rid of that name altogether, likely given his involvement in its history and also its fine reputation.
A related website is https://www.computerbasedmath.org (CBM) by Conrad Wolfram. Most likely Conrad adopts Stephen’s view on computation. It might also be that CBM finds the name “Mathematica” disinformative, as educators (i) may be unaware of what this language and program is, (ii) may associate mathematics with pen and paper, and (iii) would pay attention however at the word “computer”. Perhaps CBM also thinks: You better adopt the language of your audience than teach them to understand your terminology on the history of Mathematica.
I am not convinced by these recent developments. I still think: (1) that this is a two-sided coin (but I am no physicist and do no know about electrons and such), (2) that it is advantageous to clarify to the world: (2a) that mathematics can be used for everything, and (2b) that doing mathematics by computer is a revolution for mankind, and (3) that one should beware of people without didactic training who want to ship computer technology into the classroom. My suggestion to Stephen Wolfram remains, as I did before in (2009, 2015a), that he turns WRI into a public utility like those that exist in Holland – while it already has many characteristics of this. It is curious to see the open source initiatives that apparently will not use the language of Mathematica, now by WRI (also) called the Wolfram Language, most likely because of copyright fears even while it is good mathematics.
Apparently there are legal concerns (but I am no lawyer) that issues like 1+1 = 2 or [Pi] are not under copyright, but that choices for software can be. For example the use of h[x] with square brackets rather than parentheses h(x), might be presented to the copyright courts as a copyright issue. This is awkward, because it is good didactics of mathematics to use the square brackets. Not only computers but also kids may get confused by expressions a(2 + b) and f(x + h) – f(x). Let me refer to my suggestion that each nation sets up its own National Center for Mathematics Education. Presently we have a jungle that is no good for WRI, no good for the open source movement (e.g. R or https://www.python.org or http://jupyter.org), and especially no good for the students. Everyone will be served by clear distinctions between (i) what is in the common domain for mathematics and education of mathematics (the language) and (ii) what would be subject to private property laws (programs in that language, interpreters and compilers for the language) (though such could also be placed into the common domain).
PM. This test has been included in the May 23 2018 article on Arithmetic with H = -1. Earlier related blogs are here and here.
Colignatus, Th. (2009, 2015a), Elegance with Substance, (1) website (2) PDF on Zenodo
Wolfram, S. (1991, 1998), Mathematica. A system for doing mathematics by computer, 2nd edition, Addison-Wesley
Wolfram, S. (2013), What Should We Call the Language of Mathematica?, weblog
Wolfram, S. (2017), What Is a Computational Essay?, weblog
Wolfram, S. (2018), Launching the Wolfram Challenges Site, weblog
]]>Examples of structure would be the rise of the world population and Jared Diamond’s Guns, Germs, and Steel. Obviously, various authors have various suggestions for what they consider to be structure, but the lack of consensus generally doesn’t matter as long as the discussion continues, and as long as people are aware that there are different points of view. It is rather tricky to identify structure for the here and now because it might require the perspective of some centuries to arrive at proper evaluation.
There are also major contingent events that shaped developments. The collapse of civilisation in 1177 BC would be a perfect storm. Caesar might not have crossed the Rubicon. His Alea iacta indicates that he took a calculated risk and the outcome might have been different. If the weather had been better then perhaps the Armada had conquered England and saved the world for Catholicism.
Thus we distinguish structure and relevant and irrelevant contingency.
Brexit came with such surprise that we are still discussing how it could have happened. It very much looks like a perfect storm. The 2016 referendum result has many curious aspects. The referendum question itself doesn’t fit the requirements of a scientifically warranted statistical questionnaire – and the British Electoral Commission doesn’t mind. Even in 2017 17% of UK voters put Remain between different options for Leave, and those of them who voted Leave in 2016 might not have voted so if their preferred option might not materialise (see here). Hannes Grassegger & Mikael Krogerus point to media manipulation. Referenda are instruments of populism, and the better model of democracy is representative democracy. Chris Patten rightly remarks that the UK House of Commons had more options than Theresa May suggests:
“The Brexit referendum last June was itself a disaster. A parliamentary democracy should never turn to such populist devices. Even so, May could have reacted to the 52 per cent vote to quit Europe by saying that she would hand the negotiations to a group of ministers who believed in this outcome and then put the result of the talks in due course to parliament and the people. Instead, she turned the whole of her government into a Brexit machine, even though she had always wished to remain in the EU. Her government’s motto is now “Brexit or bust.” Sadly, we will probably get both.”
My take of the structural cause of Brexit is clarified by the following table. We distinguish Euro and Non-Euro countries versus the political structures of district representation (DR) and equal or proportional representation (EPR).
District representation (DR) | Equal or proportional representation (EPR) | |
Euro | France | Holland (natural quota) Germany (threshold 5%) |
Non-Euro | UK (Brexit) | Sweden (threshold 4%) Norway (non-EU, threshold 4%) |
Update 2018-02-27: On the distinction between DR and EPR, there are: (1) this short overview of elementary statistics with an application to votes and seats, (2) a deconstruction of the disarray in the “political science on electoral systems” (1W1V), and (3) details on the suggestion for an inequality or disproportionality measure (SDID).
In the special Brexit edition of BJPIR, Helen Thompson discusses inevitability and contingency, and concludes that the position of the UK as a non-Euro country in a predominantly Eurozone EU became politically untenable.
Thompson regards the economic angle as a sufficient structural cause. My take is that it is only necessary, and that another necessary element is the form of parliamentarian representation. In my recent paper One woman, one vote. Though not in the USA, UK and France, with the focus on this parlementarian dimension, I forward the diagnosis that the UK political system is the main cause. Brexit is not the proof of a successful UK political system but proof of its failure.
The 2007+ financial crisis highlighted irresponsible herd behaviour in economic science. Brexit highlights irresponsible herd behaviour in political science. Said paper One woman, one vote. Though not in the USA, UK and France (1W1V) shows that political science on electoral systems (on that topic specifically) is still pre-science, comparable to homeopathy, astrology and alchemy. Thus the UK finds itself in the dismal spot of being disinformed about democracy for decades.
The paper runs through the nooks and crannies of confusion and bias. At various points I was surprised by the subtleties of the particular madness. The paper is rather long but this has a clear explanation. When an argument has 100 aspects, and people understand 99% correctly and 1% wrongly, but everyone another 1%, in continuous fashion, then you really want the full picture if you want that all understand it.
But let me alert you to some points.
(1) The paper focuses on Carey & Hix (2011) on an “electoral sweet spot” of 3-8 seats per district. Particular to C&H is that they confuse “most frequent of good” with “the best“. The district magnitude of 3-8 seats appears most frequent in cases that satisfy their criteria for being good, and they turn this into the best. Since such DR would be best, say goodbye to EPR. But it is a confusion.
(2) They use fuzzy words like vote and election. But the words mean different things in DR or EPR. In DR votes are obliterated that EPR translates into seats. Using the same words for different systems, C&H suggest treatment on a par while there are strict logical differences. The Universal Declaration of Human Rights only fits with EPR. Science would use strict distinctions, like “vote in DR” and “vote in EPR”. Political science is still too close to colloquial language, and thus prone to confusion. Obviously I agree that it is difficult to define democracy, and that there are various systems, each with a historical explanation. But science requires clear terms. (See this Varieties of Democracy project, and check that they still have to do a lot too.)
(3) There is a statistical relationship between a measure of disproportionality (EGID) and a measure of the concentrated number of parties (CNP). C&H interprete the first as “interest-representation” and the latter as “accountability”. An interpretation is something else than a model. Using the statistical regularity, they claim to have found a trade-off relation between interest-representation and accountability. Instead, the scientific approach would be to try explain the statistical regularity for what it is. The suggested interpretation is shaky at best. One cannot use a statistical regularity as an argument on content and political principle (like One woman, one vote).
(4) They present a mantra, and repeat it, that there would be a trade-off between interest-representation and accountability. The best point [confusion] would be achieved at a district magnitude of 3-8 seats per district. However, they do not present a proper model and measure for accountability. My paper presents such a model, and shows that the mantra is false. Not DR but EPR is most accountable. EPR is obviously most interest-representative, so that there is no trade-off. Thus the C&H paper fails in the scientific method of modeling and measuring. It only has the method of repeating tradition and a mantra, with some magic of using interpretations. (Section 3.6 of 1W1V should start opening eyes of political scientists on electoral systems.)
(5) The C&H paper is the top of a line of research in “political science on electoral systems”. This paper fails and thus the whole line fails. Section 4.5 of 1W1V shows confusion and bias in general in political science on electoral systems, and the C&H paper is no exception to this.
The cure of Brexit might well be that it just happens, and that we must learn to live with it. The EU lives with Norway while NATO has its Arctic training there.
Seen from the angle of the cause via the political structure, it may also be suggested that both France and the UK switch from DR to EPR, and that the newly elected UK House of Commons re-evaluates Brexit or Bregret. This switch may well cause the break-up of the parties of the Conservatives and Labour into Remain or Leave parties, but such would be the consequence of democracy and thus be fine by itself. We would no longer have Theresa May who was for Remain leading the Leavers and Jeremy Corbyn who was for Leave leading the Remainers. (For an indication, see here.) The other EU member states tend to stick to the Brexit deadline of March 29 2019, but when they observe the cause for Brexit and a new objective in the UK to deal with this (fateful) cause by switching to EPR, then this deadline might be shifted to allow the UK to make up its mind in a proper way.
Obviously, a UK switch to EPR is advisable in its own right, see said paper. It would also allow the new UK House of Commons to still adopt Brexit. The advantage of such an approach and decision would be that it would have the democratic legitimacy that is lacking now.
Thompson’s article surprised me by her discussion of the 2010 UK Sovereignty Bill (that calls itself an Act). She calls it a “referendum lock”, and indeed it is. The Bill / Act states:
“2 Treaties. No Minister of the Crown shall sign, ratify or implement any treaty or law, whether by virtue of the prerogative powers of the Crown or under any statutory authority, which — (a) is inconsistent with this Act; or (b) increases the functions of the European Union affecting the United Kingdom without requiring it to be approved in a referendum of the electorate in the United Kingdom.”
The approach is comparable to the one in Ireland, in which EU treaties are subject to referenda too. In Holland, only changes in the constitution are subject to new elections and affirmation by the newly elected parliament, while treaties are exempt from this – and this is how the EU constitution of 2005 got rejected in a referendum but the Lisbon treaty got accepted in Dutch parliament. Currently a state commission is investigating the Dutch parliamentary system.
Thompson explains that the UK referendum lock had the perverse effect that EU leaders started to avoid the instrument of a treaty and started to use other ways to enact policies. For EU-minded Ireland, the instrument of a referendum was acceptable but for EU-skeptic UK the instrument was a poison pill. Why put much effort in negotiating a treaty if it could be rejected by the UK circus (partly created by its system of DR) ?
Thompson explains that while the referendum lock had been intended to enhance the UK position as a non-euro country w.r.t. eurozone UK, in effect it weakened Cameron’s position. The world noticed this and this weak position was fuel for the Brexiteers.
Brexit is mostly caused in the UK itself. Thompson doesn’t call attention to these relevant contingencies:
I am reminded of the year 1988 at the Dutch Central Planning Bureau (CPB) when we looked at the Cecchini report. One criticism was that the report was too optimistic about productivity growth and less realistic on the costs of displaced workers. An observation by myself, though not further developed, was that, with more job mobility, people might prefer a single language barrier to a double one. People from the UK might move easier to Northern European countries that speak English well. People from the rest of Europe who have learned some English might prefer to go to the UK, to avoid having to deal with two other languages. I don’t know much about migration and I haven’t checked whether the UK has a higher share of it or not, and whether this language effect really matters. Given the role in the discussion it obviously would be a relevant contingency. Perhaps the UK and Ireland might claim a special position because of the language effect, and this might encourage other countries to switch to English too. But I haven’t looked into this.
The other elephant in the room is my own analysis in political economy. It provides an amendment to Thompson’s analysis.
The main idea in DRGTPE was available in 1990 with the collection of background papers in 1992 (published by Guido den Broeder of Magnana Mu). Thus the EU might have had a different approach to EMU. The later edition of DRGTPE contains a warning about financial risk that materialised in 2007+. CSBH and MGMW provide a solution approach for the current problems.
If the EU would adopt such policies then there would be much less migration, since people would tend to prefer to remain at home (which is why I regard migration as a secondary issue and less in need for studying).
If the EU and UK would adopt such policies then there might still be Brexit or Bregret. Thus UK politicians might still prefer what they are now trying to find out what they prefer.
My impression is that the above gives a clear structural explanation for the UK decision for Brexit and an indication what contingent events were relevant. Knowing what caused helps to identify a cure. It is remarkable how large the role of denial in all of this is. Perhaps this story about the polar bear provides a way to deal with this huge denial (as polar elephants a.k.a. mammoths are already extinct).
]]>I now want to focus on Pearson’s 1897 paper Mathematical Contributions to the Theory of Evolution.–On a Form of Spurious Correlation Which May Arise When Indices Are Used in the Measurement of Organs.
The main theme is that if you use the wrong model then the correlations for that model will be spurious compared to the true model. Thus, Pearson goes at length to create a wrong model, and compares this to what he claims is the true model. It might be though that he still didn’t develop the true model. Apart from this complexity, it is only admirable that he points to the notion of such spurious correlation in itself.
One example in Pearson’s paper is the measurement of skulls in Bavaria (p495). The issue concerns compositional data, i.e. data vectors that add up to a given total, say 100%. The former entry on this weblog presented the inequality / disproportionality measure SDID for votes and seats. These become compositional data when we divide them by their sum totals, so that we compare 100% of the votes with 100% of the seats.
Pearson’s analysis got a sequel in the Aitchison geometry, see this historical exposition by Vera Pawlowsky-Glahn and Juan José Egozcue, The closure problem: one hundred years of debate. Early on, I was and still am a fan of the Aitchison & Brown book on the lognormal distribution but I have my doubts about the need of this particular geometry for compositional data. In itself the Aitchison geometry is a contribution, with a vector space, norm and inner product. When we transform the data to logarithms, then multiplication becomes addition, and powers become scalars, so that we can imagine such a vector space, yet, the amazing finding is that rebasing to 1 or 100% can be maintained. It is called “closure” when a vector is rebased to a constant sum. What, however, is the added value of using this geometry ?
It may well be that different fields of application still remain different on content, so that when they generate compositional data, then these data are only similar in form, while we should be careful in using the same techniques only because of that similar form. We must also distinguish:
An example of the latter might be the paper by Javier Palarea-Albaladejo, Josep Antoni Martın-Fernandez and Jesus A. Soto (2012) in which they compare the compositions of milk of different mammals. I find this difficult to judge on content since I am no biologist. See the addendum below on the distance function.
In a fine overview by sheets, Pawlowsky-Glahn, Egozcue & Meziat 2007 present the following example, adapted from Aitchison. They compare two sets of soil samples, of which one sample is contaminated by water. If you want to spot the problem with this analysis yourself, take a try, and otherwise read on.
When the water content in the sample of A is dropped, then the test scores are rebased to the total of 100% for B again. E.g. for the 60% water in sample 1, this becomes:
{0.1, 0.2, 0.1} / (0.1 + 0.2 + 0.1) = {0.25, 0.5, 0.25}
PM. A more complex example with simulation data is by David Lovell.
It is useful to first reproduce the example so that we can later adapt it.
In Wolfram Alpha, we can reproduce the outcome as follows.
For A, the input code is: mat1 = {{0.1, 0.2, 0.1, 0.6}, {0.2, 0.1, 0.2, 0.5}, {0.3 , .3, 0.1, 0.3}}; Correlation[mat1] // Chop // MatrixForm.
For B, the input code is: mat1 = {{0.1, 0.2, 0.1, 0.6}, {0.2, 0.1, 0.2, 0.5}, {0.3 , .3, 0.1, 0.3}}; droplast[x_List?VectorQ] := Module[{a}, a = Drop[x, -1]; a / (Plus @@ a)]; mat2 = droplast /@ mat1; Correlation[mat2] // Chop // MatrixForm.
In the former weblog entry, we had SDID[v, s] for the votes v and seats s. In this way of thinking, we would reason differently. We would compare (correlate) rows and not columns.
There is also a difference that correlation uses centered data while Sine and Cosine use original or non-centered data. Perhaps this contributed to Pearson’s view.
One possibility is that we compare sample 1 according to A with sample 1 according to B, as SDID[1A*, 1B]. Since the measures of A also contain water, we must drop the water content and create A*. The assumption is that A and B are independent measurements, and that we want to see whether they generate the same result. When the measurements are not affected by the content of water, then we would find zero inequality / disproportionality. However, Pawlowsky et al. do not state the problem as such.
The other possibility is that we would compare SDID[sample i, sample j].
Instead of using SDID for inequality / disproportionality, let us now use the cosine as a measure for similarity.
For A, the input code is: mat1 = {{0.1, 0.2, 0.1, 0.6}, {0.2, 0.1, 0.2, 0.5}, {0.3 , .3, 0.1, 0.3}}; cos[x__] := 1 – CosineDistance[x]; Outer[cos, mat1, mat1, 1] // Chop // MatrixForm.
Since the water content is not the same in all samples, above scores will be off. To see whether these similarities are sensitive to the contamination by the water content, we look at the samples according to B.
The input code for Wolfram Alpha is: mat1 = {{0.1, 0.2, 0.1, 0.6}, {0.2, 0.1, 0.2, 0.5}, {0.3 , .3, 0.1, 0.3}}; cos[x__] := 1 – CosineDistance[x]; droplast[x_List?VectorQ] := Module[{a}, a = Drop[x, -1]; a / (Plus @@ a)]; mat2 = droplast /@ mat1; Outer[cos, mat2, mat2, 1] // Chop // MatrixForm.
Since the water content differed so much per sample, and apparently is not considered to be relevant for the shares of the other components, the latter matrix of similarities is most relevant.
If we know that the samples are from the same soil, then this would give an indication of sample variability. Conversely, we might have information about the dispersion of samples, and perhaps we might determine whether the samples are from the same soil.
Obviously, one must have studied soil samples to say something on content. The above is only a mathematical exercise. This only highlights the non-transposed case (rows) versus the transposed case (columns).
Reading the Pearson 1897 paper shows that he indeed looks at the issue from the angle of the columns, and that he considers calibration of measurements by switching to relative data. He gives various examples, but let me show the case of skull measurement, that may still be a challenge:
Pearson presents two correlation coefficients for B / L with H / L. One based upon the standard definition (that allows for correlations between the levels), and one baptised “spurious”, based upon the assumption of independent distributions (and thus zero correlations for the levels). Subsequently he throws doubt on the standard correlation because of the high value of the spurious correlation.
One must be a biologist or even a skull-specialist to determine whether this is a useful approach. If the true model would use relative data with zero correlations, what is the value of the assumptions of zero or nonzero correlations for the absolute values ? What is useful depends upon the research question too. We can calculate all kinds of statistics, but what decision is intended ?
It is undoubtedly a contribution by Pearson that looking at phenomena in this manner can generate what he calls “spurious correlation”. Whatever the model, it is an insight that using the wrong model can create spurious correlation and a false sense of achievement. I would feel more comfortable though when Pearson had also mentioned the non-transposed case, which I would tend to regard as the proper model, i.e. comparing skulls rather than correlating categories on skulls. Yet he doesn’t mention it.
Apparently the Aitchison geometry provides a solution to Pearson’s approach, thus still looking at transposed (column) data. This causes the same discomfort.
The above uses soil and skulls, which are not my expertise. I am more comfortable with votes and seats, or budget shares in economics (e.g. in the Somermeyer model or the indirect addilog demand system, Barten, De Boer).
Pearson was not confused on what he defined as spurious correlation. He might have been confused about the proper way to deal with compositional data, namely looking at columns rather than rows. This however also depends upon the field of interest and the research question. Perhaps a historian can determine whether Pearson also looked at compositional data from rows rather than columns.
For geological data, Watson & Philip (1989) already discussed the angular distance. Martin-Fernandez, Barcelo-Vidal, Pawlowsky-Glahn (2000), “Measures of differences for compositional data and hierarchical clustering methods“, discuss distance measures. They also mention the angle between two vectors, found via arccos[cos[v, s]], for votes v and seats s. It is mentioned in the 2nd row of their Table 1. The vectors can also normalised to the unit simplex as w = v / Sum[v] and z = s / Sum[s], though cos is insensitive with cos[w, z] = cos[v, s].
In sum, the angular distance, or the use of the sine as a distance measure and the cosine as a similarity measure, satisfy the Aitchison criteria of invariance to scale and permutation, but do not satisfy subcompositional dominance and invariance to translation (perturbation).
This discussion makes me wonder whether there are still key differences between compositional data in terms of concepts. The compositional form should not distract us from the content. For a Euclidean norm, a translation leaves a distance unaffected, as Norm[x – y] = Norm[(x +t) – (y + t)}. This property can by copied for logratio data. However, for votes and seats, it is not clear why a (per party different) percentage change vector should leave the distance unaffected (as happens in logratio distance).
An election only gives votes and seats. Thus there is no larger matrix of data. Comparison with other times and nations has limited meaning. Thus there may be no need for the full Aitchison geometry.
At this moment, I can only conclude that Sine (distance) and Cosine (similarity) are better for votes and seats than what political scientists have been using till now. It remains to be seen for votes and seats whether the logratio approach would be better than the angular distance and the use of Sine and Cosine.
]]>Such imbalance is not uncommon. At the US election of November 8 2016, the Republicans got 49.1% of the votes and 55.4% of the seats, while the Democrats got 48% of the votes and 44.6% of the seats. At the UK general election of June 8 2017, the Conservatives got 42.2% of the votes and 48.8% of the seats while Labour got 39.9% of the votes and 40.3% of the seats (the wikipedia data of October 16 2017 are inaccurate).
This article clarifies a new and better way to measure this inequality or disproportionality of votes and seats. The new measure is called Sine-Diagonal Inequality / Disproportionality (SDID) (weblink to main article). The new measure falls under descriptive statistics. Potentially it might be used in any area where one matches shares or proportions, like the proportions of minerals in different samples. SDID is related to statistical concepts like R-squared and the regression slope. This article looks at some history, as Karl Pearson (1857-1936) created the R-Squared and Ronald A. Fisher (1890-1962) in 1915 determined its sample distribution. The new measure would also be relevant for Big Data. William Gosset (1876-1937) a.k.a. “Student” was famously unimpressed by Fisher’s notion of “statistical significance” and now is vindicated by descriptive statistics and Big Data.
Statistics has the triad of Design, Description and Decision.
Historically, statisticians have been working on all these areas of design, description and decision, but the most difficult was the formulation of decision methods, since this involved both the calculus of reasoning and the more complex mathematics on normal, t, chi-square, and other frequency distributions. In practical work, the divide between the experimental and the non-experimental (observational) sciences appeared insurmountable. The experimental sciences have the advantages of design and decisions based upon samples, and the observational sciences basically rely on descriptive statistics. When the observational sciences do regressions, there is an ephemeral application of statistical significance that invokes the Law of Large Numbers, that all error approximates the normal distribution.
This traditional setup of statistics is being challenged in the last decades by Big Data – see also this discussion by Rand Wilcox in Significance May 2017. When all data are available, and when you actually have the population data, then the idea of using a sample evaporates, and you don’t need to develop hypotheses on the distributions anymore. In that case descriptive statistics becomes the most important aspect of statistics. For statistics as a whole, the emphasis shifts from statistical decision making to decisions on content. While descriptive statistics had been applied mostly to samples, Big Data now causes the additional step how these descriptions relate to decisions on content. In fact, such questions already existed for the observational sciences like for macro-economics and national elections, in which the researcher only had descriptive statistics, and lacked the opportunity to experiment and base decisions upon samples. The disadvantaged areas now provide insights for the earlier advantaged areas of research.
The key insight is to transform the loss function into a descriptive statistic itself. An example is the Richter scale for the magnitude of earthquakes. It is both a descriptive statistic and a factor in the loss function. A nation or regional community has on the one hand the cost of building and construction and on the other hand the risk of losing the entire investments and human lives. In the evaluation of cost and benefit, the descriptive statistic helps to clarify the content of the issue itself. The key issue is no longer a decision within statistical hypothesis testing, but the adequate description of the data so that we arrive at a better cost-benefit analysis.
Let us return to the election for the House of Representatives (USA) or the House of Commons (UK). The criterion of One man, one vote translates into the criterion that the shares of seats equal the shares of votes. We are comparing two vectors here.
The reason why the shares of seats and votes do not match is because the USA and UK use a particular setup. The setup is called an “electoral system”, but since it does not satisfy the criterion of One man, one vote, it does not really deserve that name. The USA and UK use both (single member) districts and the criterion of Plurality per district, meaning that the district seat is given to the candidate with the most votes – also called “first past the post” (FPTP). This system made some sense in 1800 when the concern was district representation. However, when candidates stand for parties then the argument for district representation loses relevance. The current setup does not qualify for the word “election” though it curiously continues to be called so. It is true that voters mark ballots but that is not enough for a real election. When you pay for something in a shop then this is an essential part of the process, but you also expect to receive what you ordered. In the “electoral systems” in the USA and UK, this economic logic does not apply. Only votes for the winner elect someone but the other votes are obliterated. For such reasons Holland switched to equal / proportional representation in 1917.
For descriptive statistics, the question is how to measure the deviation of the shares of votes and seats. For statistical decision making we might want to test whether the US and UK election outcomes are statistically significantly different from inequality / proportionality. This approach requires not only a proper descriptive measure anyway, but also some assumptions on the distribution of votes which might be rather dubious to start with. For this reason the emphasis falls on descriptive statistics, and the use of a proper measure for inequality / disproportionality (ID).
A measure proposed by, and called after, Loosemore & Hanby in 1971 (LHID) uses the sum of the absolute deviations of the shares (in percentages), divided by 2 to correct for double counting. The LHID for the UK election of 2017 is 10.5 on a scale of 100, which means that 10.5% of the 650 seats (68 seats) in the UK House of Commons are relocated from what would be an equal allocation. When the UK government claims to have a “mandate from the people” then this is only because the UK “election system” is so rigged that many votes have been obliterated. The LHID gives the percentage of relocated seats but is insensitive to how these actually are relocated, say to a larger or smaller party.
The Euclid / Gallagher measure proposed in 1991 (EGID) uses the Euclidean distance, again corrected for double counting. For an election with only two parties EGID = LHID. The EGID has become something like the standard in political science. For the UK 2017 the EGID is 6.8 on a scale of 100, which cannot be interpreted as a percentage of seats like LHID, but which indicates that the 10.5% of relocated seats are not concentrated in the Conservative party only.
Alan Renwick in 2015 tends to see more value in LHID than EGID: “As the fragmentation of the UK party system has increased over recent years, therefore, the standard measure of disproportionality [thus EGID] has, it would appear, increasingly understated the true level of disproportionality.”
The new Sine-Diagonal Inequality / Disproportionality (SDID) measure – presented in this paper – looks at the angle between the vectors of the shares of votes and seats.
It appears that the sine is more sensitive than either absolute value (LHID) and Euclidean distance (EGID). It is closer to the absolute value for small angles, and closer to the Euclidean distrance for larger angles. See said paper, Figure 1 on page 10. SDID is something like a compromise between LHID and EGID but also better than both.
When we regress the shares of the seats on the shares of the votes without using a constant – i.e. using Regression Through the Origin (RTO) – then this gives a single regression coefficient. When there is equality / proportionality then this regression coefficient is 1. This has the easy interpretation that this is the diagonal in the votes & seats space. This explains the name of SDID: when the regression coefficient generates the diagonal, then the sine is zero, and there is no inequality / disproportionality.
Said paper – see page 38 – recovers a key relationship between on the one hand the sine and on the other hand the Euclidean distance and this regression coefficient. On the diagonal, the sine and Euclidean distance are both zero. Off-diagonal, the sine differs from the Euclidean distance in nonlinear manner by means of a factor given by the regression coefficient. This relationship determines the effect that we indicated above, how SDID compromises between and improves upon LHID and EGID.
There appears to be a relationship between said regression coefficient and the cosine itself. This allows for a double interpretation as both slope and similarity measure. This weblog text is intended to avoid formulas as much as possible and thus I refer to said paper for the details. Suffice to say here is that, at first, it may seem to be a drawback that such a double interpretation is possible, yet, on closer inspection the relationship makes sense and it is an advantage to be able to switch perspective.
In human psychology there appears to be a distinction between actual differences and perceived differences. This is called the Weber – Fechner law. When a frog is put into a pan with cool water and slowly boiled to death, it will not jump out. When a frog is put into a pan with hot water it will jump out immediately. People may notice differences between low vote shares and high seat shares, but they may be less sensitive to small differences, while these differences actually can still be quite relevant. For this reason, the SDID uses a sensitivity transform. It uses the square root of the sine.
(PM. A hypothesis is that the USA and UK call their national “balloting events” still “elections”, is that the old system of districts has changed so gradually into the method of obliterating votes that many people did not notice. It is more likely though that that some parties recognised the effect, but have an advantage under the present system, and then do not want to change to equal / proportional representation.)
Subsequently, the sine and its square root have values in the range [0, 1]. In itself this is an advantage, but it comes with leading zeros. We might multiply with 100 but this might cause the confusion as if it would be percentages. The second digit might give a false sense of accuracy. It is more useful to multiply this by 10. This gives values like on a report card. We can compare here to Bart Simpson, who appreciates low values on his report card.
Finally, when we compare, say, votes {49, 51} and seats {51, 49}, then we see a dramatic change of majority, even though there is only a slight inequality / disproportionality. It is useful to have an indicator for this too. It appears that this can be done by using a negative sign when such majority reversal occurs. This method of indicating majority reversals is not so sophisticated yet, and at this stage consists of using the sign of the covariance of the vectors of votes and seats.
This present text avoids formulas but it is useful to give the formula for the new measure of SDID, so that the reader may link up more easily with the paper in which the new measure is actually developed. For the vectors of votes and seats we use the symbols v and s, and the angle between the two vectors give cosine and then sine:
SDID[v, s] = sign 10 √ Sin[v, s]
For the UK 2017, the SDID value is 3.7. For comparison the values of Holland with equal / proportional representation are: LHID 3, EGID 1.7, SDID 2.5. It appears that Holland is not yet as equal / proportional as can be. Holland uses the Jefferson / D’Hondt method, that favours larger parties in the allocation of remainder seats. At elections there are also the wasted vote, when people vote for fringe parties that do not succeed in getting seats. In a truly equal or proportional system, the wasted vote can be respected by leaving seats empty or by having a qualified majority rule.
Remarkably, Karl Pearson (1857-1936) also used the cosine when he created R-squared, also known as the “coefficient of determination“. Namely:
The SDID measure thus provides a “missing link” in statistics between centered and non-centered data, and also provides a new perspective on R-squared itself.
Apparently till now statistics found little use for original (non-centered) data and RTO. A possible explanation is that statistics fairly soon neglected descriptive statistics as less challenging, and focused on statistical decision making. Textbooks prefer the inclusion of a constant in the regression, so that one can test whether it differs from zero with statistical significance. The constant is essentially used as an indicator for possible errors in modeling. The use of RTO or the imposition of a zero constant would block that kind of application. However, this (traditional, academic) focus on statistical decision making apparently caused the neglect of a relevant part of the analysis, that now comes to the surface.
R-squared is often mentioned in statistical reports about regressions, but actually it is not much used for other purposes than reporting only. Cosma Shalizi (2015:19) states:
“At this point, you might be wondering just what R-squared is good for — what job it does that isn’t better done by other tools. The only honest answer I can give you is that I have never found a situation where it helped at all. If I could design the regression curriculum from scratch, I would never mention it. Unfortunately, it lives on as a historical relic, so you need to know what it is, and what misunderstandings about it people suffer from.”
At the U. of Virginia Library, Clay Ford summarizes Shalizi’s points on the uselessness of R-squared, with a reference to his lecture notes.
Since the cosine is symmetric, the R-squared is the same for regressing y given x, or x given y. Shalizi (2015, p18) infers from the symmetry: “This in itself should be enough to show that a high R² says nothing about explaining one variable by another.” This is too quick. When theory shows that x is a causal factor for y then it makes little sense to argue that y explains x conversely. Thus, for research the percentage of explained variation can be informative. Obviously it matters how one actually uses this information.
When it is reported that a regression has an R-squared of 70% then this means that 70% of the variation of the explained variable is explained by the model, i.e. by variation in the explanatory variables and the estimated coefficients. In itself such a report does not say much, for it is not clear whether 70% is a little or a lot for the particular explanation. For evaluation we obviously also look at the regression coefficients.
One can always increase R-squared by including other and even nonsensical variables. For a proper use of R-squared, we would use the adjusted R-squared. R-adj finds its use in model specification searches – see Dave Giles 2013. For an increase of R-adj coefficients must have an absolute t-value larger than 1. A proper report would show how R-adj increases by the inclusion of particular variables, e.g. also compared to studies by others on the same topic. Comparison on other topics obviously would be rather meaningless. Shalizi also rejects R-adj and suggests to work directly with the mean squared error (MSE, also corrected for the degrees of freedom). Since R-squared is the cosine, then the MSE relates to the sine, and these are basically different sides of the same coin, so that this discussion is much a-do about little. For standardised variables (difference from mean, divided by standard deviation), the R-squared is also the coefficient of regression, and then it is relevant for the effect size.
R-squared is a sample statistic. Thus it depends upon the particular sample. A hypothesis is that the population has a ρ-squared. For this reason it is important to distinguish between a regression on fixed data and a regression in which the explanatory variables also have a (normal) distribution (errors in variables). In his 1915 article on the sample distribution of R-squared. R.A Fisher (digital library) assumed the latter. With fixed data, say X, the outcome is conditional on X, so that it is better to write ρ[X], lest one forgets about the situation. See my earlier paper on the sample distribution of R-adj. Dave Giles has a fine discussion about R-squared and adjusted R-squared. A search gives more pages. He confirms the “uselessnes” of R-squared: “My students are often horrified when I tell them, truthfully, that one of the last pieces of information that I look at when evaluating the results of an OLS regression, is the coefficient of determination (R^{2}), or its “adjusted” counterpart. Fortunately, it doesn’t take long to change their perspective!” Such a statement should not be read as the uselessness of cosine or sine in general.
I am not familiar with the history of statistics, and it is unknown to me what else Pearson, Fisher, Gosset and other founding and early authors wrote about the application of the cosine or sine. The choice to apply the cosine to centered data to create R-squared is deliberate, and Pearson would have been aware that it might also be applied to original (non-centered) data. It is also likely that he would not have the full perspective above, because then it would have been in the statistical textbooks already. It would be interesting to know what the considerations at time were. Quite likely the theoretical focus was on statistical decision making rather than on description, yet this for me unknown history would put matters more into perspective.
Part of the history is that R.A. Fisher with his attention for mathematics emphasized precision while W.S. Gosset with his attention to practical application emphasized the effect size of the coefficients found by regression. Somehow, statistical significance in terms of precision became more important than content significance, and empirical research has rather followed Fisher than the practical relevance of Gosset. This history and its meaning is discussed by Stephen Ziliak & Deirdre McCloskey 2007, see also this discussion by Andrew Gelman. As said, for standardised variables, the regression coefficient is the R-squared, and this is best understood with attention for the effect size. For some applications a low R-squared would still be relevant for the particular field.
The new measure SDID provides a better description of the inequality or disproportionality of votes and seats compared to existing measures. The new measure has been tailored to votes and seats, by means of greater sensitivity to small inequalities, and because a small change in inequality may have a crucial impact on the (political) majority. For different fields, one could taylor measures in similar manner.
That the cosine could be used as a measure of similarity has been well-known in the statistics literature since the start, when Pearson used the cosine for centered data to create R-square. For the use of the sine I have not found direct applications, but its use is straightforward when we look at the opposite of similarity.
The proposed measure provides an enlightening bridge between descriptive statistics and statistical decision making. This comes with a better understanding of what kind of information the cosine or R-squared provides, in relation to regressions with and without a constant. Statistics textbooks would do well by providing their students with this new topic for both theory and practical application.
]]>