Category Archives: Uncategorized

On electric bikes and science news

There is some debate on whether electric-assist bikes are good or bad. The argument on the negative side is that people will stop cycling and bike roads will be taken over by these (electric-)motorized vehicles. On the positive side, the claim is that electric bikes replace motor scooters, cars and public transit, which is good for the environment and perhaps health if people actually pedal their electric bike a bit. To summarize: electric bikes good if replace motor vehicles, electric bikes bad if replace bicycles or walking. It is an empirical question what the substitution sizes are.

There are many calls for scientists to communicate better, engage the public, present their results simply and interestingly etc. Whether more science news and popular science is good or bad depends on what it replaces, just like for electric bikes. If dumbed down entertainment science replaces the rigorous variety, this is bad. If science news replaces brainless news (celebrity gossip, funny animals, speculation on future events), it is good. It is an empirical question to what extent scientists switch to popular topics and crafting press releases if their evaluations rely more on outreach or policy impact. Also a question for the data is which news are left out of print to make room for more science news.

To maximize education of the public, there is a tradeoff between the seriousness of the science presented and the size of the audience. Research article level complexity is accessible to only a few experts. Entertainment is watched by many, but does not educate. The optimum must be somewhere in the middle.

Similarly, difficult courses in a university have few students taking them, but teach those few more than fun and easy subjects. The best complexity level is somewhere between standup comedy and a research seminar.

Statistics with a single history

Only one history is observable to a person – the one that actually happened. Counterfactuals are speculation about what would have happened if choices or some other element of the past history had differed. Only one history is observable to humanity as a whole, to all thinking beings in the universe as a whole, etc. This raises the question of how to do statistics with a single history.

The history is chopped into small pieces, which are assumed similar to each other and to future pieces of history. All conclusions require assumptions. In the case of statistics, the main assumption is “what happened in the past, will continue to happen in the future.” The “what” that is happening can be complicated – a long chaotic pattern can be repeated. It should be specified what the patterns of history consist of before discussing them.

The history observable to a brain consists of the sensory inputs and memory. Nothing else is accessible. This is pointed out by the “brain in a jar” thought experiment. Memory is partly past sensory inputs, but may also depend on spontaneous changes in the brain. Machinery can translate previously unobservable aspects of the world into accessible sensory inputs, for example convert infrared and ultraviolet light into visible wavelengths. Formally, history is a function from time to vectors of sensory inputs.

The brain has a built-in ability to classify sensory inputs by type – visual, auditory, etc. This is why the inputs form a vector. For a given sense, there is a built-in “similarity function” that enables comparing inputs from the same sense at different times.

Inputs distinguished by one person, perhaps with the help of machinery, may look identical to another person. The interpretation is that there are underlying physical quantities that must differ by more than the “just noticeable difference” to be perceived as different. The brain can access physical quantities only through the senses, so whether there is a “real world” cannot be determined, only assumed. If most people’s perceptions agree about something, and machinery also agrees (e.g. measuring tape does not agree with visual illusions), then this “something” is called real and physical. The history accessible to humanity as a whole is a function from time to the concatenation of their sensory input vectors.

The similarity functions of people can also be aggregated, compared to machinery and the result interpreted as a physical quantity taking “similar” values at different times.

A set of finite sequences of vectors of sensory inputs is what I call a pattern of history. For example, a pattern can be a single sequence or everything but a given sequence. Patterns may repeat, due to the indistinguishability of physical quantities close to each other. The finer distinctions one can make, the fewer the instances with the same perception. In the limit of perfect discrimination of all variable values, history is unlikely to ever repeat. In the limit of no perception at all, history is one long repetition of nothing happening. The similarity of patterns is defined based on the similarity function in the brain.

Repeated similar patterns together with assumptions enable learning and prediction. If AB is always followed by C, then learning is easy. Statistics are needed when this is not the case. If half the past instances of AB are followed by C, half by D, then one way to interpret this is by constructing a state space with a probability distribution on it. For example, one may assume the existence of an unperceived variable that can take values c,d and assume that ABc leads deterministically to ABC and ABd to ABD. The past instances of AB can be interpreted as split into equal numbers of ABc and ABd. The prediction after observing AB is equal probabilities of C and D. This is a frequentist setup.

A Bayesian interpretation puts a prior probability distribution on histories and updates it based on the observations. The prior may put probability one on a single future history after each past one. Such a deterministic prediction is easily falsified – one observation contrary to it suffices. Usually, many future histories are assumed to have positive probability. Updating requires conditional probabilities of future histories given the past. The histories that repeat past patterns are usually given higher probability than others. Such a conditional probability system embodies the assumption “what happened in the past, will continue to happen in the future.”

There is a tradeoff between the length of a pattern and the number of times it has repeated. Longer patterns permit prediction further into the future, but fewer repetitions mean more uncertainty. Much research in statistics has gone into finding the optimal pattern length given the data. A long pattern contains many shorter ones, with potentially different predictions. Combining information from different pattern lengths is also a research area. Again, assumptions determine which pattern length and combination is optimal. Assumptions can be tested, but only under other assumptions.

Causality is also a mental construct. It is based on past repetitions of an AB-like pattern, without occurrence of BA or CB-like patterns.

The perception of time is created by sensory inputs and memory, e.g. seeing light and darkness alternate, feeling sleepy or alert due to the circadian rhythm and remembering that this has happened before. History is thus a mental construct. It relies on the assumptions that time exists, there is a past in which things happened and current recall is correlated with what actually happened. The preceding discussion should be restated without assuming time exists.

 

Milk powder with milk

Dessert recipe: half a bowl of milk, equal volume of powdered milk on top. Eat with a spoon. Tasty, healthy, quick and easy to make, cheap.
The reason not to eat the powdered milk from the pack directly is that a slight breath will make the powder fly. If some of it hits the back of the throat, it causes coughing, which launches the rest of the powder all over the room. Adding liquid milk removes the choking hazard.

On simple answers

Bayes’ rule exercise: is a simple or a complicated answer to a complicated problem more likely to be correct?

Depends on the conditional probabilities: if simple questions are more likely to have simple answers and complex questions complicated, then a complicated answer is more likely to be correct for a complicated problem.

It seems reasonable that the complexity of the answer is correlated with the difficulty of the problem. But this is an empirical question.

If difficult problems are likely to have complex answers, then this is an argument against slogans and ideologies. These seek to give a catchy one-liner as the answer to many problems in society. No need to think – ideology has the solution. Depending on your political leaning, poverty may be due to laziness or exploitation. The foreign policy “solution” is bombing for some, eternal appeasement for others.

The probabilistic preference for complex answers in complicated situations seems to contradict Occam’s razor (among answers equally good at explaining the facts, the simplest answer should be chosen). There is no actual conflict with the above Bayesian exercise. There, the expectation of a complex answer applies to complicated questions, while a symmetric anticipation of a simple answer holds for simple problems. The answers compared are not equally good, because one fits the structure of the question better than the other.

Which ideology is more likely to be wrong?

Exercise in Bayes’ rule: is an ideology more likely to be wrong if it appeals relatively more to poor people than the rich?

More manipulable folks are more likely to lose their money, so less likely to be rich. Stupid people have a lower probability of making money. By Bayes, the rich are on average less manipulable and more intelligent than the poor.

Less manipulable people are less likely to find an ideology built on fallacies appealing. By Bayes, an ideology relatively more appealing to the stupid and credulous is more likely to be wrong. Due to such people being poor with a higher probability, an ideology embraced more by the poor than the rich is more likely to be fallacious.

Another exercise: is an ideology more likely to be wrong if academics like it relatively more than non-academics?

Smarter people are more likely to become academics, so by Bayes’ rule, academics are more likely to be smart. Intelligent people have a relatively higher probability of liking a correct ideology, so by Bayes, an ideology appealing to the intelligent is more likely to be correct. An ideology liked by academics is correct with a higher probability.

Economics to guide materials science

There are too many possible materials to test them all, or even simulate by computer. Materials scientists theorize what combinations of elements are likely to yield the desired properties, but still there are too many possibilities. One way to narrow the choice is to use economics.

If the goal of developing a material is to change the world or make money, the benefit of the invention must exceed the cost. The benefit comes from the improved characteristics of the material relative to existing alternatives. What the market is willing to pay for an improvement depends on its size. There may be a theoretical maximum for a property, or its historical rate of increase may be used to forecast the likely improvement. Once an approximate willingness to pay for a unit of the candidate invented material is known, this can be compared with its estimated cost.

Financial firms dealing in commodity futures forecast the prices of chemical elements over the likely commercialization time horizon. Only materials using a combination of elements that is cheap enough are commercially promising. Cheap enough means that the improved material must cost less per unit than the market is willing to pay for it. An expensive element can be used, but only in appropriately tiny quantity. The requirement that the bundle of elements cost less than some bound cuts down on the number of combinations that are worth testing. Similarly, the manufacturing method must be cheap enough, so some methods may be ignored.

The basic cost-benefit analysis is a simple idea, though the benefit estimation may be complicated in practice. Probably the companies producing various materials are already taking the potential cost and benefit of an innovation into account in their R&D, but academic materials scientists perhaps not. If the goal is to advance fundamental science and satisfy one’s curiosity, then the cost of the material may not be an issue. But for the world to use the material, it must be cheap enough.

A practical recommendation is for an application-oriented lab to put up a periodic table with the prices of the elements added. A spreadsheet with the prices of commodities can be used to calculate the cost of a candidate combination for a new material. Testing the candidates should proceed in the order of decreasing “profit” (benefit minus cost of the material). This profit is not necessarily the same as commercial profit, because the benefit may include its whole contribution to society, not just the revenue to the producer.

On journalistic privilege

Journalists have certain privileges over the average citizen – they get access to inside information, public figures and press conferences. An attack against a journalist creates more outrage, because it is seen as damaging the free press. These advantages are not given so that journalists could make money or satisfy their curiosity. The privileges are provided to help journalists serve the public interest, similarly to the delegation of decision power to politicians. People being people, some journalists abuse the privileges. They do not inform the public, only entertain to make money. This takes the form of sensationalism: covering frivolous topics that sell well, but do not provide useful knowledge. For example celebrity gossip, funny animals, manufactured controversy.

It would be fair to remove journalistic privileges from tabloid reporters and stop calling them journalists. They are just nosy people. The problem is that whoever decides on giving or taking privileges, gets power over the media. Therefore this authority should not be the government, but an independent organization. However, some rights and protections of the media are legislated, so a non-governmental body cannot change them. The legislature would have to delegate its authority in this sphere to the hypothetical independent regulator first. Given how much politicians wish to influence journalists, this decision seems unlikely.

Why taxes hit the middle class

The poor don’t have much wealth or income, so not much to tax. The rich find it worthwhile to hire accountants and lawyers to find loopholes useful for avoiding taxes. Optimizing one’s taxes is almost a fixed cost (does not depend much on the wealth or income), but the benefit is proportional to what is taxed. In other words, there are economies of scale in tax evasion. The middle class is not rich enough for professional loophole-seeking to be profitable, but has enough to tax.
This is only theoretical reasoning. Empirically, the average tax rate increases in income, as far as I know. Whether the rate of increase should be higher or lower than at present is a matter of political preference.
To avoid the unequal treatment based on the ability to hire tax advisors, the tax system should be simple. Changing income from one form to another (salary to dividends for self-employed for example), or changing the source (foreign corporation vs its domestic subsidiary) should not change the tax rate. Minimizing loopholes means no special deductions for belonging to select groups (farmers, pensioners, veterans, parents).
The existence of a profession for providing tax advice to ordinary people is a sign that the tax code is too complex. Those in a common salaried job should be able to understand and file their taxes themselves.

Publication delay provides incentives

From submitting a paper to a journal until getting the first referee reports takes about six months in economics. It is very rare to get accepted on the first try. Most papers are rejected, and an immediate acceptance implies having submitted to too weak a journal. Waiting for the referee reports on the revision and second revision takes another six plus a few months. This seems unnecessary (reading a paper does not take six months) and inefficient (creates delay in disseminating research results), but is used for incentives.
Delay discourages frivolous submissions. It forces authors to evaluate their own work with some honesty. If the referee reports were immediate, then everyone would start at the top journal and work their way down through every venue of publication until getting accepted. This would create a large refereeing and editing burden. Delay is a cost for the authors, because simultaneous submissions to multiple journals are not allowed. Trying for high-ranking journals is a risk, because the author may not have anything to show at the next evaluation. This reduces submissions to top journals. It may be optimal to start at the middle of the ranking where the chances of acceptance are higher.
A similar incentive to submit to the correct quality level of journal can be created by imposing a submission fee, forbidding further submissions for a period of time if rejected or requiring the author to write referee reports on others’ papers. A submission fee should be distinguished from publication fees, which are used at fake journals. The submission fee is paid no matter whether the paper is accepted, therefore does not create the incentive for the journal to lower its standards and publish more papers.
The submission fee would impose different costs on authors in different financial circumstances. Some have research funds to pay the fee, some do not. Similarly, delay has a larger effect on people whose evaluation is coming sooner. Being banned from a journal for a given amount of time after a rejection is worse for a researcher working in a single field. Interdisciplinary folk have a wider variety of venues to publish in. Writing referee reports as a price of having one’s work evaluated may lead to sloppy reviewing. Any mechanism to induce self-selection has a cost. Yet self-selection is needed.

Insurance in research

Most developed countries have programs to support research and encourage students to choose careers in it. This suggests scientists have a positive externality on the rest of their country that is not fully internalized in their income. Why not support research by paying the researchers its value, assuming the value can be measured? This would internalize the externality, leading to efficient effort provision.
A potential answer is different risk aversion of the organization supporting science and the scientists. If the institution is involved with many different projects, it is diversified and likely to be less risk averse than a researcher who only has a few projects. The arrangement optimal for both sides is then for the institution to offer insurance (at a cost). The researchers get paid a lower expected amount than the value of their work, but with a lower variance. Instead of the scientists taking loans to finance their work, becoming rich if the project succeeds and bankrupt if it fails, they avoid loans and get a fairly constant salary.
There is a tradeoff between incentives and insurance. If the salary does not depend on success, there is no incentive for effort, but perfect insurance. Having researchers take loans and get the full value of their work provides no insurance, but strong motivation. The compromise is that promotion and pay depend somewhat on research success, but not too much.