Archive for March, 2017

Outsiders

Recently, I’ve run across a number of articles, including some of the scholarly variety, which address the issue of “false news.” Several of them have made the points that so-called human “rationality” evolved to facilitate cooperation, not necessarily rational analyses of facts, and that the majority of human beings will accept “false news” that facilitates their inclusion in their belief group and reject verified facts that are in conflict with group beliefs.

If this is so, and from what I’ve observed, it seems to be for large groups of people, it gives rise to another question: How have human beings ever managed to evolve and develop a technological society?

The first response that came to my mind was: That’s why progress has been so slow and spotty, because you need consensus for a new way to become part of society.

One of the corollaries to this is that groups that more easily accept new and better changes will be able – still cooperatively – to outcompete groups or societies that don’t. And history tends to show that this is in fact true. When the Chinese culture essentially and gradually closed itself off to outside influences, symbolized by the government decree to destroy all ocean going ships in 1525 A.D., that marked the beginning of the long, slow, and inexorable decline of China, ending with the effective destruction of the “traditional” culture in the early twentieth century.

As I’ve noted in various previous blogs, a great number of more “modern” inventions, including the mechanical computer embodied in the antikithera mechanism, were actually developed and forgotten hundreds if not thousands of years before some society finally adopted them. Some were discarded because they were seen as uneconomic, others because people didn’t want to change existing ways of doing things, but what’s often overlooked is that economic factors aren’t entirely “rational,” but also part of a belief structure.

Some twenty-five years ago, my wife pointed out to an executive in the retail clothing world that there was a growing number of older women with money and taste who wanted professional and tasteful clothing not designed for twenty and thirty-year olds. The executive told her that there was no market for such clothing. I now know of several large retail firms making hundreds of millions, if not billions, from that market… but in the late 1980s and early 1990s, the group-belief in the clothing industry was that there was no market.

That’s where outsiders come in. They’re the people who are at least marginally part of society but who really aren’t part of a group, the ones who can set aside non-functional group “beliefs” and come up with changes.

Being an inside outsider can be dangerous, particularly in areas of belief. Many of the first theologians who started the movement that became the Reformation, like Jan Hus, ended up being executed. Although Alfred Wegener proposed the theory of continental drift in 1915, he was ignored and ridiculed for more than fifty years before conclusive evidence vindicated him.

But even inside outsiders can be trapped by believers, because, if they’re successful, they tend to attract a group of people who either share the same beliefs, or profess to share those beliefs, and, in time, that reduces the former outsider’s objectivity. Companies started by outsiders, such as Microsoft, Apple, or even Walmart, often have this problem after a while. One of Edison’s great advantages was that, for the most part, but not always, he relied on what he could prove or disprove, an attitude that often goes against group beliefs.

The problem, of course, is that for every outsider with a good idea, there are a dozen with bad ideas [which is often why they’re outsiders]. The fact that bad new ideas always outnumber good new ideas may also be why stable societies tend to be conservative. It’s also, I suspect, why when a society incorporates too much change too quickly the results are almost always disastrous, or close to it. Yet, without change, cultures stagnate and collapse… or are taken over or conquered by other cultures.

All of this is why I’m speculating, and it’s only a speculation, that societal/cultural success depends over the long run on the successful use and management of “outsiders” and their ideas.

Books – Getting There or Being There

There are so many different ways to categorize or analyze books that anything I write is likely to have been said or written many times before, but the other day something struck me, in an analytical sense, that I’ve known so intuitively that I never really ever verbalized it. It was simply that there are some books that one reads merely to get to the end in order to find out what happens, or who did what in what fashion, and there are others where each page is a delight, and one is disappointed when the book ends. The first kind of book is about “getting there,” and the second is more about “being there.”

Just as there are both kinds of books, and a great many that fall in the middle, readers also tend to fall along that spectrum as well.

Personally, I tend to like books that incorporate both aspects, and, obviously, I try my best to create both feelings, but my books, I suspect, tend to have a strong component of “being there,” and I’m reminded of that when I see reviews or comments by readers who complain about too little action or not enough battles or too many meal scenes.

But there’s more to “being there” than just language or description of mundane events. As a former Navy search and rescue pilot, I can’t help but recall the description of Naval Aviation that instructors brought up more than a few times – “ninety-nine percent routine boredom and one percent pure terror.” I also held a variety of fairly senior staff positions in national politics over nearly twenty years, from the Nixon Watergate years through Reagan years and some of the first Bush presidency. There were some tense moments there, about which the less said in this or any other public forum, the better, but the bottom line was the same. Pulse-pounding, heart-stopping action is rare and infrequent, as is political tension and true drama… and both are usually caused because someone’s screwed up the “being there” and routine parts of life [as we’re now seeing in the U.S. political arena at present].

That’s another reason why I write the way I do, because I like showing just how that can happen, and how disaster so often comes as a result of carelessness, thoughtlessness, lack of understanding, or incompetence in dealing with the routine. Seeing the protagonist fix the disaster, of course, is what most readers enjoy, but I’ve found in many novels that there’s little detail or creation involved in what causes the disaster, and that often there is problem after problem that, when considered for more than a moment, come off as improbable.

That’s why it helps to have at least at bit of “being there” because it makes the “getting there” more enjoyable and a deeper read… at least in my view.

The Wrong Healthcare Issue

Right now, the House Republicans are fighting to get enough votes to pass their bill to repeal and replace the Affordable Care Act, aka “Obamacare.” The Democrats are staunchly opposed. Both sides are arguing over the affordability of healthcare and access to healthcare insurance.

As far as I can see, they’re both circling around wrong tree, chasing each other’s tails. Insurance is only a symptom of the greater problem, and trying to deal with symptoms is not only expensive, but will also postpone dealing with the real problem, which continues to worsen. That problem? Healthcare costs. People need insurance because healthcare costs in the U.S. are effectively the highest in the world, and the vast majority of Americans don’t get as good healthcare as nations spending far less on healthcare.

In 2015, U.S. health care costs were $3.2 trillion, making healthcare one of the largest U.S. industries, nearly eighteen percent of Gross Domestic Product, but fifty-five years ago, healthcare only comprised five percent of GDP.

Part of the reason for the cost increase is emergency room treatment, the most expensive single aspect of current healthcare, making up one-third of all health care costs in America. And a significant proportion of emergency room care occurs because people can’t get or afford other treatment for various reasons.

Another component of rising costs is the continuing increase in the costs of drugs and medical devices. According to Forbes, the healthcare technology industry was the most profitable U.S. industry sector of all in 2015, notching an average profit margin of 21%, with the most profitable company of all being Gilead Sciences with a 53% profit margin. And no wonder, given that the list price for the top-20-selling drugs in the U.S. averages more than twice as much as those drugs as in the E.U. or Canada.

While the pharmaceutical industry pleads high research and development costs, a GlobalData study showed that the ten largest pharmaceutical companies in the world in 2013 spent a total of $86.9 billion on sales and marketing, as opposed to $35.5 billion on research and development, almost two and a half times as much on marketing as R&D. Those ten companies had an average profit margin of 19.4%, ranging individually from 10% to 43%, with half making 20% or more. And since Medicare is prohibited by law from negotiating drug prices for its 55 million beneficiaries, the program must pay whatever price drug makers set.

The U.S. medical technology market exceeds $150 billion a year in sales, and in 2015 the gross profit margin for the medical equipment and supplies industry averaged 12.1%, according to data from CSImarket.com.

Studies of doctors’ compensation show that over the past twenty years, that, in general, physician compensation has increased far less than all other components of healthcare. In fact, annual earnings actually declined for the typical physician between 2000 and 2010. Annual earnings for physician assistants and pharmacists have increased at a greater rate. More to the point, as a percentage of total national healthcare costs, U.S. physician wages are small – approximately 9% – a number among the lowest in the developed world.

Hospitals’ costs have increased significantly, but not because they’re making money. A Health Affairs study analyzed hospital income and costs of more than 3,000 hospitals nation-wide and found that fifty-five percent of hospitals lost money on each patient they served in 2013. This does raise the question of whether non-profit hospitals are paying more and more, possibly too much, for high-priced administrators apparently required by the bureaucratic and legal maze generated by the interweaving of private and public medical systems, government regulations, and insurance company requirements. Studies indicate that administrative costs make up twenty to thirty percent of the United States health care bill, far higher than in any other country. American insurers, meanwhile, spent $606 per person on administrative costs, more than twice as much as in any other developed country and more than three times as much as many, according to a study by the Commonwealth Fund.

Then add to that the skyrocketing costs of malpractice insurance and often excessive court judgments in medical tort claims cases.While the amount is subject to dispute, it’s not inconsiderable and also adds to costs.

Unfortunately, neither the Affordable Care Act nor any proposed Republican replacement will do anything to deal with what I’ve mentioned, and what I’ve mentioned are only the most obvious causes of ever-increasing health care costs.

Political Messaging

Everyone who follows media in the entire world likely knows that President Trump sends messages via Twitter. What’s been almost lost in the Twitter-storm, and the swirling claims and counter-claims about Russian influence in one form or another, is another, and far more ominous message.

In the United States, indeed anywhere, one form of “reality” is not what necessarily is, but what people believe is so. If people believe that foreign aid comprises twenty percent of federal spending or that public television and radio constitute five percent of the budget, then for them, that is reality, regardless of the facts. Unfortunately for actual reality, a majority of Trump supporters hold such beliefs, despite hard dollar figures to the contrary. So when Trump’s proposed budget proposes massive cuts to federal programs such as those whose total budgets in reality only comprise than five percent of federal spending, Trump’s followers truly believe that he is trying to make a significant cut in federal spending, while the observations by those who understand the numbers and the federal budget that such cuts cripple worthwhile programs while not really addressing the actual debt and deficit are largely ignored or minimized as just being politics as usual.

The problem is that Trump has no interest in confronting reality. His interest is, as is the interest of all promoters and snake-oil salesmen, to sell people on his version of reality, or to affirm their illusory version of reality to increase his own power and image. He also understands, as apparently the mainstream media doesn’t, that repetition turns anything into popular “truth.”

This is something that the mainstream media still doesn’t seem able or willing to confront. It’s one thing to argue about what national spending priorities should be. It’s another to put forth a spending plan designed solely to appease and appeal to one’s supporters, as Trump has, while totally ignoring fiscal reality. Unfortunately, even a sizable fraction of the GOP members of Congress seems unwilling to come to grips with this, and that’s understandable because Trump will turn on “defectors” and because a majority of Republicans also want to believe in Trump’s version of reality.

The media should be pointing out, daily, and loudly, that the numbers don’t add up. Have you seen a headline claiming “Trump Budget Based on Lies”? Or: “EPA Head, Oil Industry Cause 5,000 Earthquakes.” Or: “Trump Buys Off McConnell.” All of those are legitimate headlines, but you haven’t seen them, and you likely won’t, because if they show up, Trump will accuse them of being crooked liars, or the equivalent.

There seems to be a media assumption that people will see the truth on their own. Really? In a nation that requires remote controls for their televisions, ten second sound-bites, and news by Twitter, maximum 128 characters? Add to that the fact that most media publishes Trump’s proposals without strong critical analyses and worries more about his criticism than letting the public know what is occurring.

Establishing “truth” by repetition is currently winning… and all of us are losing.

Taking Credit

There are people who accomplish good or great things, and there are those who take credit for those accomplishments. As most intelligent individuals know, often the person who gets credit isn’t the one who actually did the work. Also, sometimes more than a few individuals take credit for something that was never accomplished or completed.

Over the course of my life I’ve certainly seen a lot of such instances. One of the best things – or the worse – about being a writer is that when a book is published you get the credit – or criticism. In my case, either, depending upon your point of view, is warranted, because I personally write every word that’s published, except for the few words corrected by my editor. I have been known to borrow/steal ideas from my wife, but the words are my own.

Not all books, however, are necessarily written by name on the spine of the book. While the original “Ellery Queen” mysteries were written by Frederic Dannay and Manfred Bennington Lee, more than twenty of the later Ellery Queen novels were ghost-written by others, including SF author Jack Vance.

Likewise, particularly in politics and often in business and academia, credit or blame is often taken by or placed on the wrong people. President Herbert Hoover didn’t cause the Great Depression, nor did Franklin Roosevelt end it [although he did make great efforts and did his best to mitigate its effects until the economic recovery caused by WWII kicked in]. Bill Clinton got credit for the economic recovery actually primed by the first President Bush.

Then there are the people who labor long and hard and slowly build something from virtually nothing, such as Fred Adams, who created the now well-known Utah Shakespeare Festival [which was good enough to win a Tony several years ago as the best regional theatre in the U.S.]. His wife Barbara did half the work, but only those who knew Fred and Barbara know that because Fred was not only a great builder, but a great showman. There’s also a well-known fantasy author whose wife contributed to every book, but whose name only appeared on the last few.

In more than a few cases, those who build an organization, a cause, a business just aren’t self-promoters, and often, because of that, others take credit… or the individual never gets credit.

More often than not, why someone gets credit, deserved or undeserved, is because they’re a good self-promoter, and there’s nothing wrong with that in itself – unless the self-promoter steals the credit from someone else. What is equally wrong is when the rest of us let the self-promoters who are stealing credit from those who deserve it reward the deceptive self-promoter.

National Identity and Anger

A recent poll from The Associated Press-NORC Center for Public Affairs Research showed that seventy percent of Americans felt that the country was “losing its identity.” Unfortunately, what the poll also revealed was that Americans couldn’t agree on what were the important components of that “identity.”

Although there are some points of agreement among Democrats, Republicans and independents about certain aspects of what makes up the country’s identity, such as a fair judicial system and rule of law, the freedoms enshrined in the Constitution, and the ability to get good jobs and achieve the American dream, recent political developments make it clear that the consensus on these points is overshadowed by the differences.

Fifty-seven percent of Republicans thought one of the most important parts of the national identity was a Christian belief structure, as opposed to twenty nine percent of Democrats. On the other hand, sixty-five percent of Democrats thought that that the mixing of global cultures in the U.S. was important, compared to thirty-five percent of Republicans.

According to the poll, seventy-four percent of Democrats say that the ability of immigrants to come to the U.S. to escape violence and persecution is very important, as opposed to fifty-five percent of Republicans. Forty-six percent of Republicans agreed the culture of the country’s early European immigrants was very important, versus twenty-five percent of Democrats.

Putting these findings together suggests that, in general, Republicans think that the national identity should be based on an enshrined Christian faith and the Anglo-centric patriarchal culture of the first immigrants, while Democrats emphasize a more global-culture, welcoming to immigrants, and more concerned with the present than the past. Obviously, that’s an oversimplification, but there’s still a basic conflict, almost between the past and the present.

That conflict was definitely revealed in the last election, with the Republicans essentially claiming that the country was turning from its white, European, and totally Christian roots, and that such a turn was destroying and/or diminishing not only the United States, but the position of middle-class white American males.

As both the AP-NORC Poll and the Women’s March on Washington [with millions of women in hundreds of cities and towns across the country] showed this Republican “traditional” society is not endorsed by a significant percentage of the country.

Yet the Founding Fathers attempted to hold together thirteen colonies of very different belief structures, some with the [to me] abhorrent idea that slavery was morally acceptable, and they crafted a government based on shared principles that did not require a specific religious belief, or indeed, any belief in a supreme deity at all. For the time, this was an extraordinarily radical enterprise, so radical that the American Revolution equally merits the title of the Anglo-American Civil War.

So why is there so much disagreement about national identity and national priorities?

The election results and the vitriolic rhetoric from the right reflect, among other things, that there are fewer and fewer well-paid unskilled and semi-skilled jobs, and those jobs already lost to out-sourcing and technology, but mainly to technology, removed some eight million largely white men from the middle class. Those men and their families and relatives look to a past of more secure and prosperous employment and believe that the country has lost its way… and its traditional identity, and they’re angry.

On the other hand, there are over forty million African Americans in the U.S., and while the Civil War that resulted in their freedom ended over 150 years ago, those blacks still face discrimination and other barriers to rights equal to other white ethnicities. After 150 years they’re angry, and getting angrier, especially given the number of young black males killed and incarcerated, particularly when study after study shows discrimination still exists and that blacks receive harsher jail sentences for the same offense as do whites… among other things.

Educated women of all ethnicities are angry that they do not receive even close to equal pay for the same jobs as men and that the male-imposed glass ceilings in business, government, and politics still remain largely unbroken.

Since women and minorities are getting more and more vocal, and since minorities are becoming a bigger and bigger share of the American population, I foresee some very “interesting” years ahead, and I’d suggest that the largely white male Congress consider those facts very carefully.

The “Other” Culture

There are several definitions of “culture.” One is the development of microorganisms in an artificial media. Another is “the refinement of mind, morals, or taste.” A third is “the cultivation of plants or animals.” But there are two other definitions that tend to get overlooked: (1) the specific period or stage in the development of a civilization and (2) the sum total of the attainment and learned behavior patterns of any specific period or group of people regarded as expressing a way of life. The second of those latter definitions is the one that tends to get overlooked in government and politics, and yet the problems caused by the “learned behavior patterns” of smaller groups within a society represent one of the principal reasons for societal unrest.

That is largely because quite a few nations, including the United States, are in fact composed of various subcultures. In the U.S., those subcultures, especially those disliked by the majority, are often minimized or denigrated in racial or religious terms. An important point, and one consistently ignored by ideologues, businesses, and particularly politicians, is that “culture,” as exemplified by learned patterns of behavior, trumps “race” or religion. By that I mean that the good or bad traits of group or subgroup of people have virtually nothing to do with their religion or their skin color or ethnicity. What determines how people act is their “learned patterns of behavior.”

And while religion is definitely a learned behavior, how people of a certain religion act can and does vary enormously from cultural group to cultural group. It also varies over time. Some 500 years ago, good “Christian” countries in Europe were slaughtering each other in a fashion even more brutal than that in which the Sunni and Shia factions of Islam are now doing. Yes, religion is a critical part of “culture,” but it ranges from being the primary determinant of a culture to being merely one of many factors, and in the history of certain civilizations, the impact of a religion can and has changed the culture drastically.

As I’ve also noted before, likely more than a few times, history is filled with examples of both great and failed societies and nations identified as being predominantly of one race or religion. There have been great empires in all parts of the world – except, so far, Antarctica, and there have been failed societies everywhere in the world, regardless of race or religion.

Certain cultural practices seem to work better than others, one of which is that cultures that allow religion to control society tend to stagnate and become ever more brutal. Cultures with great income inequality tend to be more likely to be oppressive, and a greater percentage seem to have either de jure or de facto polygamy. A good sociologist could likely carry this much farther, but the basic point is that it’s not only morally wrong to claim that a given race or ethnicity or religion is “stupid” or “inferior” (or any other number of pejorative terms), but also such unthinking “type-casting” totally misses the point. Culture – not race, genes, skin color, or religion – determines how people behave. More to the point, one can change a toxic culture [although it takes time] and a beneficial culture is always only a cultural change or two away from becoming toxic.

The Threat from Radical Islamic Terrorists

I’m fed up with the propaganda from the White House about the “overwhelming” danger to U.S. citizens from radical Islamic terrorists. Yes, there are radical Islamic terrorists, but here in the United States, radical Islamic terrorists are far less of a threat than home-grown right-wing terrorists. Overseas, that’s another question, which is why so many law-abiding members of the Islamic faith want to get to the U.S. – or did before Donald Trump became President.

While consensus on hard numbers is difficult to come by, and numbers vary by source, whatever the source, those numbers suggest that radical Islamic terrorists are not the major threat to Americans – not even close. Other Americans are.

On terms of terrorist attacks in the United States, the numbers are lopsided, to say the least. According to a study by the United States Military Academy’s Combating Terrorism Center, domestic right-wing extremists averaged 337 attacks in the U.S. in the decade after 9/11, accounting for 254 fatalities, while, depending on the study and the definitions, between a total of 20 and 24 terrorist attacks in the U.S. were carried out by Islamic radicals with between and 50 and 123 fatalities.

In the last several years, the vast majority of police deaths have come from domestic extremists. An ADL report states that of the forty-five police officers killed by extremists since 2001, ten were killed by left-wing U.S. extremists, thirty-four by right-wing U.S. extremists, and one by domestic Islamic extremists.

As far as Trump’s proposed travel ban goes, there has not been one single terrorist attack on U.S. soil in the last four decades that has been carried out by citizens of any the seven countries on Trump’s ban list. Out of the 240,000 Americans who have been murdered since the attacks on the Twin Towers in 2001, exactly 123 of those deaths were linked to Muslim-American extremists. In other words, .05123 percent of the murders in the United States in a sixteen year period were carried out in the name of radical Islam. Even figures from the right-wing National Review only list 88 deaths in the U.S. from radical Islamic terrorists since 2009.

Yet at the same time that Trump is citing the danger from radical Islamic terrorists, reports have surfaced that he plans to shut down Homeland Security’s watch list for domestic extremists. Not only that, but bowing to the NRA, he decided to void an Executive Order by former President Obama that would have put people declared to be mentally incompetent by a court on a do-not-buy list for firearms. The NRA argued that mentally incompetent people should have the same right to firearms as anyone else.

And we’re worrying about Islamic terrorists?

Education and the Business Model

More and more state legislators across the country are demanding that education be run in a more business-like fashion. While greater efficiency is likely necessary in many educational systems, running higher education “like a business” is not only counter-productive, but it’s more likely to create more problems than it purports to solve – and business-like approaches so far haven’t shown much success at the university level.

One of the tools employed by both business and educational systems run by the “business model” is to reduce costs by reducing the number of employees and their compensation in relation to “output.” In the business area, this has given us outsourced manufacturing or high-tech automated manufacturing or, on the other hand, in retailing, lots and lots of underpaid, part-time employees without benefits. In education, a similar change is occurring, particularly in higher education, where university faculties have shifted from those primarily comprised of full-time dedicated professors to faculties where the majority of teaching faculty are part-time adjuncts, many of them far less qualified or experienced than seasoned full-time faculty. At the same time, administrations spouting the “business model” mantra have burgeoned.

At virtually all public universities, administrative overhead and full-time administrative positions have increased two to threefold over the past twenty plus years, while full-time faculty positions have decreased, except in the case of some smaller state universities that have expanded massively so that full-time positions have increased somewhat, even though the percentage of full-time positions has decreased to the same level as at other state universities, if not more.

The chief reason for this emphasis on part-time teaching positions is cost. As I’ve noted before, fifty years ago, on average, state legislatures supplied the bulk of higher education funding. In 1974, states provided 78% of the cost of educating a student. Today, although total funding is actually higher, because almost four times as many students attend college today, the amount of state funding per student averages around 20%, although it varies widely by state, and in some cases it is around 10%.

For example, almost until 1970, California residents could attend the University of California [Berkeley] tuition-free. Today, tuition and fees for in-state students are around $15,000 a year. This trend, if anything, is accelerating. Just since 2008, state funding of higher education has dropped by 20% per student

The response by legislatures, predictably, is to push for more efficiency. Unhappily that has translated into “get lower costs however you can.” The problem with this is that the emphasis, no matter what administrators say, is to turn out the most graduates at the lowest cost. Universities also tend to phase out departments with small numbers or high costs, and expand departments with large numbers and low costs, even if students that major in that area have difficulty getting jobs.

In addition, political pressure, both to “keep” students in school for budgetary reasons and to graduate a higher percentage of students, has inexorably decreased the academic rigor of the majority of publicly funded universities and colleges. This, in turn, has led to more and more businesses and other employers demanding graduate degrees or other additional qualifications, which further increases the tuition and debt burden on students. That’s scarcely “economic” on a societal basis because it pressures students to aim for high income professions or high income specialties in a profession, rather than for what they’re good at doing and what they love. It’s also created an emphasis on paper credentials, rather than the ability to do a job. On top of that, it’s meant more highly qualified individuals are avoiding professions such as teaching, library science, music, art, government civil service, and others; and those professions, especially teaching, are being filled by a greater percentage of less highly qualified individuals.

The end result of the combination of stingy state legislatures and the “business model” is less rigorous academic standards and watered down curricula at the majority of public colleges and universities, skyrocketing student debt, a smaller and smaller percentage of highly qualified, excellent, and dedicated full-time professors, and a plethora of overpaid administrators, the majority of whom heap even more administrative requirements on full-time teaching faculty.

No efficient business actually operates this way, and why higher education gets away with calling what it’s doing “the business model” has baffled me for more than two decades.