Archive for the ‘General’ Category

Harassment and Scandal

Bill O’Reilly is now out at Fox News, following by a few months the ouster of Roger Ailes, each removal the result of the revelation of a long and continuing pattern of inappropriate sexual behavior including sexual harassment. Does such behavior, as well as the long-term retention of two such individuals by the highly conservative Fox organization, have anything to do with the political outlook of Republicans and conservatives?

Certainly, a great number of liberals think so, especially some in my own family, but are conservatives really more likely to behave badly in the sexual/gender area than are liberals?

I can name a number of liberals and Democrats who have engaged in what most would call sexual improprieties, going all the way back to President Grover Cleveland, who fathered a child out of wedlock, or Franklin Roosevelt who had several affairs while in public office, or John Kennedy, or Bill Clinton. On the Republican side, the most obvious were Warren Harding who had a fifteen year affair with Carrie Fulton Phillips, Dwight Eisenhower who had a brief affair with his military driver, or Nelson Rockefeller,who had quite a few indiscretions, but the list of sexual political improprieties among national political figures is long and it includes roughly the same numbers of Republicans and Democrats. What it doesn’t include is many women [I could only find one out of more than 100 Republicans and Democrats listed for sexual crimes and improprieties], which suggests that women in power are either far less likely to engage in sexual indiscretions or less likely to be found out.

On the issue of harassment, however, conservatives and Republicans seem to do more of that, or at least they’ve been found out and charged with it more often. In addition to the Ailes and O’Reilly cases, there was Senator Bob Packwood (R-ORE) who resigned in 1995 under a threat of public senate hearings related to 10 female ex-staffers accusing him of sexual harassment. As candidates for the Republican Presidential nomination, both Donald Trump and Herman Cain were charged with sexual harassment. Clarence Thomas was also accused of sexual harassment. Then there was the Oklahoma state representative who used state funds to pay off a judgment against him for sexually harassing a staffer, or the Texas Congressman who tried to gut the Office of Congressional Ethics, who had earlier been charged with sexual harassment by an employee whom he later fired, or the Wisconsin state assembly Republican majority leader who was convicted of sexual assault, or… the list is very long.

Democrats certainly aren’t blameless, starting with Teddy Kennedy, former Senator Brock Adams, Congressmen Tim Mahoney, Jim Bates, and Mel Reynolds, but the list of actual Democratic harassers is about a fourth the size and length of the list of Republican harassers.

And then I came across an interesting chart of criminal misconduct by Presidential administration. Since Richard Nixon, whose administration resulted in 76 criminal indictments, 55 convictions and 15 prison sentences for members of his administration, there have been four Republican administrations and three Democratic administrations. The Democratic administrations had three criminal indictments, one conviction, and one prison sentence. The Republican administrations had 44 criminal indictments, 34 convictions, and 19 prison sentences.

The way it seems to stack up is that political viewpoint doesn’t make much difference in terms of consensual or semi-consensual sexual indiscretions, but the Republican/conservative outlook seems to result in more abuses of power and position.

But, from what we’ve seen recently, is that really surprising? Or is it that Democrats are really better politicians and are better at sexual persuasion?

All That Different?

Because human beings don’t have chlorophyll and a few other physio-chemical adaptations, for us to survive, we need to eat either other forms of life or the products of other forms of life. We’ve bred forms of both plant and animal life to provide food for us, and we’ve become better and better at it.

But there’s an underlying assumption behind our agricultural achievements, and that assumption is that human beings are not only superior to other forms of life on earth, but that we are fundamentally different in the way we interact with our environment.

One of the early beliefs was that human beings were the only tool-users on the planet. Now, after a raft of studies over the past fifty years or so, we know that there are quite a few other species that make and use tools. While those tools are incredibly crude compared to our tools, they are tools, and for a species to make and use a tool requires a certain amount of thinking and forethought beyond blind instinct or environmentally programmed responses. We’ve also discovered that animal tool use is, at least in a number of cases, “cultural,” in that some groups of a species use tools and others don’t, or make different tools.

Then came the questions dealing with whether animals could actually think, especially in dealing with “theory of mind” matters, that is, is the ability to attribute beliefs, intents, desires, pretending, knowledge, etc., to oneself and others and to understand that others have beliefs, desires, intentions, and perspectives that are different from one’s own. Experiments with mirrors and images have shown that certain species do indeed have that ability. Crows, ravens, elephants, and certain primates behave in ways that show they are very much aware of possible differences and mental motivation and states of others of their species and sometimes, even of other species.

But what we’ve learned doesn’t stop there. For a long time, most biologists dismissed the idea that plants did anything but grow and reproduce in some fashion. In the last few decades, however, they’ve discovered that plants aren’t nearly as simple as once had been thought. Experiments have shown that plants of the same species communicate with each other, and can warn other plants about insect attacks and other changes in the environment. They can also muster defenses against certain attacks. Unhappily, at times these defenses can be fatal if the attackers also adapt, as in the case of the spruce and pine bark beetles, who are attracted to both the warning signals and pitch secreted by the trees in an effort to repel beetles.

At the same time, more and more experiments and evidence show that plants do learn and adapt to changes in their environment. An evolutionary ecologist at the University of Western Australia, Monica Gagliano, actually trained plants to grow in specific directions based on which way a fan blew.

What’s the bottom line of all this? That while human beings are currently the best tool-users and thinkers on the planet, we’re not the only ones, and that we’re not fundamentally different from the rest of life, just better at taking advantage of all other life-forms – except maybe bacteria and viruses, but that’s another blog.

Plastic Perfect

On Tuesday, I laughed, if ruefully, at one of the headlines in the local paper – “Plastic Surgery High in Utah” – especially after reading the article, in which researchers noted that Utah had one of the highest rates of cosmetic plastic surgery, especially breast implants and “tummy tucks.” The researchers did observe that plastic surgery rates are greater in areas where women’s higher education levels lag more behind that of men than the national average, and one was even bold enough to suggest that it might have something to do with the Mormon faith, and the emphasis on “female perfection.”

Might have something to do with the LDS faith? Is that an understatement! This is the state where the rate of Prozac usage by married women is the highest in the nation. This is the only state where the achievement of higher education rates by women has essentially hit a stone wall, or ceiling – call it the LDS celestial glass ceiling. And, after all, with all those women having five children and their husbands still clamoring for Barbie-doll-figures, how could women not feel pressured into having a tummy-tuck? Or certain other “enhancements”?

As I’ve noted before, I walk, with occasional short stretches of running, most mornings, and the time I set out varies by as much as two hours, but no matter what time I walk, whether it’s at 6:30 or 8:30, or occasionally later, who do I see walking and running? Women, and most of them are decades younger than I am, often pushing baby strollers of the type suited to being propelled by more than walking speeds. Gym memberships are predominantly female as well. I do see a very few men, but those few are gray or white haired, likely out there on doctor’s orders.

But bring this up among the “faithful,” just like the “holy number” of preferred children, and it’s emphatically denied, even as the cult of the plastic perfect continues to dominate the lives of young LDS women.

Who Knew…

That the first true Greek language came about because someone wanted to write down the orally transmitted works of Homer, but couldn’t because none of the existing languages in that part of the Mediterranean had any vowels – and you can’t accurately transcribe poetry [or song, either, my wife the voice professor informs me] without vowels. So this original transcriber (according to Archaeology magazine) took the vowelless Phoenician alphabet and added Greek vowels, and within a hundred years ancient Greeks became literate on a wide scale.

Now, for purists, there were two prior Greek languages, known as Linear A, which has yet to be deciphered/translated, and Linear B, but it is likely Linear A was without vowels, and Linear B was definitely without vowels and was exclusively used by a very limited number of bureaucrats and merchants for record-keeping, primarily of commodities and taxes. Definitely not for poetry or literature, or even science fiction or fantasy. Roughly a hundred to two hundred years before the Greek introduction of vowels, the same transition took place in ancient Israel [so, yes, the Jews were first to add vowels to the Phoenician alphabet, but word, literally, traveled slowly in those days].

Apparently, the Greek version of language with vowels was more effective than the Hebrew version, possibly because even then entertainment topped scripture, but that also might have been because Alexander the Great conquered more territory and imposed Greek on more people. The Romans, the great practical engineers, adopted/stole everything Greek, including the idea of vowels, but streamlined and simplified the alphabet in the Latinate letters that the majority of the world uses today.

And that’s why, when I include poetry and flowery language in my books, to the dismay of the action-preferred readers, everyone can read it… all because [take your pick], ancient Hebrews wanted more descriptive language in their scriptures or ancient Greeks wanted to be able to preserve the works of Homer.

Road Kill

A report released last month by the Governors Highway Safety Association shows that the number of pedestrians killed in traffic jumped eleven percent last year, to nearly 6,000, the largest single-year increase in pedestrian fatalities ever, and the highest number in more than two decades.

And this wasn’t just because the number of traffic deaths went up due to increased driving. While overall traffic deaths increased six percent in 2016, reversing slightly a ten year decline, pedestrian deaths increased by nearly 12%. But it wasn’t just in those two years. Since 2006, pedestrian deaths have increased from 11% of all traffic fatalities in 2006 to over 15% in 2016, an increase of 25%. The increase in pedestrian deaths over the past decade occurred at time when total traffic deaths dropped by almost 17%. According to a number of sources, the greatest component of this increase is distracted walking.

Over the past year or so, I’ve occasionally commented on the increasing functional stupidity of students and others who blithely cross streets, their heads in their cell phones, not paying attention to traffic or much else. Well… now there’s some evidence that there is a cost to such stupidity, and that those who engage in it are candidates for the Darwin Awards, whose not-quite-tongue-in-cheek criterion for receiving the award states, “In the spirit of Charles Darwin, the Darwin Awards commemorate individuals who protect our gene pool by making the ultimate sacrifice of their own lives. Darwin Award winners eliminate themselves in an extraordinarily idiotic manner, thereby improving our species’ chances of long-term survival.”

Lack of intelligence around moving vehicles isn’t, unhappily, confined to homo sapiens, as a recent report in Royal Society Open Science confirms, by noting that the highest percentage of birds killed by moving vehicles were those with the smallest brains relative to their overall size.

In short, small brains makes it more likely that birds will die as road kill.

I have to wonder if we’d find the same thing if we looked at pedestrian traffic deaths.

Equal Pay

On January 29, 2016, the Obama Administration proposed a change to EEOC reporting requirements. Currently, all employers with 100 or more workers, roughly 60,000 employers with 63 million employees, already complete the EEO-1 form on an annual basis, providing demographic information to the government about race, gender, and ethnicity, but the proposed change would require employers to complete a revised EEO-1 form that included salary and pay information.

Almost immediately, the business community objected, claiming that the additional information was unnecessary, useless, and a burden. The EEOC made revisions to the proposal, which included defining “pay” as the total W-2 compensation paid to an employee, since businesses already have to compile and report that figure, and issued the revised rule in September, 2016, while extending the compliance date from March 2017 to March 31, 2018.

Business interests, led by the U.S. Chamber of Commerce are pressing the Trump administration hard to revoke the rule, saying that there’s no merit in the requirement. Trump’s Director of the Office of Management now says the matter is under review.

This is despite a huge amount of data that would appear to indicate the opposite, that, in particular, there is significant overall pay discrimination based on race and gender. The difficulty is that while statistics show that women are paid roughly twenty percent less than men, those are aggregate statistics, and both sides dispute them for different reasons.

What I find interesting is the Chamber of Commerce statement that the data would be useless. It seems to me that the data could be incredibly useful. It would go a very long way to either establishing or rejecting the idea that gender and racial pay discrimination exists.

In earlier comments, some businesses objected to the use of W-2 total compensation in the report, claiming that “base pay” was more accurate. Equal Pay advocates countered by pointing out that bonuses and other additional compensation go far more often to white males, and that total compensation – the measure adopted in the final rule – was a more accurate indicator.

The Chamber of Commerce’s opposition, at least to me, smacks of trying to keep everyone in the dark about what’s happening in the pay area, especially since business has to make the basic report anyway. It’s similar to the idea that, if the government stops funding climate research, global warming will just go away… but then, the head in the sand attitude has always been a favorite of those who don’t want things to change.

NOTE: At court hearing last Friday [April 7th], a U.S. Department of Labor regional director announced that, in investigating Google, the DOL had “found systemic compensation disparities against women pretty much throughout the entire workforce.” Google, of course, vehemently denied the charges. This was the second Silicon Valley tech company that DOL had charged with such gender pay discrimination, the first being Oracle earlier this year.

Outsiders

Recently, I’ve run across a number of articles, including some of the scholarly variety, which address the issue of “false news.” Several of them have made the points that so-called human “rationality” evolved to facilitate cooperation, not necessarily rational analyses of facts, and that the majority of human beings will accept “false news” that facilitates their inclusion in their belief group and reject verified facts that are in conflict with group beliefs.

If this is so, and from what I’ve observed, it seems to be for large groups of people, it gives rise to another question: How have human beings ever managed to evolve and develop a technological society?

The first response that came to my mind was: That’s why progress has been so slow and spotty, because you need consensus for a new way to become part of society.

One of the corollaries to this is that groups that more easily accept new and better changes will be able – still cooperatively – to outcompete groups or societies that don’t. And history tends to show that this is in fact true. When the Chinese culture essentially and gradually closed itself off to outside influences, symbolized by the government decree to destroy all ocean going ships in 1525 A.D., that marked the beginning of the long, slow, and inexorable decline of China, ending with the effective destruction of the “traditional” culture in the early twentieth century.

As I’ve noted in various previous blogs, a great number of more “modern” inventions, including the mechanical computer embodied in the antikithera mechanism, were actually developed and forgotten hundreds if not thousands of years before some society finally adopted them. Some were discarded because they were seen as uneconomic, others because people didn’t want to change existing ways of doing things, but what’s often overlooked is that economic factors aren’t entirely “rational,” but also part of a belief structure.

Some twenty-five years ago, my wife pointed out to an executive in the retail clothing world that there was a growing number of older women with money and taste who wanted professional and tasteful clothing not designed for twenty and thirty-year olds. The executive told her that there was no market for such clothing. I now know of several large retail firms making hundreds of millions, if not billions, from that market… but in the late 1980s and early 1990s, the group-belief in the clothing industry was that there was no market.

That’s where outsiders come in. They’re the people who are at least marginally part of society but who really aren’t part of a group, the ones who can set aside non-functional group “beliefs” and come up with changes.

Being an inside outsider can be dangerous, particularly in areas of belief. Many of the first theologians who started the movement that became the Reformation, like Jan Hus, ended up being executed. Although Alfred Wegener proposed the theory of continental drift in 1915, he was ignored and ridiculed for more than fifty years before conclusive evidence vindicated him.

But even inside outsiders can be trapped by believers, because, if they’re successful, they tend to attract a group of people who either share the same beliefs, or profess to share those beliefs, and, in time, that reduces the former outsider’s objectivity. Companies started by outsiders, such as Microsoft, Apple, or even Walmart, often have this problem after a while. One of Edison’s great advantages was that, for the most part, but not always, he relied on what he could prove or disprove, an attitude that often goes against group beliefs.

The problem, of course, is that for every outsider with a good idea, there are a dozen with bad ideas [which is often why they’re outsiders]. The fact that bad new ideas always outnumber good new ideas may also be why stable societies tend to be conservative. It’s also, I suspect, why when a society incorporates too much change too quickly the results are almost always disastrous, or close to it. Yet, without change, cultures stagnate and collapse… or are taken over or conquered by other cultures.

All of this is why I’m speculating, and it’s only a speculation, that societal/cultural success depends over the long run on the successful use and management of “outsiders” and their ideas.

Books – Getting There or Being There

There are so many different ways to categorize or analyze books that anything I write is likely to have been said or written many times before, but the other day something struck me, in an analytical sense, that I’ve known so intuitively that I never really ever verbalized it. It was simply that there are some books that one reads merely to get to the end in order to find out what happens, or who did what in what fashion, and there are others where each page is a delight, and one is disappointed when the book ends. The first kind of book is about “getting there,” and the second is more about “being there.”

Just as there are both kinds of books, and a great many that fall in the middle, readers also tend to fall along that spectrum as well.

Personally, I tend to like books that incorporate both aspects, and, obviously, I try my best to create both feelings, but my books, I suspect, tend to have a strong component of “being there,” and I’m reminded of that when I see reviews or comments by readers who complain about too little action or not enough battles or too many meal scenes.

But there’s more to “being there” than just language or description of mundane events. As a former Navy search and rescue pilot, I can’t help but recall the description of Naval Aviation that instructors brought up more than a few times – “ninety-nine percent routine boredom and one percent pure terror.” I also held a variety of fairly senior staff positions in national politics over nearly twenty years, from the Nixon Watergate years through Reagan years and some of the first Bush presidency. There were some tense moments there, about which the less said in this or any other public forum, the better, but the bottom line was the same. Pulse-pounding, heart-stopping action is rare and infrequent, as is political tension and true drama… and both are usually caused because someone’s screwed up the “being there” and routine parts of life [as we’re now seeing in the U.S. political arena at present].

That’s another reason why I write the way I do, because I like showing just how that can happen, and how disaster so often comes as a result of carelessness, thoughtlessness, lack of understanding, or incompetence in dealing with the routine. Seeing the protagonist fix the disaster, of course, is what most readers enjoy, but I’ve found in many novels that there’s little detail or creation involved in what causes the disaster, and that often there is problem after problem that, when considered for more than a moment, come off as improbable.

That’s why it helps to have at least at bit of “being there” because it makes the “getting there” more enjoyable and a deeper read… at least in my view.

The Wrong Healthcare Issue

Right now, the House Republicans are fighting to get enough votes to pass their bill to repeal and replace the Affordable Care Act, aka “Obamacare.” The Democrats are staunchly opposed. Both sides are arguing over the affordability of healthcare and access to healthcare insurance.

As far as I can see, they’re both circling around wrong tree, chasing each other’s tails. Insurance is only a symptom of the greater problem, and trying to deal with symptoms is not only expensive, but will also postpone dealing with the real problem, which continues to worsen. That problem? Healthcare costs. People need insurance because healthcare costs in the U.S. are effectively the highest in the world, and the vast majority of Americans don’t get as good healthcare as nations spending far less on healthcare.

In 2015, U.S. health care costs were $3.2 trillion, making healthcare one of the largest U.S. industries, nearly eighteen percent of Gross Domestic Product, but fifty-five years ago, healthcare only comprised five percent of GDP.

Part of the reason for the cost increase is emergency room treatment, the most expensive single aspect of current healthcare, making up one-third of all health care costs in America. And a significant proportion of emergency room care occurs because people can’t get or afford other treatment for various reasons.

Another component of rising costs is the continuing increase in the costs of drugs and medical devices. According to Forbes, the healthcare technology industry was the most profitable U.S. industry sector of all in 2015, notching an average profit margin of 21%, with the most profitable company of all being Gilead Sciences with a 53% profit margin. And no wonder, given that the list price for the top-20-selling drugs in the U.S. averages more than twice as much as those drugs as in the E.U. or Canada.

While the pharmaceutical industry pleads high research and development costs, a GlobalData study showed that the ten largest pharmaceutical companies in the world in 2013 spent a total of $86.9 billion on sales and marketing, as opposed to $35.5 billion on research and development, almost two and a half times as much on marketing as R&D. Those ten companies had an average profit margin of 19.4%, ranging individually from 10% to 43%, with half making 20% or more. And since Medicare is prohibited by law from negotiating drug prices for its 55 million beneficiaries, the program must pay whatever price drug makers set.

The U.S. medical technology market exceeds $150 billion a year in sales, and in 2015 the gross profit margin for the medical equipment and supplies industry averaged 12.1%, according to data from CSImarket.com.

Studies of doctors’ compensation show that over the past twenty years, that, in general, physician compensation has increased far less than all other components of healthcare. In fact, annual earnings actually declined for the typical physician between 2000 and 2010. Annual earnings for physician assistants and pharmacists have increased at a greater rate. More to the point, as a percentage of total national healthcare costs, U.S. physician wages are small – approximately 9% – a number among the lowest in the developed world.

Hospitals’ costs have increased significantly, but not because they’re making money. A Health Affairs study analyzed hospital income and costs of more than 3,000 hospitals nation-wide and found that fifty-five percent of hospitals lost money on each patient they served in 2013. This does raise the question of whether non-profit hospitals are paying more and more, possibly too much, for high-priced administrators apparently required by the bureaucratic and legal maze generated by the interweaving of private and public medical systems, government regulations, and insurance company requirements. Studies indicate that administrative costs make up twenty to thirty percent of the United States health care bill, far higher than in any other country. American insurers, meanwhile, spent $606 per person on administrative costs, more than twice as much as in any other developed country and more than three times as much as many, according to a study by the Commonwealth Fund.

Then add to that the skyrocketing costs of malpractice insurance and often excessive court judgments in medical tort claims cases.While the amount is subject to dispute, it’s not inconsiderable and also adds to costs.

Unfortunately, neither the Affordable Care Act nor any proposed Republican replacement will do anything to deal with what I’ve mentioned, and what I’ve mentioned are only the most obvious causes of ever-increasing health care costs.

Political Messaging

Everyone who follows media in the entire world likely knows that President Trump sends messages via Twitter. What’s been almost lost in the Twitter-storm, and the swirling claims and counter-claims about Russian influence in one form or another, is another, and far more ominous message.

In the United States, indeed anywhere, one form of “reality” is not what necessarily is, but what people believe is so. If people believe that foreign aid comprises twenty percent of federal spending or that public television and radio constitute five percent of the budget, then for them, that is reality, regardless of the facts. Unfortunately for actual reality, a majority of Trump supporters hold such beliefs, despite hard dollar figures to the contrary. So when Trump’s proposed budget proposes massive cuts to federal programs such as those whose total budgets in reality only comprise than five percent of federal spending, Trump’s followers truly believe that he is trying to make a significant cut in federal spending, while the observations by those who understand the numbers and the federal budget that such cuts cripple worthwhile programs while not really addressing the actual debt and deficit are largely ignored or minimized as just being politics as usual.

The problem is that Trump has no interest in confronting reality. His interest is, as is the interest of all promoters and snake-oil salesmen, to sell people on his version of reality, or to affirm their illusory version of reality to increase his own power and image. He also understands, as apparently the mainstream media doesn’t, that repetition turns anything into popular “truth.”

This is something that the mainstream media still doesn’t seem able or willing to confront. It’s one thing to argue about what national spending priorities should be. It’s another to put forth a spending plan designed solely to appease and appeal to one’s supporters, as Trump has, while totally ignoring fiscal reality. Unfortunately, even a sizable fraction of the GOP members of Congress seems unwilling to come to grips with this, and that’s understandable because Trump will turn on “defectors” and because a majority of Republicans also want to believe in Trump’s version of reality.

The media should be pointing out, daily, and loudly, that the numbers don’t add up. Have you seen a headline claiming “Trump Budget Based on Lies”? Or: “EPA Head, Oil Industry Cause 5,000 Earthquakes.” Or: “Trump Buys Off McConnell.” All of those are legitimate headlines, but you haven’t seen them, and you likely won’t, because if they show up, Trump will accuse them of being crooked liars, or the equivalent.

There seems to be a media assumption that people will see the truth on their own. Really? In a nation that requires remote controls for their televisions, ten second sound-bites, and news by Twitter, maximum 128 characters? Add to that the fact that most media publishes Trump’s proposals without strong critical analyses and worries more about his criticism than letting the public know what is occurring.

Establishing “truth” by repetition is currently winning… and all of us are losing.

Taking Credit

There are people who accomplish good or great things, and there are those who take credit for those accomplishments. As most intelligent individuals know, often the person who gets credit isn’t the one who actually did the work. Also, sometimes more than a few individuals take credit for something that was never accomplished or completed.

Over the course of my life I’ve certainly seen a lot of such instances. One of the best things – or the worse – about being a writer is that when a book is published you get the credit – or criticism. In my case, either, depending upon your point of view, is warranted, because I personally write every word that’s published, except for the few words corrected by my editor. I have been known to borrow/steal ideas from my wife, but the words are my own.

Not all books, however, are necessarily written by name on the spine of the book. While the original “Ellery Queen” mysteries were written by Frederic Dannay and Manfred Bennington Lee, more than twenty of the later Ellery Queen novels were ghost-written by others, including SF author Jack Vance.

Likewise, particularly in politics and often in business and academia, credit or blame is often taken by or placed on the wrong people. President Herbert Hoover didn’t cause the Great Depression, nor did Franklin Roosevelt end it [although he did make great efforts and did his best to mitigate its effects until the economic recovery caused by WWII kicked in]. Bill Clinton got credit for the economic recovery actually primed by the first President Bush.

Then there are the people who labor long and hard and slowly build something from virtually nothing, such as Fred Adams, who created the now well-known Utah Shakespeare Festival [which was good enough to win a Tony several years ago as the best regional theatre in the U.S.]. His wife Barbara did half the work, but only those who knew Fred and Barbara know that because Fred was not only a great builder, but a great showman. There’s also a well-known fantasy author whose wife contributed to every book, but whose name only appeared on the last few.

In more than a few cases, those who build an organization, a cause, a business just aren’t self-promoters, and often, because of that, others take credit… or the individual never gets credit.

More often than not, why someone gets credit, deserved or undeserved, is because they’re a good self-promoter, and there’s nothing wrong with that in itself – unless the self-promoter steals the credit from someone else. What is equally wrong is when the rest of us let the self-promoters who are stealing credit from those who deserve it reward the deceptive self-promoter.

National Identity and Anger

A recent poll from The Associated Press-NORC Center for Public Affairs Research showed that seventy percent of Americans felt that the country was “losing its identity.” Unfortunately, what the poll also revealed was that Americans couldn’t agree on what were the important components of that “identity.”

Although there are some points of agreement among Democrats, Republicans and independents about certain aspects of what makes up the country’s identity, such as a fair judicial system and rule of law, the freedoms enshrined in the Constitution, and the ability to get good jobs and achieve the American dream, recent political developments make it clear that the consensus on these points is overshadowed by the differences.

Fifty-seven percent of Republicans thought one of the most important parts of the national identity was a Christian belief structure, as opposed to twenty nine percent of Democrats. On the other hand, sixty-five percent of Democrats thought that that the mixing of global cultures in the U.S. was important, compared to thirty-five percent of Republicans.

According to the poll, seventy-four percent of Democrats say that the ability of immigrants to come to the U.S. to escape violence and persecution is very important, as opposed to fifty-five percent of Republicans. Forty-six percent of Republicans agreed the culture of the country’s early European immigrants was very important, versus twenty-five percent of Democrats.

Putting these findings together suggests that, in general, Republicans think that the national identity should be based on an enshrined Christian faith and the Anglo-centric patriarchal culture of the first immigrants, while Democrats emphasize a more global-culture, welcoming to immigrants, and more concerned with the present than the past. Obviously, that’s an oversimplification, but there’s still a basic conflict, almost between the past and the present.

That conflict was definitely revealed in the last election, with the Republicans essentially claiming that the country was turning from its white, European, and totally Christian roots, and that such a turn was destroying and/or diminishing not only the United States, but the position of middle-class white American males.

As both the AP-NORC Poll and the Women’s March on Washington [with millions of women in hundreds of cities and towns across the country] showed this Republican “traditional” society is not endorsed by a significant percentage of the country.

Yet the Founding Fathers attempted to hold together thirteen colonies of very different belief structures, some with the [to me] abhorrent idea that slavery was morally acceptable, and they crafted a government based on shared principles that did not require a specific religious belief, or indeed, any belief in a supreme deity at all. For the time, this was an extraordinarily radical enterprise, so radical that the American Revolution equally merits the title of the Anglo-American Civil War.

So why is there so much disagreement about national identity and national priorities?

The election results and the vitriolic rhetoric from the right reflect, among other things, that there are fewer and fewer well-paid unskilled and semi-skilled jobs, and those jobs already lost to out-sourcing and technology, but mainly to technology, removed some eight million largely white men from the middle class. Those men and their families and relatives look to a past of more secure and prosperous employment and believe that the country has lost its way… and its traditional identity, and they’re angry.

On the other hand, there are over forty million African Americans in the U.S., and while the Civil War that resulted in their freedom ended over 150 years ago, those blacks still face discrimination and other barriers to rights equal to other white ethnicities. After 150 years they’re angry, and getting angrier, especially given the number of young black males killed and incarcerated, particularly when study after study shows discrimination still exists and that blacks receive harsher jail sentences for the same offense as do whites… among other things.

Educated women of all ethnicities are angry that they do not receive even close to equal pay for the same jobs as men and that the male-imposed glass ceilings in business, government, and politics still remain largely unbroken.

Since women and minorities are getting more and more vocal, and since minorities are becoming a bigger and bigger share of the American population, I foresee some very “interesting” years ahead, and I’d suggest that the largely white male Congress consider those facts very carefully.

The “Other” Culture

There are several definitions of “culture.” One is the development of microorganisms in an artificial media. Another is “the refinement of mind, morals, or taste.” A third is “the cultivation of plants or animals.” But there are two other definitions that tend to get overlooked: (1) the specific period or stage in the development of a civilization and (2) the sum total of the attainment and learned behavior patterns of any specific period or group of people regarded as expressing a way of life. The second of those latter definitions is the one that tends to get overlooked in government and politics, and yet the problems caused by the “learned behavior patterns” of smaller groups within a society represent one of the principal reasons for societal unrest.

That is largely because quite a few nations, including the United States, are in fact composed of various subcultures. In the U.S., those subcultures, especially those disliked by the majority, are often minimized or denigrated in racial or religious terms. An important point, and one consistently ignored by ideologues, businesses, and particularly politicians, is that “culture,” as exemplified by learned patterns of behavior, trumps “race” or religion. By that I mean that the good or bad traits of group or subgroup of people have virtually nothing to do with their religion or their skin color or ethnicity. What determines how people act is their “learned patterns of behavior.”

And while religion is definitely a learned behavior, how people of a certain religion act can and does vary enormously from cultural group to cultural group. It also varies over time. Some 500 years ago, good “Christian” countries in Europe were slaughtering each other in a fashion even more brutal than that in which the Sunni and Shia factions of Islam are now doing. Yes, religion is a critical part of “culture,” but it ranges from being the primary determinant of a culture to being merely one of many factors, and in the history of certain civilizations, the impact of a religion can and has changed the culture drastically.

As I’ve also noted before, likely more than a few times, history is filled with examples of both great and failed societies and nations identified as being predominantly of one race or religion. There have been great empires in all parts of the world – except, so far, Antarctica, and there have been failed societies everywhere in the world, regardless of race or religion.

Certain cultural practices seem to work better than others, one of which is that cultures that allow religion to control society tend to stagnate and become ever more brutal. Cultures with great income inequality tend to be more likely to be oppressive, and a greater percentage seem to have either de jure or de facto polygamy. A good sociologist could likely carry this much farther, but the basic point is that it’s not only morally wrong to claim that a given race or ethnicity or religion is “stupid” or “inferior” (or any other number of pejorative terms), but also such unthinking “type-casting” totally misses the point. Culture – not race, genes, skin color, or religion – determines how people behave. More to the point, one can change a toxic culture [although it takes time] and a beneficial culture is always only a cultural change or two away from becoming toxic.

The Threat from Radical Islamic Terrorists

I’m fed up with the propaganda from the White House about the “overwhelming” danger to U.S. citizens from radical Islamic terrorists. Yes, there are radical Islamic terrorists, but here in the United States, radical Islamic terrorists are far less of a threat than home-grown right-wing terrorists. Overseas, that’s another question, which is why so many law-abiding members of the Islamic faith want to get to the U.S. – or did before Donald Trump became President.

While consensus on hard numbers is difficult to come by, and numbers vary by source, whatever the source, those numbers suggest that radical Islamic terrorists are not the major threat to Americans – not even close. Other Americans are.

On terms of terrorist attacks in the United States, the numbers are lopsided, to say the least. According to a study by the United States Military Academy’s Combating Terrorism Center, domestic right-wing extremists averaged 337 attacks in the U.S. in the decade after 9/11, accounting for 254 fatalities, while, depending on the study and the definitions, between a total of 20 and 24 terrorist attacks in the U.S. were carried out by Islamic radicals with between and 50 and 123 fatalities.

In the last several years, the vast majority of police deaths have come from domestic extremists. An ADL report states that of the forty-five police officers killed by extremists since 2001, ten were killed by left-wing U.S. extremists, thirty-four by right-wing U.S. extremists, and one by domestic Islamic extremists.

As far as Trump’s proposed travel ban goes, there has not been one single terrorist attack on U.S. soil in the last four decades that has been carried out by citizens of any the seven countries on Trump’s ban list. Out of the 240,000 Americans who have been murdered since the attacks on the Twin Towers in 2001, exactly 123 of those deaths were linked to Muslim-American extremists. In other words, .05123 percent of the murders in the United States in a sixteen year period were carried out in the name of radical Islam. Even figures from the right-wing National Review only list 88 deaths in the U.S. from radical Islamic terrorists since 2009.

Yet at the same time that Trump is citing the danger from radical Islamic terrorists, reports have surfaced that he plans to shut down Homeland Security’s watch list for domestic extremists. Not only that, but bowing to the NRA, he decided to void an Executive Order by former President Obama that would have put people declared to be mentally incompetent by a court on a do-not-buy list for firearms. The NRA argued that mentally incompetent people should have the same right to firearms as anyone else.

And we’re worrying about Islamic terrorists?

Education and the Business Model

More and more state legislators across the country are demanding that education be run in a more business-like fashion. While greater efficiency is likely necessary in many educational systems, running higher education “like a business” is not only counter-productive, but it’s more likely to create more problems than it purports to solve – and business-like approaches so far haven’t shown much success at the university level.

One of the tools employed by both business and educational systems run by the “business model” is to reduce costs by reducing the number of employees and their compensation in relation to “output.” In the business area, this has given us outsourced manufacturing or high-tech automated manufacturing or, on the other hand, in retailing, lots and lots of underpaid, part-time employees without benefits. In education, a similar change is occurring, particularly in higher education, where university faculties have shifted from those primarily comprised of full-time dedicated professors to faculties where the majority of teaching faculty are part-time adjuncts, many of them far less qualified or experienced than seasoned full-time faculty. At the same time, administrations spouting the “business model” mantra have burgeoned.

At virtually all public universities, administrative overhead and full-time administrative positions have increased two to threefold over the past twenty plus years, while full-time faculty positions have decreased, except in the case of some smaller state universities that have expanded massively so that full-time positions have increased somewhat, even though the percentage of full-time positions has decreased to the same level as at other state universities, if not more.

The chief reason for this emphasis on part-time teaching positions is cost. As I’ve noted before, fifty years ago, on average, state legislatures supplied the bulk of higher education funding. In 1974, states provided 78% of the cost of educating a student. Today, although total funding is actually higher, because almost four times as many students attend college today, the amount of state funding per student averages around 20%, although it varies widely by state, and in some cases it is around 10%.

For example, almost until 1970, California residents could attend the University of California [Berkeley] tuition-free. Today, tuition and fees for in-state students are around $15,000 a year. This trend, if anything, is accelerating. Just since 2008, state funding of higher education has dropped by 20% per student

The response by legislatures, predictably, is to push for more efficiency. Unhappily that has translated into “get lower costs however you can.” The problem with this is that the emphasis, no matter what administrators say, is to turn out the most graduates at the lowest cost. Universities also tend to phase out departments with small numbers or high costs, and expand departments with large numbers and low costs, even if students that major in that area have difficulty getting jobs.

In addition, political pressure, both to “keep” students in school for budgetary reasons and to graduate a higher percentage of students, has inexorably decreased the academic rigor of the majority of publicly funded universities and colleges. This, in turn, has led to more and more businesses and other employers demanding graduate degrees or other additional qualifications, which further increases the tuition and debt burden on students. That’s scarcely “economic” on a societal basis because it pressures students to aim for high income professions or high income specialties in a profession, rather than for what they’re good at doing and what they love. It’s also created an emphasis on paper credentials, rather than the ability to do a job. On top of that, it’s meant more highly qualified individuals are avoiding professions such as teaching, library science, music, art, government civil service, and others; and those professions, especially teaching, are being filled by a greater percentage of less highly qualified individuals.

The end result of the combination of stingy state legislatures and the “business model” is less rigorous academic standards and watered down curricula at the majority of public colleges and universities, skyrocketing student debt, a smaller and smaller percentage of highly qualified, excellent, and dedicated full-time professors, and a plethora of overpaid administrators, the majority of whom heap even more administrative requirements on full-time teaching faculty.

No efficient business actually operates this way, and why higher education gets away with calling what it’s doing “the business model” has baffled me for more than two decades.

Economics, Politics, Business, and Regulation

To begin with, economics and business are not the same, although they share much of the same terminology, because the economics of business center on business, while the study of economics, at least in theory, encompasses all of society, and just not business, even though some business types have trouble comprehending that a nation’s economy consists of more than business, or more than government and business.

And no matter what they claim, most business people really don’t understand economics, or choose not to. Likewise, very few economists really understand business. Politicians, for the most part, understand neither, and most Americans understand even less than the politicians. This is, I submit, one of the fundamental problems facing the U.S. today.

Let’s just look at why in terms of fundamentals. Supposedly, the basis of economics and business rests on the interaction of supply and demand. In general terms, “supply” means the amount of a good sellers are willing to provide at a given price. Demand is what buyers will purchase at a given price. In a relatively free market [there are no totally free markets, and never can be, a point too many business types fail to acknowledge publicly], the going price of a good or service is set when supply and demand meet. If there is greater demand or a lesser supply, usually prices rise. If demand falls, or supply increases significantly, prices usually fall, again in a relatively free economy.

Of course, no economy is completely free because to have a working economy requires a working society, and human beings have yet to create a working society without government, and government, for various reasons, always imposes restrictions on the market. Some of those restrictions, given human nature, are necessary. Why? Because of the intersection of the way business operates and human nature.

As some have pointed out, the price of a good or service is not necessarily its cost plus remuneration to the supplier, but over time, price has to consist, at the least, of the amount necessary to cover costs of production plus enough above that to keep the supplier or business going. But the devil is in the details, and one of those details is how one defines “costs of production.”

There are all sorts of costs – fixed costs, marginal costs, operating costs, external diseconomies [otherwise known as negative externalities], etc. The cost that matters most to a business is whatever costs the business is required to pay by both the demands of the marketplace (i.e., supply and demand) and the government. If a business has to pay taxes, that’s a cost imposed by government. So are wage, benefit, safety, and environmental standards.

So… by what right, in a supposedly free market economy, is government imposing those costs on business?

The reason for government action is because: (1) the marketplace doesn’t include all the costs of production and (2) a totally “free” marketplace creates wage levels and working conditions virtually all western governments have declared unacceptable, and, therefore, governments have set minimum standards for wages, safety, and worker health conditions.

In addition, some of those government taxes provide for the highways and airways on which business goods are transported, for the national defense which protects business and everyone else from enemies from coming in and seizing businesses and properties and which allows U.S. businesses to conduct operations elsewhere in the world, for regulation and continuance of a stable banking system, for public safety, and so forth, all of which make the operation of businesses possible.

One of the reasons that, years ago, the Cuyahoga River next to the Republic steel mill in Cleveland caught fire was because the marketplace cost, and thus the price of a good, didn’t include costs passed on to others in society in the form of polluted air or water, and thus, any manufacturer who did restrict the emissions of pollutants incurred higher costs compared to producers who didn’t. Consequently, marketplace “discipline” effectively encouraged pollution, or at the very least, certainly didn’t discourage it. Costs inflicted on others are usually termed negative externalities [the older term is external diseconomies], but such terms tend to gloss over the fact that pollution and other degradation of the environment caused by manufacturing is not reflected in the cost of production unless government requires it.

So, when a manufacturer claims that environmental or worker safety regulations are stifling the economy, what that manufacturer really is saying is that he or she can’t compete with manufacturers in other countries that have fewer environmental regulations, and thus, often lower costs of production… and when that manufacturer demands less regulation, it is a demand to allow more pollution so that the manufacturer can make more money – or even stay in business.

Balancing economic output and worker and environmental health and safety is a trade-off. Although some regulations have been ill-thought-out, in general, stricter regulations result in a better environment for both workers and society, but if the rest of the world has lower levels, those U.S. industries competing in a global market will suffer higher costs, unless they have other cost advantages, such as better technology or far more productive workers. Because environmental control technology is expensive, most industries tend to oppose regulations requiring more technology.

In certain industries, workers, such as coal miners, often oppose environmental rules because those rules raise costs, and higher costs may result in the loss of their jobs. The question in such cases is whether continuing such jobs is worth the environmental and health damage, both to workers and to others. The Trump administration is working to remove an Obama administration rule that put stricter limits on how close to watercourses coal mining and chemical wastes could be placed, claiming that the rule will cost jobs, which it likely would to some degree. But the rule would also cut the number of coal and chemical industrial storage and waste disposal sites near rivers and streams in an effort to eliminate slurry and waste accidents such as the one along the Elk River in West Virginia in 2014 that fouled miles of streams and rivers, poisoned hundreds of people who drank the water unknowingly, and left more than 300,000 people without drinkable water for months.

History has shown, convincingly, for all who are willing to look at the facts, actual deaths, poisonings, and worse, that, without government regulations, a significant proportion, sometimes all, manufacturers in an industry will commit unspeakable wrongs in the search to maximize profit. Remember when the Ford Motor Company tried to cover up the faulty design of the gas tank in the Ford Pinto, deciding that it was cheaper to pay legal costs for deaths [which Ford estimated at $49 million] rather than produce a more expensive gas tank, which would have cost $113 million. Ford decided against the fix on a cost-benefit basis, then ended up paying out much more in legal settlements, in addition to a costly recall.

This kind of business cost-benefit analysis continues today, and that’s why the “business model” can’t be allowed without oversight and regulation. The question is not whether to regulate or not to regulate, but how much regulation is appropriate in what circumstances. Or put another way, is your business more important than my health? Except that business owners would say, an increase in regulations will kill my business and probably won’t measurably improve your health. Both are likely exaggerating, and that’s why verifiable science and facts – scientific, financial, and economic – are critical, and why political slogans and political pressure brought by outside interests have no place in determining whether a regulation is necessary, and if so, the degree of regulation required.

Bookstore Idiocy

Last weekend, I attended a fantasy and science fiction literary symposium in northern Utah, called LTUE (or, after a noteworthy writer, Life, The Universe, and Everything). As more of a literary symposium than a standard convention, LTUE attracts a great number of writers and editors, and an even greater number of would-be or beginning F&SF writers. Over the years, the guests of honor have included best-selling authors, F&SF publishers, and noted editors in the field (and, yes, I’ve been a GOH twice).

One of the highlight events of LTUE is a “mass signing” of all attending authors on Friday night, and this is facilitated by a book-selling site in the same enormous room as the mass signing, which means that those who are attending can run over and buy a book for an author to sign if they attended panels or discussions and realized that they really wanted to try an author’s work.

For over twenty years, the Barnes & Noble in Orem operated this on-site temporary book-selling venue, and, from what I’ve observed in the years I’ve attended, they seemed to do very well indeed. I know that my books have always sold moderately well at LTUE, and often the works of bigger name authors sold in the hundreds of copies over three days.

This year, however, the B&N store was unable to continue this activity, not because it didn’t sell books and make money, but because B&N recently adopted a chain-wide policy that banned “satellite events.”

To me, such a blanket policy makes no sense. I could understand a policy that declared that satellite events must cover their costs or come close, but a blanket ban? This reeks of accounting bean-counting. The business of a bookstore is, at least ostensibly, to sell books. If LTUE gets readers to try reading authors new to them, at least a proportion of those readers will buy more books by those authors. This increases sales, and since B&N is the only large set of bookstores in Utah, at least some of those sales will come from B&N. What’s not to like about making a bit of money at the symposium and increasing overall sales?

Tom Doherty, the publisher of Tor, came up through the sales ranks, and he’s told me more than a few times about the role the small mall stores – Waldenbooks and B. Dalton, now both defunct – played in developing readers, because they were convenient places for people to pick up books, not destination stores like the current B&N megastores, or the vanished Borders stores. Now, most of those convenient places are gone, whether it’s the vanished rack in the drugstore, the small mall bookstore, or the like, and those bookselling venues that are left are stocked by computer on based on national sales that often have little to do with the community where the sales outlet is located. Along that line, the ability of B&N store managers to customize their inventory has been reduced, if not eliminated.

I know B&N is having financial problems, but focusing on almost mindless cost-cutting when the effect of cost-cutting is to reduce sales is counter-productive. Success is measured by increasing sales in a cost-effective manner, not by cutting costs and doing less. You don’t turn around a financial down-turn just by cutting costs; you also have to increase sales, and doing things like a blanket ban on satellite events cuts down on sales. It also leaves a bad taste in the mouth of the symposium regulars and organizers, as far as B&N is concerned – and those people are all heavy readers. Does this really make sense, economically or in PR terms?

By the way, a small book vendor did step up at the last moment and set up an on-site book store, and she certainly sold a number of my books, as well as those of quite a few other authors.

Political Appeal and Innumeracy

U.S. federal spending in 2016 was roughly $4 trillion, and revenues were slightly over $3.4 trillion, leaving a deficit of around $600 billion. Out of total spending, $2.6 trillion was mandatory spending on programs such as Social Security, Medicare, and Medicaid. Spending on these programs cannot be cut without major changes in federal law, and since 77% of all Americans oppose such cuts, it’s highly unlikely that major cuts will occur any time soon. Then add to that some $260 billion in mandatory payments on the federal debt, and essentially 72% of federal spending cannot be effectively cut, at least at present. That leaves $1.1 trillion in discretionary spending, that is, spending that can be increased or decreased by Congress.

Unhappily, the vast majority of Americans have no real understanding of even these basic numbers, especially Fox News viewers, 49% of whom declared in a recent poll that cutting “waste and fraud” would eliminate “the national debt” [which now stands at $14.4 trillion]. A number of polls over the year have shown that most Americans believe that 25% of the federal budget goes to foreign aid [it’s less than one percent], and that five percent of all federal spending goes to PBS and NPR [in fact, roughly a tenth of one percent does].

The real numbers are more daunting. The largest component of discretionary spending is defense, and while the DOD “official” budget is slightly under $600 billion, various contingency funds and defense activities funded in other forms and by other agencies [for example, the Coast Guard is funded by the Treasury Department], brought the total annual cost of U.S. defense much higher, as high as $900 billion, according to some sources, but even assuming $600 billion for defense, that leaves $500 billion for everything else, including agriculture, energy, education, transportation, federal lands management, national parks, environmental protection, veterans benefits, welfare payments, and a whole lot more.

Trump’s proposed tax cut would reduce federal revenues by $500 billion, according to the Tax Foundation, on top of that $600 billion deficit, so even if he could persuade Congress to cut non-defense discretionary spending by 50% — in essence gutting most federal agencies, the deficit would increase to nearly $900 billion, and that doesn’t count the additional spending he’s proposed for infrastructure spending – which initial estimates suggest range from $500 billion to over a trillion dollars, over ten years, or $50 billion to $100 billion a year.

Proponents of the Trump plans claim that all the new investment and jobs will increase tax revenues, and some probably will, but not anywhere close to enough to deal with the federal deficit that increases the national debt – and the interest that must be paid on it – each year.

Based on a 2014 study by Standard & Poor’s, if Congress were to pass a $50 billion a year infrastructure bill, that legislation would create an additional 1.1 million jobs. Construction workers make an average of around $35,000 a year, and, under the best estimate of the Trump tax plan, those million workers would pay around $4,000 in federal income taxes each, thus adding up to an additional $4.5 billion. Economists like to point to the multiplier effect, i.e., how many additional jobs are created by one new job. According to the IMF, under present conditions, the multiplier effect is hovering around one, one additional job created somewhere in the economy for each new job created by investment. So… fifty billion dollars of infrastructure investment might create somewhere over two million jobs and possibly add $10 billion in tax revenues while costing $50 billion. Even if the multiplier effect is five times as much as the IMF says, the infrastructure proposal is at best a break-even proposition, and, as such, might be a good idea. BUT… it won’t do much for reducing the current deficit, let alone the increase in the deficit that will be occurring as a result of more federal spending on defense, and the likely coming increase in interest rates.

The other bottleneck in increasing jobs is the mismatch between available workers and the available jobs. According to research from human resources consultancy Randstad Sourceright, a survey of more than 400 U.S. executives found a skills gap impacting their businesses. Four-fifths of those executives said that a shortage of sufficiently skilled workers will affect their companies in the next 12 months. Complaints of hard-to-fill factory jobs are backed up by Bureau of Labor Statistics data: 324,000 manufacturing spots were open in November, up from 238,000 a year earlier.

Another problem that the Trump approach doesn’t address is that jobs creation isn’t equal. Right now, employees of high-tech companies receive almost 12% of all employee compensation, but there are only seven million of them and the average salary is close to $105,000, more than double the salary of the average industrial or manufacturing employee, or triple that of a construction worker. In addition, the tech industries are only adding about 200,000 employees a year. That doesn’t do much for the nearly 15 million unemployed or underemployed Americans, or the roughly three million college graduates each year. The largest numbers of jobs are in the lower paid service industries, and all the investment money putatively freed up by the tax cuts will be going to tech-heavy companies, and those jobs comprise less than 5% of total U.S. employment.

Massive tax cuts, more defense spending, a major infrastructure initiative… all to be paid for by new jobs and cuts in such federal programs as PBS, NPR, the Endowments for the Arts and Humanities, foreign aid, and the like? The numbers don’t add up, even if the political appeal does, perhaps because most Americans don’t seem to understand the numbers, or care to.

The Education/Business Fallacy

Recently, a semi-prominent president of an educational institution told a group of music professors that they shouldn’t complain about the fact that they were paid less than professors in other disciplines or that they were required by the institution to work longer hours and more days than most other professors because they “knew what they were getting into.” Besides the arrogance of the statement, I also found the sheer ignorance behind those words even more disturbing.

First off, when the vast majority of students on the collegiate or graduate level begin their academic preparation for their careers – whatever those careers may be – they have only the vaguest understanding of the scope of that career or of the demands it will make on them. Those only become truly apparent AFTER students graduate and move into the professional fields. That’s one of the reasons why something like 50% of all teachers drop out of teaching within five years. It’s why professionals change careers or leave them behind totally.

Second, this kind of attitude is typical of those who regard education from the “business” mindset and contributes to such factors as pushing to obtain as many students as possible, regardless of whether the students are ready or suited for college and where there’s a huge push to “steer” students toward “STEM” education and careers, as if students are organic robots that can simply be programmed toward the most lucrative careers, or those that will at least allow them to repay their often-massive student loans. As both a parent of a number of children who have been successful in various fields and careers and as a former faculty member on the collegiate level, I find the idea that students can be successfully “programmed” for specific careers or even careers in a general field totally ludicrous. People have different levels of ability in differing fields and different mind-sets.

For someone to have suggested that I might have a career in music because pop music stars make lots of money would have been both criminal and deceptive, given that I can’t carry a tune and have no sense of rhythm. In turn, to suggest that a good music student who can barely pass basic chemistry or physics, and for whom calculus is akin to magic, would be better served by going into a science or technology career would also be criminal and deceptive.

Third, the emphasis on college as vocational training, particularly on the undergraduate level, ignores reality. Even today, most college-educated individuals change jobs and often entire career paths seven to ten times in their professional lives. Those who make those transitions most successfully are those who have learned how to keep learning. Even those who remain in the same field have found that the requirements of their positions continue to change as technology advances.

Fourth, available jobs and job requirements are constantly changing as the result of shifting economic factors and technological advancement, and “guiding” students to the current “jobs du jour” may serve those not strongly motivated to enter that field poorly indeed.

Fifth, while employment “supply and demand” does in fact determine compensation levels, those levels have increasingly less and less to do with the skills needed by society. At least at present, scarce skills, even those that aren’t all that necessary to the functioning of society, are more highly valued than many necessary occupations and services. No matter what the financial types say, we need very few hedge fund managers for a successful civilization. We need a lot more of the practical and mundane skills, from electricians and plumbers to good classroom teachers and more doctors in general practice, but fewer and fewer doctors want to be in internal medicine or general practice because those fields usually pay half what specialized medical fields do and require longer hours, making it far harder to pay off the medical school loans.

Finally, what drives personal success in any field is the love of what one is doing combined with the education and capability to do the job at hand. “Training” a student for a theoretically more remunerative field that disregards the student’s abilities and interests serves neither the society’s interests nor the student’s. It’s a sad commentary on higher education when a university president suggests that because economics lowers the comparative compensation of professionals in certain disciplines and because the university takes advantage of that to the point of requiring more of those individuals, it’s all the fault of those professionals because they “chose” to pursue the field in which their talents lie.

This administrative mindset is also why more and more universities hire fewer and fewer expert and dedicated full-time professionals and more and more underpaid part-time adjuncts, because the quality of the instruction has become increasingly less and less important than the push to lower “people” costs, or at least the people costs associated with actual learning, as opposed to those associated with collegiate athletics.

Decline of Fictional Uniqueness?

As some of my readers know, these days I binge-read fiction on business trips or other travels, and, for the most part, I make an effort to search out books and authors I haven’t read, as well as books that deal with what I’d call interesting subjects or more familiar subjects addressed in a unique fashion.

The problem is, at least for me, that, beneath the veneer of “new and different” claimed by publishers and authors, I’m finding that there really isn’t all that much truly new and different. Oh, there are definitely books that deal with “new and different,” but not nearly so many as the publishing hype might suggest. Perhaps that’s always been the case, and perhaps when an author gets older, and has read as many books in the field as I have, it’s just harder to find something that’s truly different.

But I’m not so certain about that. Tolkien re-invented heroic fantasy with The Lord of the Rings, and I can’t even count the number of follow-ons and knock-offs. As far as I can determine Fred Saberhagen re-invented the vampire genre with The Dracula Tape in 1975 [Ann Rice’s Interview with the Vampire wasn’t published until May of 1976], although one could also claim that Richard Matheson’s I Am Legend [1954] was the first of the true twentieth century vampire “re-births,’ but Matheson’s blood-suckers were more “generic.’ Saberhagen also pioneered the whole idea of malevolent, non-gendered cyber beings with his berserker stories, something that tends to get overlooked in all the hoopla about Ann Leckie’s Ancillary Justice and its sequel.

Certainly James Tiptree, Jr., [Alice Bradley Sheldon], Joanna Russ, Sheri Tepper, and Ursula K. LeGuin were questioning gender roles and societal norms some thirty years ago, and even in 1987 Melissa Scott wrote The Kindly Ones, a masterful work in which it is impossible to determine with any certainty the gender of the protagonist.

The Wheel of Time and Game of Thrones are essentially huge-scale epic fantasies, with a few twists, that, in my mind, at least, fall into the post-Tolkien follow-on school.

Now, as I’ve noted in some of my comments on what I’ve read, there are still books with unique twists on old themes and some few with new themes, and I’m still looking, but it just could be that, as I’m getting older, it’s just harder to surprise me.

What do you think… and what books have struck you as unique… and why?