Author Archives: susanjustwrites

Hey There, Handsome!

Hey, handsome!  You know who you are.  You’re a charitable donor to at least one worthy cause you support.

Say what? 

In a recent article in The Wall Street Journal, Arthur C. Brooks, the head of a nonprofit organization, accumulated a wealth of data to support the conclusion that giving to charity makes us happier, healthier, and yes, even better-looking.

First, according to one study cited by Brooks, happiness and giving are strongly correlated.  A survey by the University of Chicago showed that charitable givers are 43% more likely to say they are “very happy” than non-givers.  By contrast, non-givers are 3.5 times more likely to say they are “not happy at all.”  Wow!

But is it really charitable giving that makes us happier, or is it the reverse?  Another study provided one answer.  Researchers from Harvard and the University of British Columbia found that the amount of money subjects spent on themselves was “inconsequential for happiness,” but spending on others resulted in significant gains in happiness. 

In another study, University of Oregon researchers asked people to divide $100 between a food pantry and their own wallets.  The researchers used a brain-scanner to see what happened.  It turned out that choosing the charitable option lighted up the brain’s center of pleasure and reward, the same center that lights up because of pleasurable music, addictive drugs, and a mother’s bond with her children.

Are we also healthier when we act in a charitable way?  Brooks cited several studies that say we are.  A University of Buffalo psychologist recently studied more than 800 residents of Detroit and found that volunteering for a charity significantly lowered the association between stressful life-events and death. 

Two studies conducted in California lent further support to this notion.  When researchers at Stanford and the Buck Institute for Research on Aging tracked nearly 2000 older Americans over a nine-year period, they found that the dedicated volunteers in the group were 56% more likely to have survived all nine years than non-volunteers who started out in identical health.  A study of teenagers yielded even more support.  In 2008, the University of California reported that altruistic teenagers were physically and mentally healthier later in their lives than their less generous peers.

And now we get to our most intriguing question:  Does being charitable do anything for the way you look?  Dutch and British researchers recently showed women college students one of three videos featuring the same good-looking actor.  In the first video, he gave generously to a man begging on the street.  In the second, he appeared to give only a little money.  In the third, the actor gave nothing to the panhandler.  The result? The more he gave, the more handsome he appeared to the women in the study.

Brooks concluded that this finding explains why men loosen their wallets in an attempt to impress women.  And he uncovered one more study to support his conclusion.  A 1999 experiment conducted by the University of Liverpool showed that “eager men” on first dates gave significantly more to a panhandler than men who were already in comfortable long-term relationships.

In short, giving generously to the causes we support really does appear to boost our well-being and our esteem—even our appearance–in the eyes of others.  Although I have reservations about some of the techniques used by charities to pry money from us (see “Why Am I Suddenly a Member?” found elsewhere on this blog), I wholeheartedly support charitable giving and volunteering on behalf of worthy causes. 

The charitable men in my life have always looked good to me, and as I’ve gotten older, I find they’re looking better and better.

As for me, in addition to my feeling good about giving, I now know that it helps me look good, too.

That reminds me…where’s my checkbook?

 

 

The Demise of the Granada

When they tore down the Granada movie theater, a large chunk of me crumbled with it.

As the wreckers began dismantling the magnificent old movie palace on Chicago’s Far North Side, other moviegoers must have felt the same sense of loss.  For those of us who came of age in the ’50s and ’60s, it was a wrenching reminder of the idyllic world we inhabited back then.

I grew up at the Nortown Theater, two or three miles west of the Granada.  It was the theater we could walk to, and nearly every Saturday afternoon we made our way to the Nortown to sit beneath its dark-sky ceiling filled with scores of glittering stars, our eyes glued to the larger-than-life stars who glittered on the screen.

Saturday afternoons at the Nortown expanded my otherwise limited horizons.  I learned about the Wild West from John Wayne, criminal pursuits from Bogart and Mitchum, romance from Taylor, Monroe, and Bacall, song and dance from Garland, Kelly, and Astaire.  But when our parents finally consented to our taking the Devon Avenue bus alone, a whole new world opened up:  the world of the Granada Theater.

Life became more complicated on the screen of the Granada.  At one remarkable double-feature in 1956, I encountered both the happiness and the sorrow of a woman’s search for love.  Katharine Hepburn’s spunky heroine, in love with a very-married Rossano Brazzi in “Summertime,” and Jennifer Jones’s strong woman doctor, in love with war journalist William Holden in “Love Is a Many-Splendored Thing,” bravely returned to their careers when their doomed love affairs ended.  An early portent of women’s liberation?  Judging from the masses of wet Kleenex we produced, I doubt it.

“The Man Who Knew Too Much” introduced me to the heart-pounding suspense stirred up by Hitchcock.  I watched “Anastasia” aware of the firestorm Ingrid Bergman’s scandalous love affair had ignited in Hollywood.   And a powerful statement about the criminal justice system, “12 Angry Men,” forced me to think about the possibility of injustice in America and whether I might someday do something about it.

As I grew older, the Granada became a place to go on dates.  Teenaged boys in that era liked taking dates to movies, where their eager sweaty hands would reach out in the dark in hopes of touching something soft, warm, and female.  They had limited success, at least with me.  My date and I once watched a shockingly bad movie with Tab Hunter and Natalie Wood, “The Burning Hills.”  It was so awful that we laughed too hard to do anything else.

In the ’60s, I rarely patronized the Granada.  I left Chicago for college and grad school, and when I returned, I lived in another part of the city.  On my last visit, just before leaving Chicago once again in 1970, the theater seemed rundown and much dirtier than I remembered.  Was the Granada on the skids?

Five years later, I returned to Chicago with a husband and a baby.  Living in a suburb north of the Granada, we passed it now and then, but my busy new life left no time to seek out old haunts.  Then one day it suddenly closed.  No warning, no notice announced in the newspapers, allowed its former patrons one last chance to see it.  The doors were locked, and entry barred.

Repeated efforts to save the Granada failed, and the wrecking ball finally arrived.  As I drove by the theater on my way to teach a law-school class, I saw the wall behind the screen fall to pieces and the two-story terra cotta columns crash to the floor.  The balcony seats were exposed to view, then destroyed.  At the end, a sodden ugly mass of tangled beams and columns, entwined with an array of aging construction materials, became a hideous pile awaiting disposal.

An era had ended.  TV, VCRs, and the proliferation of movie theaters in the suburbs all played their part.  Most of the opulent movie palaces that once thrived in American cities had become dinosaurs.  And so, in 1990, the Granada died.

But like the best of the movies that appeared on its screen for more than 50 wonderful years, the memories it created have never died.

An earlier version of this piece appeared in the Chicago Sun-Times.

 

Go p(nuts)! PB is actually good for you

Peanut-butter lovers of the world, rejoice!  This humble legume, in the form of an easy-to-eat spread, has recently earned some noteworthy praise.

First, one of the food industry’s harshest critics, the Center for Science in the Public Interest (CSPI), has just celebrated the virtues of peanut butter.  In the October 2013 issue of its publication, Nutrition Action, CSPI notes that peanut butter–a lunchbox classic and a staple in 90 percent of U.S. households–is loaded with unsaturated fat, vitamin E, and magnesium, and it supplies some copper, fiber, and zinc as well.  (Some must steer clear of PB because of peanut-related allergies, but most of us can eat it with abandon.)

True, CSPI acknowledges that there’s one small problem with peanut butter:  it’s also loaded with calories. Most people probably eat about 250 calories’ worth in the typical sandwich.  According to CSPI, that’s much more than the 50 to 80 calories in the equivalent amount (roughly 2 ounces) of turkey, ham, or a quarter cup of tuna.  These alternatives also offer more protein:  10 to 12 grams as compared with the 7 or 8 grams in peanut butter.

For the 90 percent of us who relish eating peanut butter, CSPI suggests some new ways to trim the calories.  For starters, there’s powdered PB.  It’s made by slow-roasting and pressing peanuts to remove 85 percent of the oil.  You just mix the powder with water and stir.  According to CSPI, the result is a creamy texture and rich peanut taste for just 50 calories per serving (with roughly the same amount of protein as regular PB).

Two other new products are whipped PB (fewer calories but less protein) and Better ‘n Peanut Butter (defatted peanut flour, mixed with PB and sugars, also cuts both calories and protein).

Traditionalists might want to stick with “natural” PB or even oldies like Jif and Peter Pan.  Happily, none of them have trans fat any more.  Just watch out for the new “artisan” varieties that add chocolate and other sweet ingredients, upping the usual 1 or 2 grams of sugar all the way to 9 grams.  Who needs it?  If you crave PB infused with chocolate, go for broke and have a candy bar instead.

But wait, there’s more good news for peanut-butter lovers!  In addition to CSPI’s focus on PB as a healthy sandwich-filler, the medical community has just declared an even more significant finding.  A study announced in September by Washington University School of Medicine in St. Louis (along with Harvard Medical School) revealed that girls ages 9 to 15 who regularly ate peanut butter or nuts were 39 percent less likely to develop benign breast disease by age 30.  Although benign breast disease is noncancerous, it increases the risk of breast cancer later in life.

Over 9,000 U.S. girls were part of the study, which began in 1996.  The researchers followed the girls until they were 18 to 30 years old.  This study is significant because it’s the first one that actually recorded what the girls were eating during their adolescent years (instead of relying on their recalling later what they had eaten years before).

The senior author of the study is Graham Colditz, M.D., a disease-prevention expert at Washington University’s School of Medicine.  Professor Colditz is an epidemiologist with a longstanding interest in cancer prevention, particularly among women.

According to Colditz, the findings in the recent study “suggest that peanut butter could help reduce the risk of breast cancer in women.”  He recommends that girls snack on peanut butter or nuts instead of reaching for high-calorie junk food and sugary beverages.

Wow!  Lots of great news about peanut butter!  I feel totally vindicated.  My instincts were right all along.

All those mornings making countless peanut-butter sandwiches for my daughters may have actually led to their staying healthy longer.  My choice to eschew fillings like bologna and head cheese (what was that stuff anyway?) probably didn’t hurt either.

A personal reminiscence about PB:  When my husband had a month-long sabbatical in Paris during the 1980s, we brought a jumbo jar of peanut butter from home because we knew it wasn’t readily available in France.  We wanted our small daughters to have a familiar food to eat while we otherwise attempted to live like Parisians.  I can still see myself in our tiny Paris apartment, spreading peanut butter on scores of French biscotti so our unfamiliar surroundings would feel a little more like home.

Like almost everything I’ve done (and still do) for my daughters, it was worth it.

Thinking about peanut butter has, not surprisingly, made me want some.  I’m ready to munch on a PB sandwich right this minute.  Want to join me?

“Paper or…?” Drying your hands has unexpected consequences

We’re all familiar with the following question:  Paper or plastic?

For decades, every purchase in a supermarket or drugstore has led to this question.  And for decades, many of us have wondered:  Is it better—for the environment, for my pocketbook, for my overall well-being—to request paper or plastic?  The answer hasn’t always been clear.

Never mind.  Today, in San Francisco and an increasing number of other cities, the question is moot.  Local ordinances ban plastic bags and require customers to pay for paper ones, thus encouraging shoppers to carry their own reusable bags.  The “paper or plastic” question is fast disappearing.

But now we’re confronted with a new but even more troubling question:  When we use a restroom in a public place and we wash our hands (as we’re repeatedly urged to do), should we use paper towels or an air blower?

In this case, we usually don’t have a choice.  Restaurants, stores, theaters, museums, and other institutions with restrooms for their patrons generally confront us with only one way to dry our hands:  paper towels OR air blowers.  A few establishments offer both, thereby giving us a choice, but most do not.

I’m a strong proponent of paper towels, and my position recently garnered support from an epidemiologist at the Mayo Clinic, Rodney Lee Thompson.

According to a story in the Wall Street Journal last December, the Mayo Clinic has published a comprehensive study of every known hand-washing study done since 1970.  The conclusion?  Drying one’s skin is essential to staving off bacteria, and paper towels are better at doing that than air blowers.

Why?  According to this study, paper towels are more efficient, they don’t splatter germs, they won’t dry out your skin, and most people prefer them (and therefore are more likely to wash their hands in the first place).

Thompson’s own study was one of those included in the overall study, and he concurs with its conclusions.  He observed people washing their hands at places like sports stadiums.  “The trouble with blowers,” he says, is that “they take so long.”  Most people dry their hands for a short time, then “wipe them on their dirty jeans, or open the door with their still-wet hands.”

Besides being time-consuming, most blowers are extremely noisy.  Their decibel level often strikes me as deafening.  Like Thompson, I think these noisy and inefficient blowers “turn people off.”

But, he adds, there’s “no downside to the paper towel,” either psychologically or environmentally.  Thompson states that electric blowers use more energy than producing a paper towel, so they don’t appear to benefit the environment either.

The air-blower industry argues that blowers reduce bacterial transmission, but studies show that the opposite is true.  Much to my surprise, these studies found that blowers tend to spread bacteria from 3 to 6 feet.  To keep bacteria from spreading, Thompson urges using a paper towel to dry your hands, opening the restroom door with it, then throwing it into the trash.

A recent episode of the popular TV series “Mythbusters” has provided new evidence to support Thompson’s conclusions.  The results of tests conducted on this program, aired in June 2013, demonstrated that paper towels are more effective at removing bacteria from one’s hands and that air blowers spread more bacteria around the blower area.

In San Francisco, many restrooms have posted signs stating that they’re composting paper towels to reduce waste.  Because San Francisco has embarked on an ambitious composting scheme, we’re not even adding paper towels to our landfills or recycling bins.  Other cities may already be doing the same, and still others (like New York City, where composting has already been proposed) will undoubtedly follow.

I strongly advocate replacing air blowers with paper towels in public restrooms.  Political leaders, including those who’ve already compelled their constituents to abandon plastic bags for the sake of the environment, should carefully review this issue as well.  If they conclude, as overwhelming evidence suggests, that paper towels are better both for our health and for the environment, they should enact local ordinances requiring that public restrooms use paper towels.

Paper or…?  The answer, my friend, is blowin’ in the wind.  The answer is blowin’ in the wind.

“One” Small Step for Humankind

Eliminating gender-bias in the English language has been a preoccupation of mine for many years.  During the 1980s, I came up with an idea that I believed would be a useful remedy for one kind of gender-bias rampant in English nouns.

My hopes that this idea would gain wide acceptance escalated when the Chicago Tribune in 1986 published a piece I wrote that advocated this change.

But despite my hopes that this idea would catch on (and although a number of people told me how much they liked it), it went nowhere.  During the intervening 27 years, everyone (including me) has moved on and dealt with this issue in some other way.  But I still think my idea is a good one.  Here’s why.

Back in 1986, I observed that few people were concerned with gender-bias in our language. Those who preferred gender-free language were using “person” instead of “man,” and “he or she” instead of “he” alone, and pretty much leaving it at that.  Others had refused to make even those substitutions and continued to use traditional parlance, much of which had an undeniable male bias.

Maybe the underlying problem was sexism, pure and simple.  But I preferred to have a more generous view and came up with another conclusion.

The fundamental question, then and now:  Why do so many people continue to use words like “man” to mean men and women?  I think I know the answer.  The persistent use of the word “man” and all of the words that use “man” as a suffix—policeman, fireman, repairman, deliveryman, Congressman, and all the rest—can be blamed on a simple fact:  It’s easier to say “man” than to say “person” or any other word or suffix of more than one syllable.

Why haven’t words like “Congressperson” caught on?   Because it’s easier to say “Congressman.”  (I’d be satisfied with “Member of Congress,” like “Member of Parliament” in the U.K., but that hasn’t caught on either.)

Let’s face it.  Even the most dedicated feminists among us would prefer to say mailman instead of mail carrier or repairman instead of repairperson–IF the shorter versions were gender-neutral–because they’re easier to say.  But any word that includes the suffix “man” cannot be gender-neutral.

People can tell me that “man” includes both men and women until their faces turn blue, but I don’t believe it for one second.  Why?  Because the word “man” conjures up the image of a man.  Just think back to your childhood.  Do you remember when you were being taught the meanings of the most basic words?  Who did your parents point to when they said the word “man”?  A woman?  C’mon!  All of us were taught that “man” means a man, not a woman.  No wonder we’re confused a few years later when we encounter writing or speech using “man” that purports to mean both men and women.

It’s almost impossible to say “businessman” and envision a woman.  The same goes for every other word using the suffix “man.”  But if we balk at such infelicitous terms as “businessperson,” what substitutes can we use?

I propose using the suffix “one.”  This suffix is commonly used as part of our everyday speech.  Think of the words “anyone,” “everyone,” “someone,” and “no one.”  No one quibbles about those.  On the contrary, the use of “anyman” or “everyman” would be comical (they sound downright medieval to me).  Why not, then, extend use of the suffix “one” to other words where we have traditionally used “man”?

In recent years, some gender-neutral terms, like “firefighter,” have caught on, and for that I’m grateful.  But many people still cling to the old descriptions, like policeman and Congressman.  If we adopted my proposal, the word “policeman” could become “police-one.”  “Congressman” could become “Congress-one.”  (I’ve added a hyphen for clarity until people become accustomed to the new usage, but that hyphen could disappear as the usage became second-nature.)  Adding “one” instead of “man” or “person” is both gender-neutral and easy to say.

These new words sound strange at first (and in print they look even stranger, especially if we omit the hyphen).  But after you use “one” a few times, it begins to trip off your tongue as easily as “man” because, like “man,” it’s only one syllable.  It also doesn’t sound very different from “man” (“person” does), and you don’t have to stop to think about which suffix to use every time you want to describe someone.  Instead of trying to remember “mail carrier” and “police officer,” you could just use “one” for all of them.

Why don’t you try it?  Try saying mail-one, police-one, sports-one, Congress-one.  Even…snow-one!  It might take a bit of getting used to, but I think it could work.

In my view, it’s never too late to do the right thing.  So even though we’ve come up with ways to get around sexist language, using awkward words like “spokesperson,” it’s not too late to make things even better.

Do you agree with me that we should stop using language that excludes half of humanity?  If you do, we could try this new approach.  If each of us tries it, maybe we can start a trend that changes the English language in this small but important way.

Why don’t we do it?  Let’s be bold.  Let’s target one day to try this experiment in gender-free language and see what happens.  I propose trying it on the first day of the first month of 2014.

Please join me.  On January 1, 2014, let’s try using this short and easy way to include absolutely everyone.

In Praise of San Francisco’s Weather

I moved to San Francisco eight years ago, and there’s much about the city that I truly love:  the breathtaking vistas, the natural beauty surrounding the city, the warmth of its inhabitants, and the rich assortment of parks, museums, theaters, concert halls, movie houses, restaurants, and shops.

There’s one more thing:  the weather.

I revel in the weather I’ve encountered in San Francisco.  After decades of living in a cold climate (mostly in Chicago), dealing with snow and ice for much of the year and heat and humidity for much of the rest, I relish the sunshine and cool breezes that San Francisco offers year-round.  People who’ve never lived in a cold climate can’t begin to imagine how difficult life there can be.  On many cold mornings I found myself crossing the bridge over the Chicago River, headed from the commuter train station to my office, snow and sleet blowing in my face.  No matter how many warm layers of clothing I’d wrap around my body, my face was largely exposed, bearing the brunt of the cold wind that persisted in hurling snow in my direction.

If you’ve never confronted them, let me assure you that icy sidewalks and streets are extremely treacherous.  Many of those attempting to walk on icy sidewalks have slipped and fallen, breaking bones and suffering concussions.  Driving on icy streets is equally hazardous, resulting in countless collisions.  Luckily, snow and ice are non-existent in San Francisco, relieving us of the challenges and pitfalls of negotiating on ice both on foot and in a vehicle.  Even rainy days don’t bother me, and locals who complain about the occasional chilly weather strike me as almost comically unaware of the reality faced by Americans in almost every other region of the county.

Unless you love hot weather and can’t wait to hit the beach, summers in San Francisco are delightful.  The temperature almost never rises above 80 degrees, and humidity barely reaches a noticeable level.  The contrast with places like Chicago, Boston, and New York is striking.  On recent trips to those cities, I encountered uncomfortably high humidity, thunderstorms, and temperatures in the 90s.  Extreme heat and humidity has plagued much of the nation this summer, but here in San Francisco, we’ve been as cool as cukes.

Air conditioning?  In San Francisco, we almost never need it, while most other regions of the country, including many parts of California, rely on air conditioning to survive.  I remember some vivid examples.  On one sweltering summer day in Ann Arbor, Michigan, my husband and mother joined me at Ann Arbor’s famed outdoor art festival.  I was surveying the artwork when l glanced at my mother’s face.  It was bright red. The thermometer on a nearby building read 99 degrees, and the humidity felt just as high.  We quickly abandoned the art festival and fled to our air-conditioned apartment.  On a recent trip to Boston, I was barely able to drag myself from the Harvard Square “T” to my daughter’s air-conditioned Cambridge apartment just a few blocks away, when both the temperature and the humidity hit 90-plus.  And don’t get me started on places like Arizona and Texas.

Here in San Francisco we save the financial cost of air conditioning, not to mention any feelings of guilt arising from  its demand on our energy resources.  And we don’t have to suffer the physical jolt of going from intense heat to intense cold every time we enter a super-air-conditioned building.

Our weather has another stellar feature.  Because San Franciscans can revel in sunshine and moderate temperatures all year long, we can spend much more time outdoors than most other Americans.  We’re not confined to exercising in sterile gray-walled fitness centers.  We have much better options.  I wake up every day almost certain that I’ll be able to take a walk, hike, or bike ride before the sun sets.

I don’t even mind the San Francisco fog that occasionally envelops the city.  Au contraire.  I think it creates a kind of magical aura over the city.  So long you remember to carry a light jacket, and drivers are careful maneuvering their vehicles in the fog, it really doesn’t have much of a downside.  Besides, if you want to escape the fog, you need travel only a short distance from the city in any direction.  The microclimates surrounding us are almost always fog-free.

Of course, life in San Francisco has its flaws.  For one thing, housing is more expensive than that in other cities (with the possible exception of NYC).  Rents are high, and on the rise as the city’s economy gets better and better, while buying a house in the most desirable neighborhoods has become more costly than ever.  And San Franciscans are constantly under the shadow of “The Big One.”  Perched as we are on the Pacific Rim, the threat of a major earthquake never really goes away.

But those of us who live here are willing to take those negatives along with all the positive features of life in the city.  Count me in.  I’m genuinely happy in my new hometown and especially delighted with its weather.  And when I recently came across the following story, reprinted from the San Francisco Chronicle of July 15, 1937, I realized that my reaction to the city’s weather is very much like that of a famous writer’s over 75 years ago.

The Chronicle reported:

Ernest Hemingway arrived in San Francisco yesterday ‘to get cool.’ On his first visit to [the city], he gulped in a few cubic yards of fog shortly after stepping from a …plane at [the airport] and sighed: ‘Say, this is great. After frying in New York, stewing down in Florida and sweltering in Los Angeles, this is something….  I can’t for the life of me see why anybody would ever move out of San Francisco, particularly in the summertime.’

Hey, Papa, when it comes to weather, we’re on the same page!

Quote

Audrey Hepburn and Me

I never thought I had a single thing in common with Audrey Hepburn.  She was tall and decidedly slim.  I’m short and, uh, not exactly slim.  She was a brunette with enormous brown eyes.  I’m a redhead with almond-shaped but not-so-enormous hazel eyes.  She was a famed film star who won an Oscar at 24 (for 1953’s Roman Holiday) while my adolescent dreams of becoming an actress never became reality.

So I never saw myself as having anything in common with this glamorous star of the ’50s and ’60s.  But a quick glance at a recent magazine article has convinced me that I have a few things in common with Audrey after all.

The article, appearing in the May issue of Vanity Fair, is based on a new book, Audrey in Rome, written by her younger son, Luca Dotti.  Luca lived with Audrey in Rome from the time of his birth in 1970 until she left for Switzerland (and he went off to a Swiss boarding school) in 1986.  As the magazine cover proclaims, in his book he recalls “the secrets of her iconic style.”

What were some of these secrets?  Well, for one thing, she was “fond of kerchiefs tied under the chin (not wound around and fastened in back in the French manner).”  Her love of sous-chin kerchiefs is apparent in a 1970 photo showing Audrey in a fabulous Givenchy coat and a scarf tied under her chin.

According to Luca, Audrey’s scarves were “a bit of a vice.”  Although she wasn’t “like Imelda Marcos and shoes,” she had “maybe 30 or 40” scarves.  In Rome, she often wore them along with big sunglasses as a disguise, enabling her “to do her shopping without having…crowds” following her.

This is one style-revelation I share with Audrey Hepburn.  My love of scarves, like hers, could be called a vice, but in view of the small amount of space they occupy and the small sums of money they cost, they’re a pretty harmless one.  I have a colorful collection in every possible fabric, suitable for every season, some bestowed on me as charming gifts, others purchased by me in a weak moment.

I admit I’ve never had crowds following me.  But I wear scarves (usually tied under my chin) for my own reasons.  In chilly weather, they keep my head warm.  On warmer days, they shield my curly hair from humidity and wind.

Childhood photos taken by my father show me, like Audrey, wearing scarves tied beneath my chin.  Ever since then, I’ve worn scarves no matter where I’ve made my home—from Chicago to Boston to Los Angeles.  Now, living in breezy San Francisco, I almost never leave home without a scarf in my jacket pocket, prepared to withstand whatever breezes the ocean blows my way.

Some have ridiculed my penchant for wearing scarves.  A friend once muttered that I liked to wear “babushkas.”  That hurt.  But now I can point to Audrey Hepburn as a scarf-loving style icon who, like me, wore scarves tied beneath her chin.

Another secret revealed by Luca is Audrey’s choice of footwear.  Generally basing her style choices on “simplicity and practicality,” she preferred to wear ballerina flats and low heels.  Vanity Fair claims that she wore them partly to accentuate her long feet, “adding to her elegant attenuation.”  (Huh?  Do you know any women with long feet who want to accentuate them?)  But even VF admits the far more likely reason:  she wore them so she “could walk comfortably.”

So here’s another preference I share with Audrey.  Long ago I gave up wearing high heels.  Like Audrey, I like to stride purposefully through the city, and wearing anything but low heels makes that impossible.  Every day I see women struggling with high heels that inhibit their freedom to move through life with ease.  I ache to tell them to forgo those high heels, and like Audrey and me, walk comfortably and safely wherever they go.

[Please note:  I’ve written another post on this blog, “High Heels Are Killers,” explaining at greater length my opinion of high heels.]

If truth be told, when I was younger, I wasn’t a big fan of Audrey Hepburn.  Maybe it was the way Hollywood portrayed her that was to blame.  After Roman Holiday (in which she fell in love with reasonably age-appropriate Gregory Peck), she was paired with male leads who were far too old for her.  At 28 she was supposedly smitten by Gary Cooper, then 56 (and looking even older), in Love in the Afternoon and by 58-year-old Fred Astaire in Funny Face.  I found these pairings simply baffling.  Why would radiant young Audrey fall for men twice her age?  At the time, I was unaware of the way Hollywood worked back then.  It’s clear to me now that she was complying with the demands of the movie moguls who dictated most of the roles she played.

No wonder she confided to friends that her favorite role was that of the nun in The Nun’s Story.  No superannuated men were slobbering over her in that role!

My view of Audrey Hepburn evolved as I learned more about her.  In her later years, she became an activist on behalf of UNICEF, traveling to more than 20 countries around the globe to advocate for the world’s most vulnerable children.  Her advocacy has endeared her to me, a fellow advocate for the underprivileged.

Moreover, during those years, she openly chose to welcome growing older.  Luca remembers that she “was always a little bit surprised by the efforts women made to look young.”  By contrast, “she was actually very happy about growing older because it meant more time for herself, more time for her family, and separation from the frenzy of youth and beauty that is Hollywood.”  She saw aging as part of the circle of life.

Audrey liked to say that “true beauty in a woman is reflected in her soul. It’s the caring that she lovingly gives, the passion that she shows. The beauty of a woman only grows with passing years.”

Some may remember Audrey Hepburn as a stunning style icon, but in my view, she should be remembered for much, much more.

“All in the Family” Revisited

Are you old enough to remember the TV sitcom “All in the Family”?  Or have you managed to catch an episode or two on late-night TV?

This sitcom was a number-one hit on TV in the 1970s (it debuted in ’71 and lasted till ’79), and it became an honest-to-goodness phenomenon.   Produced by Norman Lear, it featured movie and TV actor Carroll O’Connor as the irascible Archie Bunker.  Archie was a working-class bigot, openly racist and sexist.  Sitting in his favorite living-room chair (now enshrined in the Smithsonian), chomping on a cigar, he belittled gays and intellectuals and anyone else who lived outside his narrow world in Queens, New York.

Why did a sitcom revolving around this character become an Emmy- and Golden Globe-winning hit?  Probably because audiences felt comfortable laughing at Archie’s appalling antics.  Audiences could watch Archie clumsily try to maneuver through a rapidly changing world, feeling smugly superior to him while we grappled with many of the same troubling issues in our own lives.

Archie was surrounded by a memorable family, notably his long-suffering wife Edith, whom Archie called “the dingbat.”  Edith was played by Jean Stapleton as a somewhat flaky but kind-hearted helpmeet who tolerated her husband’s offensive behavior because she truly loved him.  Archie’s daughter Gloria, portrayed by Sally Struthers at the outset as a miniskirted twenty-something with corkscrew blonde curls, and Gloria’s husband, college student Mike, played by Rob Reiner in much slimmer times, completed the family circle.  Without enough money to afford their own place, Gloria and Mike lived with Archie and Edith, creating a situation rife with conflict.

Archie and Mike (dubbed “Meathead” by Archie) constantly clashed, their different world-views colliding on a daily basis.  Gloria was caught in the middle, sometimes siding with Archie but usually backing up her husband Mike.  This dynamic provided considerable comic fodder for the audience.  When, later in the show’s run, Mike and Gloria left Archie’s house and moved to their own place, a lot of the comedy went with them.

A few years back, someone asked me which TV show I would choose to inhabit if I was suddenly transported into a TV sitcom.  My choice was easy:   “All in the Family.”

I knew precisely where I wanted to go: 704 Hauser Street, Queens, New York, plunked down in the Bunker household as a newly minted version of Archie’s daughter, Gloria.

In my version of the Bunker family, Gloria would no longer be Archie’s relentlessly cute but somewhat uncomplicated daughter, declaring “Oh, Daddy!” whenever Archie did something that baffled or annoyed her.  I’d be a smarter, savvier Gloria, bringing a dose of common sense and a measure of sensitivity to the Bunker household.

Instead of running off to Mike, as Gloria was wont to do, I’d give Archie a hug, then sit down with him and offer him my empathy.  I’d let him know I understood how hard it was to be a blue-collar white male in a world that was spinning around him, changing by the hour.  I’d try to reassure him that he still had his place in that world, and that nothing would ever change my daughterly love for him.

I’d empathize with Edith, too, trying to reassure her as well.  I’d let her know it was okay for her to be content–for the moment–in her current role, that of a housewife whose focus was cooking, cleaning, and helping her husband deal with his daily defeats at home and at work.  At the same time, I’d encourage her incipient efforts to become more assertive, no longer entirely dependent on Archie and therefore no longer the willing target of his insults and disparaging attitude.

As for the Meathead, I’d struggle to keep our marriage intact, constantly reminding myself how much he loved me, calming him down whenever Archie was on the warpath, serving as a buffer between the two of them more effectively than Gloria ever did.

In sum, I’d bring tranquility and order to the Bunker household, thereby transforming the Bunker family into the kind of family I always tried to create in my own home.

There’s just one problem: “All in the Family” wouldn’t be funny anymore.  The Archie that I loved to laugh at would be buried under a cloak of rationality, with only bits and pieces of funny stuff breaking through now and then.

My family shared a house much like the Bunkers’, but our dynamic was nothing like theirs.  We bounced ideas off of each other, not always in total agreement but open to what each of us had to say.  As my children grew and the world evolved, we evolved, too.  We shared a home full of love and a minimal amount of conflict.

So, although we had loads of fun together, we were pretty boring compared to the Bunkers—not at all the stuff of a successful TV sitcom.  I guess I would have liked to see the Bunkers become more like us, but let’s face it:  The result would have been much closer to “Little House on Hauser Street” or “The Waltons of Queens,” and nothing like the very funny “All in the Family.”

Does Jury Duty Matter?

Have you ever served on a jury?  As a lawyer, I’ve observed juries over the years and found the whole process fascinating.  But although I’ve been called and questioned for jury service several times, I’ve never actually sat on a jury.

A few years ago, I wrote a book review recounting one juror’s experience sitting on the jury in a particularly salacious trial in New York City (please see my review in the Chicago Daily Law Bulletin in 2002 of A Trial by Jury by D. Graham Burnett).

More recently I read another book about jury duty.  Conceding that many of us try to avoid serving on a jury whenever we can, it makes a compelling argument that jury duty is absolutely vital in our democracy.

Here’s a review I’ve written of this new book, Why Jury Duty Matters.

  Book Review:  WHY JURY DUTY MATTERS

by Andrew Guthrie Ferguson 

         Does jury duty matter?  Anyone who’s seen the 1957 film “12 Angry Men” can answer that.  In that film, a single member of the fictional jury derails the conviction of a murder defendant when he persuades the other jurors that there’s reasonable doubt of the defendant’s guilt.

Real-world jury duty may not have the impact it does in that film, but it still matters—a lot.  In “Why Jury Duty Matters:  A Citizen’s Guide to Constitutional Action,” Andrew Ferguson tells us just how consequential jury duty is.

Ferguson writes from his perspective as a former public defender with the Public Defender Service for the District of Columbia, where for seven years he represented adults and juveniles in serious felony cases.  Now a professor of law at the David A. Clarke School of Law at the University of the District of Columbia, he has focused on how vital jury duty is to our democracy.

Ferguson notes that the significance of the jury was enshrined from the beginning of our country in the United States Constitution.  He points out that jury duty is “one of the last unifying acts of citizenship”—our “recurring civil obligation to head down to the courthouse and participate in resolving a criminal or civil case involving members of the community.”

He skillfully weaves in references to history, tracing the evolution of juries, first in England and later in the U.S.  For example, the unfair proceedings during the 1603 treason trial of Sir Walter Raleigh led to the right to confront witnesses, later enshrined in our Sixth Amendment.  Likewise, the 1735 libel trial of John Peter Zenger probably inspired the Sixth Amendment’s right to a public jury trial, as well as the First Amendment’s protection of free speech and press.  And the revolt against much unfairness by the British, which led to the American Revolution, led in turn to this theme in the Constitution:  protection against arbitrary police power.  The jury is a bulwark against that power.

Ferguson carefully reviews the many roles that jurors play:

  • deliberating—reviewing the evidence  and making a collective decision; thinking together, using reason and informed discussion to reach a decision
  • protecting dissenting voices—allowing each juror to dissent from the majority’s conclusion
  • judging accountability—assuming the responsibility to hold someone accountable
  • giving equal treatment to information and ideas, and above all,
  • ensuring fairness.

The key is fairness.

As Ferguson emphasizes, each of these roles, when taken to heart by jurors, leads to faith in our jury system.  And the faith we have in our juries is the cornerstone of our democracy.

Why Am I Suddenly a “Member”?

When did I become a member of groups I never knowingly joined?

Maybe I’m mistaken, but I always thought the word “member” meant that one had purposefully joined a club or similar entity.  Joining such a group can be a good idea.  For example, membership in a social club offers a range of pleasant benefits, like golf, swimming, or dining in an exclusive setting.

Being a member of a professional organization generally offers career-oriented perks, along with help climbing a possibly shaky ladder to professional success.  And then there’s membership in a political party, which has always meant sharing an outlook (more or less) on public policy.

But now it seems that my charitable instincts have turned me into a “member” of a host of other groups.  Charities to which I’ve donated a few stray bucks have anointed me as a member, and they are now pursuing me to “renew” my membership!

Jeez, I didn’t realize that when I sent Charity X a small donation I would suddenly become a “member” of that admirable group.  But apparently I did because here’s an envelope it recently sent me, announcing that there are “10 reasons to renew [my] membership”!

The X Foundation, another commendable organization to which I’ve donated a small sum on occasion, has now written me to ask for more.  The shocker is the envelope’s threat that this is my “final renewal notice” this year.  But please tell me, folks, when exactly did I join in the first place?

Yes, I’m a softie, and I have perhaps foolishly sent dribs-and-drabs donations to a wide array of worthy groups that tug at my heart.  But please let me know, Charities A, B, and C, why that constitutes “membership” in your organization?  Yes, you do such wonderful work on behalf of people with diabetes, cancer, or blindness.  But seriously, all I did was send you a check or two.  Yet now you’ve sent me a “renewal reminder.”

Some of these groups have even issued “membership cards,” complete with member numbers.  And a few have gone over the edge and sent me membership “statements” that strongly resemble bills.  I just got one in today’s mail.  C’mon, people!  Do you think that scares me into sitting down and writing you a check?

Truthfully, I resent being lumped into the category of “member” by groups like these, with which I have no real connection other than a desire to add a small amount to their coffers.  Being threatened with a “final renewal reminder” doesn’t induce me to respond.  On the contrary, it makes me wonder about the professional fundraisers these groups have hired.  Do their ominous reminders work on anyone?  They certainly don’t work on me.  Instead of loosening my purse strings, they encourage me to tighten them.  The threatening envelopes get tossed into my recycling more often than not.

All this tossing makes me think hard about the charitable world today.   For one thing, I’ve been warned about an insidious trend.  It appears that if you donate only a small amount, charities tend to sell your name to other groups, so you can then be hounded by ten or twelve charities instead of just one.  “Your name is worth more to these charities than the $25 you give them,” a friend confided.  If that’s true, it simply compounds the problem.  And it seems to account for the plethora of solicitations I find in my overstuffed mailbox every day.

While I’m at it, I’ll go further and denounce some of the other tactics these groups employ.  Even those that don’t call me a “member” are guilty of some pretty odious practices.  First, I am now the recipient of endless “free gifts.”  The proliferation of address labels has gotten totally out of hand.

My daily mail includes countless address labels from groups I’ve never even heard of.  I’ve received enough of these labels to last at least two more lifetimes, and that’s assuming I never move from my current address.  Other freebies include ballpoint pens, note pads, greeting cards, and calendars, many more than I can ever use (and many so unappealing that I never would use them).  Most of these freebies end up in one of my charity donation bags (one hand washes the other?) or, even worse, the trash, adding to our overflowing landfills and our overburdened recycling centers.

Honestly, I’d much prefer that these groups (who are clearly playing the “guilt” card) spend my cash quite differently.  Hey, folks, please use that money to search for a cure, actively fight racism, directly lift women out of poverty.  Note pads and calendars?  I can buy those myself.

Some of us have gotten wise to this endless pursuit of donations. Internet websites can now tell us just how efficiently most charities are run.  For example, Charity Navigator evaluates charities, distinguishing among them by giving four stars to those that operate efficiently, while giving only one or two stars to the charities that spend too much of their revenue on fundraising and other administrative costs.  This kind of ranking may not be perfect, but it’s helped me weed out some of the groups I used to support.  If enough of us did that, we might have an impact on their most deplorable tactics.

The solution for me may be to become more selective.  Instead of making small donations to a wide range of worthy groups, I may focus on a handful of them and send a larger check to each.  But I fear that my name may stay on the same old mailing lists ad infinitum, adding pounds and pounds to recycling as I continue to toss.  I’ve recently started sending back some solicitation forms, demanding that my name be removed from these lists, but so far I haven’t tracked whether that approach has done any good. I truthfully doubt that it has.

This blizzard of charity solicitations has to stop.  Where I once was charitable, I am now more likely to be hostile, vowing never to contribute a dime to most of the charities that pursue me with such zeal.

Don’t these groups realize that they’ve literally reached a point of no return?

[A version of this commentary previously appeared as an op-ed in the San Francisco Chronicle.]