I’m taking a short break from my Pandemic Chronicles entries to indulge in some deep nostalgia, prompted by a Facebook ad touting the online revival of Tower Records, the one-time brick & mortar retail shrine for music lovers. The announcement immediately set me off on a time travel journey going back some 38 years.
Tower’s massive store on 4th Street and Broadway in Manhattan’s Greenwich Village appeared in 1983. In the early 1980s, Broadway was the dividing line between the Village proper and the “frontier” of the East Village. I was a law student at New York University back then, and the law school’s Mercer Street residence hall happened to be only a few minutes walk from Tower. Even though discretionary spending under my tight budget was mainly devoted to exploring New York’s many wonderful bookstores, Tower became a draw as well.
Of course, back then I had no real music set-up, not even a boom box. Throughout law school, my cassette Walkman was my stereo system. Nevertheless, that didn’t stop me from periodic visits to Tower in search of music bargains. I was in awe of the selection. Imagine the endless rows of cassette tapes in every musical category!
I was hardly alone in recognizing Tower’s significance. In a 2016 piece for Medium, “When Tower Records was Church,” David Chiu waxed nostalgic about visits to Tower in the Village:
When you walked into the Tower Records store in New York City’s Greenwich Village neighborhood back in the day, you just didn’t go in there to buy an album and then rush off to leave. To me, going to Tower was like visiting the Metropolitan Museum of Art or attending a baseball game — it required a certain investment of time.
Sometimes it was the overall experience of being inside the store that mattered more than the purchases: the act of walking through the aisles and aisles of music, finding out what the new releases displayed out front, and hoping to meet an musician who was doing an in-store appearance. There was always a sense of anticipation as you went through Tower’s revolving doors underneath the the large sign displaying its distinctive italicized logo because you just didn’t know what you’d discover.
Sigh. The new Tower is online, and even though the variety may well exceed what the brick-and-mortar stores had to offer, it’s not the same. Similar to how I’m feeling about online booksellers and music & video streaming these days, what’s missing now is that wondrous sense of anticipation that came with entering a record store, bookstore, or video store, and making new discoveries. I can’t say that I’d trade in the vast warehouse of popular culture available to us today in return for that on-the-ground retail experience, but it’s a closer call than first meets the eye.
Have you ever moved to another part of your city, state/province, or country? Have you ever relocated to another nation? Why did you do it, and how did you get there?
NPR’s TED Radio Hour had me contemplating this topic during a feature on migration (link here), exploring why and how people have uprooted themselves from their original surroundings to less familiar ones. If you’ve made a big move or two during your life, or are contemplating doing so, this hour offers an interesting set of reflections and insights.
Location and the pandemic
Of course, the idea that location matters has become very significant during the coronavirus pandemic. One’s experience of this pandemic and public responses to it are based in part on where we live. Infection rates, medical and public health resources, population density, and beliefs in science and prevention vary widely by location.
Here in Boston, after a brutal year we are allowing ourselves to take literal and figurative breaths of relief. Our vaccination rates are trending upward, our infection rates and fatalities are in decline, and we’re gradually moving towards some resumption of living more normally.
Yesterday, however, I was on a webinar with law students and lawyers in India. I knew very well that they are reeling from a terrible surge in infections that, for now, shows no signs of abating. We may have been in the same virtual room together, but our experiences of that event were no doubt shaped by our respective perceptions of safety and health.
During my lifetime, I’ve made two bigger moves, a temporary move abroad, and a smaller move that felt like a huge one.
Going in reverse order, the small move that felt very big was leaving my hometown of Hammond, Indiana to attend Valparaiso University, all of one county and a 45-minute drive away. To an 18-year-old young man who wasn’t very worldly, it felt like I had moved halfway across the country, even though I remained squarely in northwest Indiana.
The temporary move abroad was in the form of a collegiate semester spent in England. As I’ve written before on this blog, those five months opened the world to me. Even before that study abroad experience, I had aspirations of moving to the West Coast or East Coast for law school. My semester abroad basically cemented that intention.
A year after returning from England, I would pack my bags for a much longer stay — twelve years in New York City — starting with law school at New York University. In 1994, an opportunity for a tenure-track teaching appointment at my current affiliation, Suffolk University Law School in downtown Boston, prompted a move to my current hometown.
With New York City and me, it was love at first sight. I will never again be as taken with a sense of place in the way that New York captivated me. With Boston, it has been more of an evolving affection, marked by the city’s insularity and parochialism slowly giving way (uh, sometimes kicking and screaming) to a growing cosmopolitan culture. It also helps that Greater Boston remains a place where ideas, invention, creativity, and books still matter. (Two years ago, I reflected on a quarter century of living in Boston. You may go here to read that.)
Many academics, even tenured ones, opt to be somewhat nomadic, moving from university to university as perceived greener pastures present themselves. While I’ve received periodic invitations to apply for teaching jobs elsewhere, I’ve opted to remain in Boston. Whether or not any more big moves remain for me, I cannot guess. But over the years, I’ve also taken countless plane and train trips to places far and near, and I expect that I’ll resume doing so as public health circumstances permit.
I was listening to a favorite album the other day, a collection of Gershwin songs by Michael Feinstein, one of the most devoted and talented keepers of the Great American Songbook. I reminded myself that I first owned this album in cassette format, back in the early 90s.
I further recalled a trip to London in 1992, a quick spring break visit across the pond when I had just began teaching as an entry-level legal skills instructor at New York University, my law school alma mater. Feinstein’s Gershwin album was among four or five tapes that I dropped into my backpack, along with my Sony Walkman portable cassette player. As I traipsed around London that week, I marveled at how entertainment technology now allowed me to listen to favorite albums in the palm of my hand. All I had to do was flip and swap out the tapes!
Fast forwarding to today, I’ve got several dozen albums loaded onto to my iPhone and iPad, downloaded via the MP3 platform. With this latest technology, we can carry a huge digital music collection in our pockets, bags, or backpacks. Way cool.
But here’s the rub: Suddenly, my need for such portability has decreased markedly. I venture out of my home infrequently. I have no idea when I’ll hop onto a plane again.
I’m sure that many of you can relate. If you’re in a part of the world heavily hit by the coronavirus, then you know how our lives suddenly became very small when stay-at-home advisories and social distancing became our everyday norms.
It’s hard for me to grasp that we’ve been at this for only two months or so. This has become a self-experiment of sorts, observing my daily moods while remaining mostly within the confines of my modest condo. So far, I’m doing okay, better than I expected, in fact. Over the past few years, I’ve done a lot of traveling and spent so much time out of my home. It has left me feeling exhausted at times. So in some ways, this solitude has been good for me.
I know I won’t feel that way for much longer, but future choices are largely out of my control. Advancements in public health and medicine will disproportionately shape those options, and for now the timeline is uncertain. While I am genuinely optimistic that we will get a handle on this virus, like most everyone else I must strive to be patient.
I know I’m hardly alone in spending more time watching television during this public health crisis. As I wrote a couple of a weeks ago, I’ve sharply reduced my watching of TV news, and that decision has held. Instead, I’ve been spending time with assorted series, especially highly-regarded police procedurals and historical dramas. Last night, however, I checked out the first episode of “The Last Dance,” a 10-part ESPN documentary series about the Chicago Bulls of the National Basketball Association, centering around the final championship season (1997-98) of its iconic, superstar guard, Michael Jordan.
The series is being televised in weekly installments, rather than being released in its entirety. That said, I already understand why “The Last Dance” is drawing accolades from sports writers and fans desperate to feed the beast while professional and college leagues are shut down due to the pandemic. (As further evidence, the just-completed National Football League annual draft of collegiate standouts earned its highest-ever ratings.) It’s a basketball junkie’s delight. If you’re a sports fan, and especially if you followed the great 1990s Bulls teams, then you’re in for a treat.
For me, “The Last Dance” is prompting a major nostalgia trip. The Jordan-era Bulls teams overlapped with important events and transitions in my life. Jordan first joined the Bulls for the 1984-85 season, which happened to cover my final year of law school at New York University. Even in New York, the sometimes snobby sports intelligentsia knew that this guy in Chicago was something special. Jordan immediately became one of the league’s best players. I began closely following his career and the fortunes of the Bulls from afar.
Alas, Jordan had joined a team in a deep state of mediocrity. The Bulls of the late 1970s and early 1980s were a pretty sad bunch. It would take several years of key player acquisitions and coaching changes — most notably star swingman Scottie Pippen and head coach Phil Jackson — before the team became a serious playoff contender. In fact, not until 1991 would the Bulls win their first NBA championship, the first of six during the halcyon 90s.
By then, I had been practicing law for six years in New York City, first as a Legal Aid lawyer, then as an Assistant Attorney General for New York State. But in 1991, my career was about to shift. I had accepted an appointment as an entry-level instructor in NYU’s Lawyering Program, an innovative legal skills curriculum for first-year law students, starting that fall. I was tremendously excited to be returning to my legal alma mater,as a faculty member, no less! I didn’t know it at the time, but it was the start of an academic career.
I would decamp from New York to Boston in 1994 to accept a tenure-track position at Suffolk University Law School, where I’ve remained since. My devotion to the Bulls followed me, and watching the team’s successes provided welcomed breaks from the demanding workload of a new assistant professor.
The academic calendar would provide greater flexibility in my own schedule, with added opportunities for travel. My fond memories of that team include visits to home in Indiana. My mom, of all folks, had become an ardent Bulls as well. We would watch games together in the TV room, cheering on what would become one of the sport’s legendary dynasties.
As a lifelong Chicago sports fan who puts those great Bulls teams on a pedestal, I look forward to watching the rest of “The Last Dance.” I’m sure it will continue to inspire nostalgic episodes as well. It’s all good, as we comb the memories of our lives during this challenging time.
It has taken a global pandemic to get me to a point where I feel limited by not owning a car.
Here in Boston, we’re experiencing a predicted surge in COVID-19 cases. Sheltering-in-place and social distancing remain the recommended best practices for those of us not working in essential businesses, and I’m taking these directives seriously. Thank goodness that my local grocery store and a number of area eateries continue to offer reliable delivery. But a car would make it easier to take occasional trips for other goods.
It has been over a month since I’ve taken the subway, which during 26 years in Boston and 12 years in New York City has been my primary way of getting around besides walking. I haven’t ordered a taxi or Uber since then, either.
As for having a car, well, I haven’t had a car of my own since 1982, when I left my home state of Indiana to attend law school at New York University, in the heart of Manhattan’s Greenwich Village neighborhood. During college, I owned a 1968 Buick LeSabre, a hand-me-down from my parents. A quick visit to New York during the summer before starting law school easily persuaded me that keeping a car there was neither practical nor affordable. I decided that the gas guzzling Buick would remain in Indiana.
The seeds of my new lifestyle had been deeply planted a year before, during a formative semester abroad in England through Valparaiso University, my undergraduate alma mater, which included a post-term sojourn to the European continent. Walking, buses, subways, trains, and the occasional boat trip became my modes of transportation, fueled by a sense of adventure. In addition, I didn’t have to worry about stuff like parking, upkeep, and insurance.
So, upon moving to New York, I became a happy city dweller and a creature of public transportation. I’ve never lamented a lack of wheels to take a quick trip to the country. In fact, since relocating to Boston, I’ve never traveled to Cape Cod or Nantucket, and I don’t have a burning curiosity to visit either.
In other words, for well over three decades, I’ve felt quite free bopping around cities without a car.
Until now, that is.
This afternoon, I left the immediate area of my home for the first time in a couple of weeks, to walk over to the drugstore for various provisions. Donning safety mask and gloves, I walked up the street, maintaining distance from the handful of others on the sidewalk. With a car, I could’ve completed a more ambitious shopping trip, and maybe hunted around a few other places for those elusive rolls of toilet paper and paper towels.
Honestly, though, I wasn’t unhappy about that. I did, however, feel genuine sadness at the eerie quiet in my neighborhood and the occasional sight of other masked pedestrians on what normally would’ve been a livelier Friday afternoon.
Okay, I’m not about to buy or lease a car because of this. I just hope that between various delivery options and occasional short walks to shop for necessities, I can continue to obtain the goods and supplies I need during this shutdown and any similar stretches, as we wrestle down this damnable virus.
Tara Isabella Burton, in a feature for The Economist’s 1843 magazine last year, serves up a human interest story on an iconic Manhattan institution, the 24-hour diner:
Londoners have their pubs. Parisians have their cafés. New Yorkers have diners – altars to cheap coffee and mayo-spackled pastrami, where you can order a mug at dawn and stay until dusk, where you can hurl invective at the waiters and where they’ll hurl them right back. New Yorkers may be brusque, but at the diner counter, they’ll tell you every one of their secrets before the second cup of coffee.
. . . The diner, after all, is at once the result of New York’s loneliness and its solution. It’s a place where social rules among strangers – no eye contact, no smiling, especially no conversation – are suspended. The greatest diners, like Chelsea Square, are the 24-hour ones that cater to morning workers and midnight drunks, and to the people who find themselves in those sunrise spaces in between.
Yeah, it’s something of a clichéd piece, characterizing the NYC diner as a refuge for loners and eccentrics in a sort of romanticized, 1940s kind of way. Nevertheless, I enjoyed reading it, because it pushes my nostalgia buttons: The 24-hour diner ranks high among the institutions I miss most about living in New York City, where I lived from 1982 to 1994.
During that time, two such places were regular stops for me, the Washington Square Diner on West 4th Street and 6th Avenue, and the Cozy Soup ‘n’ Burger on Broadway and Astor Place. It’s no accident that both are in the heart of Greenwich Village, near the buildings of New York University, where I went to law school. The Washington Square Diner was a short walk from Hayden Hall, then the primary dorm for first-year law students. The Cozy Soup ‘n’ Burger was close to the Mercer Street residence hall, where most second and third year law students lived.
When I visit New York, a meal at the Cozy is a required pilgrimage. I usually order the same thing: A cup of their incredible split pea soup with croutons and a delicious turkey burger. Some of the same guys who worked behind the counter in the 1980s are still there. I also make occasional visits to the Washington Square Diner, where their challah bread french toast remains one of my favorites.
For most of my life I have been a night owl type. Coming from northwest Indiana, the 24-hour city diner was a revelation to me. Good, basic comfort food at decent prices, available around the clock. Awesome!
I’ve been in Boston for some 24 years. While NYC is the city that never sleeps, Boston tends to go to bed early. Although there are many things I like about Boston, how wonderful it would be to see a bunch of 24-hour diners pop up. After all, sometimes a burger or plate of eggs at 2 a.m. just hits the spot.
Even though I’ve been teaching for some 27 years, I don’t get overly enthused about semester breaks. They usually involve a fair amount of grading exams and papers, followed by catching up on other work tasks and getting ready for the next term’s classes. They’re all good, but they’re more of a respite from teaching than a break.
Nevertheless, as the weather gets colder here in Boston and classes come to an end, I do get especially nostalgic about two semester breaks that date back to my own student days.
The first was during my senior year (1980-81) at Valparaiso University in northwest Indiana. Had I planned to spend my final semester of college on campus, I would’ve been at serious risk for developing a bad case of “senioritis” — i.e., playing out the home stretch of my undergraduate career without a lot of enthusiasm. However, I was about to spend my last collegiate semester at VU’s Cambridge, England study center. As I’ve written on this blog, it turned out to be a deeply formative experience.
Of course, I didn’t anticipate how life-changing that semester abroad would be as I completed my fall semester papers and final exams and then left my Brandt Hall dorm room for home. The nostalgia trip for me today is recalling how totally, utterly, completely clueless I was about the experience that awaited me.
In keeping with my procrastinating nature back then (less so today), my preparations for the trip were last minute and minimal. Honestly, I wasn’t even all that curious about England and Europe. I had signed up for the Cambridge semester largely because friends with whom I worked on the VU student newspaper were going. I also welcomed a change of scenery from our small town Indiana campus. (Of course, today I also get nostalgic about those days in Valparaiso. Click here for an essay I wrote, “Homecoming at Middle Age,” published in The Cresset, VU’s journal of the arts, literature, and public affairs.)
As a collegian, although I managed to maintain a certain confident front, in reality I was a jumble of ambition, insecurity, immaturity, and uncertainty over the future. I wouldn’t trade my current level of wisdom (umm, still a work in progress!) for said jumble of that stage of my life. However, it’s kind of neat to look back at that time with the gift of hindsight. As I pondered what to stuff into a suitcase and a backpack, I had no idea that the next five months would shape my personal culture, worldview, way of living, and base of friendships for a lifetime.
The second memorable semester break was during my third and final year of law school (1984-85) at New York University. I was in the job hunt, and my hope was to secure a public interest legal position in New York City for after graduation. During my short time in NYC, I had fallen in love with the city. New York of the 80s was a much grittier and affordable place than it is today. It was possible to enjoy the city on a tight budget. I badly wished to stay.
In addition, I was committed to working in the public interest field. During the previous summer, I was a summer associate at a large commercial law firm in Chicago. The money was great, and the firm treated its lawyers and staff with respect. But my heart wasn’t into corporate legal work, and so I would end up turning down the firm’s offer of a full-time associate attorney position for after graduation. Instead, I returned to the reasons that attracted me to law school in the first place, doing some type of public interest work in the non-profit or public sector.
I interviewed with a wide variety of public interest employers during the fall, and things started to develop during the semester break. During the break I received and accepted an offer for an attorney position from the New York City Legal Aid Society in downtown Manhattan. I was going to be a public interest lawyer in New York City, and I couldn’t have been happier about it! (The realities of paying rent and repaying student loans on a $20,000 salary would come later.) I recall spending a chunk of that break diving into my growing little collection of books about New York City, delighted that I would be staying in my adopted hometown.
Major junctures and events in our lives often don’t appear significant until we can look back at them via the rear-view mirror. Then they become part of our personal narratives. As mythologist Joseph Campbell observed, “when you look back on your life, it looks as though it were a plot, but when you are into it, it’s a mess: just one surprise after another” (Diana K. Osbon, ed., A Joseph Campbell Companion, 1991). That’s how these two semester breaks fit into my story.
Both of these remembrances embrace a post-Second World War, American middle class ideal that has valued higher education as a stepping stone to a better life. I was not fully appreciative of these gifts back then, but I certainly am grateful for them today.
For middle class and working class folks in the U.S., the path to upward mobility that I enjoyed is narrowing sharply. The “college experience” of going away to school, while cobbling together enough money from financial aid, summer and part-time jobs, and parental assistance to make it relatively affordable, has too often given way to sky-high tuition and costs subsidized by significant student loan debt. Many students and their families are pursuing less pricey alternatives as a result, such as two-year colleges and distance learning programs.
Indeed, it may be that Generation Jones (born 1954 through 1965) was the last major cohort to have higher education opportunities that didn’t come with enormous price tags. That reality should inform our potential choices for charitable giving and at the ballot box. Those of us who work in higher education should also be advocates for reducing student debt. We need to ease the financial burdens of higher learning, so that more may have such life-changing experiences.
Okay, college graduates, if you could continue dormitory-type living even after leaving school, would you opt to do so? If your answer is “yes,” then you may be pleased to see this option developing in certain cities.
WeWork, a company that has pioneered the concept of co-working rental office space for entrepreneurial start-ups, is now branching out with WeLive, “communal housing” rentals aimed at recent graduates and young professionals who may find themselves priced out of the housing market in expensive urban areas. Melody Hahm, writing for Yahoo! Finance, explored the new WeLive space in Manhattan:
I thought my college years were behind me. But I’m seriously reconsidering the dorm life since visiting Manhattan’s first-ever location of communal living startup WeLive.
Of course, the concept of communal housing isn’t novel. . . .
But this isn’t your typical dorm situation: You have your own apartment but get access to a chef’s kitchen, yoga studio, conference room, laundry/arcade room, and neighbors who actually want to talk to you.
In many ways, WeLive looks and sounds like a post-graduate residence hall, at a premium price:
The layouts in WeLive’s 400 units range from small studios to four-bedrooms, and all apartments come fully furnished. Per-tenant pricing begins at $1,375 but if you want a bit more privacy, you’ll have to dole out at least $2,000 per month. The most common setup is the “studio plus,” which comes with two beds (one is a Murphy hidden in the wall); these range from $2,500 to $2,800. A flat monthly utilities payment of $125 covers electric, water, cable, wifi and cleaning costs (yes, housekeeping is included).
Here’s how WeLive describes itself on its website:
WeLive is a new way of living built upon community, flexibility, and a fundamental belief that we are only as good as the people we surround ourselves with. We know life is better when we are part of a community that believes in something larger than itself. From mailrooms and laundry rooms that double as bars and event spaces to communal kitchens, roof decks, and hot tubs, WeLive challenges traditional apartment living through physical spaces that foster meaningful relationships. Whether for a day, a week, a month, or a year, by joining WeLive – you’ll be psyched to be alive.
If you’ve read this blog for any length of time, then you know that I’m fond of sharing nostalgic moments from my college and law school years. I can even get a little soggy over memories of dorm life. At Valparaiso University in northwest Indiana, my undergraduate alma mater, I lived in dormitories, with the exception of a final semester spent in a study abroad program. At New York University in Manhattan, my legal alma mater, I lived in law school residence halls throughout my stay there.
When I graduated from NYU Law, bound and determined to save the world as a Legal Aid lawyer (and with a $20,000 salary to remind me of my lofty idealism), my Manhattan housing options were practically non-existent. Consequently, I followed the trail blazed by other young denizens of the city’s non-profit sector and crossed the bridge into Brooklyn for a relatively cheap apartment share and a long subway ride to work. My first place was a three-bedroom apartment share. I believe the total rent, split three ways, was $1,000.
Those affordable Park Slope apartment shares are no more. The brownstone rentals so popular among my fellow Legal Aid colleagues and others similarly situated are now homes commanding high six and even seven figures in the current real estate market.
And so comes the market opportunity for WeLive. With more bohemian living options no longer available in places like New York, WeLive steps into the void and offers young, hip, and conveniently located housing options aimed at Millennials. Measured against the cost of living standards of almost any other area, WeLive is still pretty expensive. But to find a comparable rental in New York, your daily commute might start to resemble a sojourn.
When I moved to New York in the 1980s, gentrification and higher living costs were very much a part of the civic dialogue. Today, however, the housing costs are mind boggling. New York is not alone in this reality, at least among high demand urban places. This is definitely the case here in Boston.
It’s why ventures like WeLive are getting attention. In reality they are expensive versions of what 50 or 75 years ago would’ve been called boarding houses, with a dose of social selectiveness built into the marketing: WeLive is a new way of living built upon community, flexibility, and a fundamental belief that we are only as good as the people we surround ourselves with.
Personally, I’d rather have affordable apartment shares in Brooklyn, but I realize that time has passed.
If you’re wondering whether we’re still in a Golden Age of television drama, then look no further than “The Americans” on FX. The show stars Keri Russell and Matthew Rhys as deep-planted Soviet spies, fronting as parents of an All-American family living outside of Washington D.C. during the early 1980s.
During its first three years, I’ve regarded it as an excellent drama, though perhaps a step below iconic classics such as “The Wire.” In the current season four, however, “The Americans” has moved to the next level. It is delivering other-worldly acting, morally complex storylines, and the look-and-feel of 1980s America and the last decade of the Cold War. It is riveting entertainment.
(SPOILER ALERT: If you’re a regular viewer but haven’t caught the recent episodes, you might want to come back to this later!)
In the last episode, many of the major characters find themselves gathered around their television screens, watching the November 1983 airing of “The Day After,” an ABC made-for-TV movie depicting the devastating effects of a Russian nuclear attack on America, centering on the small city of Lawrence, Kansas. Here’s how reviewer Hank Stuever described the episode for the Washington Post:
On “The Americans,” the characters all watch in stunned silence, including secret Soviet spies Philip and Elizabeth Jennings (Matthew Rhys and Keri Russell) and their kids, Paige (Holly Taylor) and Henry (Keidrich Sellati), and their friendly next-door FBI agent, Stan Beeman (Noah Emmerich) and his son, Matthew (Daniel Flaherty). Even the Russians who work at the Rezidentura in Washington tune in – Oleg Burov and Tatiana Ruslanova (Costa Ronin and Vera Cherny) watched it curled up on the bed.
In real life, watching “The Day After” was a very similar experience. The Post‘s Stuever writes:
As seen on . . . “The Americans,” people really did set everything aside on the night of Sunday, Nov. 20, 1983, to watch ABC’s depressingly sober TV movie “The Day After.” It told the story of a handful of people in and around Lawrence, Kan., who had the misfortune of surviving an all-out nuclear war between the United States and the Soviet Union.
Please understand: This was a television event in a way that we’ll probably never see again. You see, when it comes to watching stuff on screens, we’ve become increasingly atomized (pun intended). A century ago, a theater would bring together hundreds of people to watch movies and newsreels. Fifty years ago, folks watched programs with family and friends in front of television sets. Today, we often watch TV programs and movies in different rooms of the house or stream them on our various devices.
Americans viewed “The Day After” huddled together in front of our television sets. I was in my second year of law school at NYU, and a big group of us gathered to watch it in one of our dorm rooms. Today, with so many actual horrors captured on video and posted to social media, it may be difficult to grasp that this fictional movie was jarring and upsetting to many. But such was the case. We talked about it for days.
The threat of nuclear annihilation was very real during this final decade of the Cold War. That same academic year, I was on the staff of NYU’s international law journal, one of several student-edited law reviews published at the law school. A few weeks before the airing of “The Day After,” our journal hosted a panel discussion on nuclear arms control, moderated by McGeorge Bundy (National Security Advisor to President Kennedy) and featuring a prominent group of diplomatic and military policy experts.
The subject matter of the panel extended way beyond mere academic speculation, and I recall a tone of seriousness in the dialogue among the speakers. The U.S. and the Soviet Union were playing a dangerous cat-and-mouse game around the world. “The Day After” would show us what that meant with more dramatic effect. This episode of “The Americans” brilliantly brought back those memories.
A personal note: At NYU Law, I volunteered to help out with that nuclear arms program. By doing so, I got to join the pre-panel dinner with the speakers in the faculty dining room. I mostly kept my mouth shut out of fear of saying something stupid. Nevertheless, I was thrilled — I was sharing a meal with some heavy hitters — this was cool stuff for me!
In return for this perk, however, I was handed a lot of grunt work by the senior editors, for our main post-panel task was to transcribe and edit the proceedings for publication. I was assigned to proofread the transcript and to generate dozens of footnotes documenting facts and events described by the panelists, a hefty pre-Internet research project that ate up countless hours in the dungeon level of the law library.
Working on this journal was my introduction to graduate-level scholarly work. I can’t say that the experience of developing footnotes from scratch to verify the accuracy of the panelists’ remarks was all that enjoyable! In fact, it persuaded me to pass up an opportunity to serve as a senior editor during my third and final year of law school. I didn’t want to spend my last year at NYU editing more manuscripts and chasing down more sources in the library.
Now, of course, I’m writing hundreds of footnotes for my own scholarly articles. What a twist! Let’s just say that this was not a foreseeable development during my more anti-authoritarian law school days.
For many years I’ve quipped that Introduction to Typing and Driver’s Education were the two most valuable courses I took in high school. Actually it’s more than a quip. If you toss my junior year American History course into the mix, I think you’d have the academic holy trinity of my high school career. (Yes, I was something of a rebellious underachiever in high school.)
Anyway, back to typing class: I really wanted to learn how to type. Even as an adolescent, I felt that typing out my thoughts and ideas would somehow render them more, well, significant. Once I learned how to type, I would use my mom’s old Royal manual typewriter to bang out term papers for school. And when I got involved in the student council, I would learn how to cut mimeograph stencils for printing out the council newsletter.
Of course, just because I enjoyed typing doesn’t mean I was good at it. I made lots of mistakes…and still do. In the ancient era before word processing programs and home computers, that usually meant using either liquid paper or Ko-rec-type to cover up one’s mistakes and then type over them. I did this a lot, and it slowed down my typing speed.
Off to college
When I went off to college at Valparaiso University, my main off-to-school present was a Smith-Corona electric typewriter. Whoa…..I was now moving up in the world! This model used ribbon cartridges instead of old-fashioned spooled ribbons. If you made a typing error, you could swap out the ribbon cartridge for a correcting cartridge that would white out the mistake. It is a miracle that I did not develop a repetitive stress injury swapping out those cartridges.
My typing life changed when I joined the staff of my college newspaper, The Torch. You see, the newspaper office had two IBM correcting Selectric typewriters. Typing on those machines was a sublime experience. During down times when folks weren’t working on stories, we were free to commandeer the typewriters for our papers and projects. The presence of those typewriters is one of the reasons why that office became our unofficial hangout, even when we weren’t working on the newspaper.
Now, those of later generations might not fully appreciate these challenges, but writing term papers and other assignments in the B.C. era (Before Computers) was a very, very different experience, especially when minimum or maximum page limits were in play. Most of us would first write out our papers in long hand, and then estimate if the cumulative sheafs of paper would, when typed up, potentially run afoul of the page limits. If you didn’t have a good sense of how your cursive writing translated into typed pages, you might be in for some unpleasant surprises, leading to late nights before papers were due.
Lugging it to NYC
I took my Smith Corona with me to law school at NYU. I cannot recall how I got that heavy, bulky machine to its destination, but I may have even checked it as part of my baggage for the flight from Chicago to New York. In some ways, these challenges have not changed; even in the digital era, there are only so many ways to move one’s belongings from here to there.
This was right before the home computer revolution, and very few of my classmates had PCs. Most of us continued to type our papers, with added challenges in terms of margins and page length when writing out practice versions of legal documents. By this time, we were overlapping with the emerging age of computers. At NYU I worked on one of our scholarly law journals and on the law school student newspaper, and we had computer word processing capabilities for both publications.
A computer of my own
I would not own a personal computer until several years after graduating from law school, a Commodore 64 that supported a superb game library and rudimentary word processing programs. I would later move up to an IBM PC compatible machine, and at that point I transitioned from typewriter to word processing. I became enamored of the wonderful, awesome WordPerfect 5.1 for DOS program, which remains to me the best ever software package for writing productivity. In fact, ever since being more or less forced into using the tyrannical, control-freakish, and cumbersome Microsoft Word, my writing efficiency has declined.
Today, I’ve morphed over to Apple products, but I’m still stuck with Microsoft Word. Someday I’d like to give a serious tryout to Scrivener, a word processing program that has a fiercely devoted following. As for my blogs, I use the WordPress platform, which I find easy to navigate.
Changing technologies aside, it’s clear to me that my original motivation for learning how type — to share my thoughts and ideas — remains the main reason why I’m sitting before a keyboard today. And thank goodness that you, kind reader, get to read what’s on my mind with (most of) the typos cleaned up.