Good news first, I have a new essay out in Uncanny Magazine, “Expanding Our Empathy Sphere Using F&SF, a History,” where I talk about my term ’empathy sphere’ meaning the collection of beings we consider coequally a person with ourselves, something which historically has expanded over time, and which is useful in thinking about why when we read old utopias, like More’s Utopia, or early SF utopias, they often don’t feel utopian to us anymore if they don’t have freedom for groups that are inside our empathy sphere but weren’t inside More’s (like lower classes, women, certain races, clones, A.I.s etc.). It’s a useful analytic term and one several people have asked me to write about, and I also give a history of how SF has helped expand this sphere over time. I hope you enjoy reading it!
Less good and more personal news next, my health has taken a bad turn, bad enough that I have taken medical leave and had to cancel my fall teaching. My medical team is still running tests (U Chicago has an exceptional hospital), and they don’t think it’s life-threatening, but it’s probably a circulatory system issue, with symptoms including severe dizziness, faintness, stumbling & falling, all of which make it very hard to do anything, including teaching. They’re still running tests, and generally hopeful that things will improve, but on a scale of months, not weeks or days. I hope to be well enough to teach in spring. As for writing, I’m doing some, since one of the hard things in this situation is to keep my morale up and nothing nothing nothing makes me happier than writing, but it’s still being slowed, alas, though may pick up a bit as the glut of start-of-leave tasks diminishes.
So I wanted to share some reflections on this.
One is that it is amazing how much of the resistance to taking medical leave came from me, not others. Even when friends, colleagues, disability staff at the university, and family were all encouraging it, even when I confirmed my employer policies meant I could do it w/o a bad hit to income etc., even when I was in the doctor’s office and the doctor checked a couple things and the first words out of her mouth were, “Well, you can’t work!”, even when the doctors took it so seriously they wouldn’t let me walk out of the office but insisted I wait for a wheelchair, I still immediately started protesting about, “Well, if I teach remotely from lying down… but this course is special… but if I have X accommodation…” etc. arguing back even against such reasonable arguments as, “Your body is failing to deliver oxygen to your brain! You know what you need to do anything?! Oxygen for your brain!” Nonetheless, it took many days, much encouragement, and many repetitions of exhaustion & collapses for me to decide that, yes, everyone urging me to take medical leave did indeed mean I should take medical leave. (Important principle: in teaching all courses are special/unique, if you make exceptions for that you’ll never stop making exceptions.)
Where did my resistance to taking medical leave come from, when I was in the extraordinarily fortunate position of my employers, doctors, and family all being 100% supportive? (a rare and lucky thing). Partly it came from not wanting to let others down, partly from not wanting to admit to myself that it was serious, but a big part of it also comes from narratives, from The Secret Garden, from Great Expectations, from a hundred other narratives, some classic some recent, in which chronic illness/weakness/invalidness is all in one’s head, or where it’s “overcome” by force of will or powering through the pain, so that even in the fortunate case where everyone around me was being supportive and great, those narratives of powering through were unconsciously deep inside me feeding my resistance to accepting that my doctors and employer aren’t exaggerating when they say, “Don’t work.” This connects to something I discussed in my second-most-recent Uncanny essay, on the Protagonist Problem, that it’s very important to have a variety of narratives and narrative structures, and it can do real harm if one type of narrative or structure dominates depictions of a topic. Some versions of this have been discussed a lot recently: back pre-Star Trek, when close to 100% of black women depicted on TV were housemaids, it did harm by reinforcing bad stereotypes & expectations; similarly today when a very high percentage of immigrant characters depicted on TV are shown committing crimes, it feeds bad expectations. In the Protagonist Problem essay I argue that it also does harm when a large majority of our stories show the day being saved by individual special (often chosen one or superpowered) heroes, since it feeds a variety of bad impulses, including the expectation that teamwork can’t save the day, and feelings of powerlessness if we don’t feel like heroes; the argument isn’t that protagonist narratives are bad, it’s that protagonist narratives being the vast majority of narratives is bad, because any homogeneity like that is bad, just as it’s important for us to depict many kinds of people being criminals on TV, not a few kinds overrepresented and others erased.
Thus, for disability, we also have a problem that depictions of disability tend to repeat a few stock narratives, not one but three really, which together drown out others and dominate our unconscious expectations. One form is is the disabled/disfigured villain, a holdover from pre-modern ideas about Nature marking evil with visible indicators (and virtue with beauty). Another is a person falling ill and dying, a tragedy, which ends up focusing on the friends and loved ones who help along the way, or who survive. Another is ‘inspiration porn’ (David M. Perry has great discussions of this) which has a few varieties but tends to focus on how heroic an abled person is for helping a disabled person achieve a thing (like Secret Garden where she gets him out of the chair) instead of on the disabled person’s achievements/experience, or to present “Look a disabled person did a thing!” but in a weirdly dehumanizing way, the same way you would write “Look, this monkey can play chess!” All of these make people resistant to accepting the label disabled, since, even though it’s really useful once you have (I had trouble for a long time) we associate it with being morally bad, being doomed, or being helpless and dehumanized.
The disability narrative most relevant in my recent situation, though, are the stories of ‘overcoming’ disability, where a person is either cured (through their own efforts or others’), or works hard and pushes through, so the disability becomes a problem of the past, that has been left behind. This often-repeated narrative (present in fiction and nonfiction) encourages the attitude of seeing disability’s disruptions to life as temporary and surpassable. It means that, when I get a new diagnosis, my first thoughts even this many years into having chronic illness, are always about how long it’ll be until I overcome it, what I need to do to get past it, the expectation that it’ll be normal by spring/summer/December/whatever. This often leads me to delay by weeks or months or longer taking steps to, for example, adapt my home to be more comfortable (like getting a lap desk so I can work lying down), and other changes dependent on expecting the condition to be here to stay. I think, as a culture, we really hate telling stories about illnesses and disabilities that are here to stay.
I remember a conversation with a friend once about a situation where a medication good at treating their particular condition was taken off the market, and the parents of a kid with the condition contacted my friend to ask how to advocate or find other ways to get more of the medication, and the friend had to keep saying no that wont’ work, no you can’t get it, no you really can’t get it, no your doctor can’t write a special note, until finally they asked directly, “So what do we do now?” to which my friend answered, “Accept a lower quality of life.” That phrase crystalized things for me. I think in many ways no ending is scarier for us in narrative than accept a lower quality of life. It isn’t a one-time tragedy like death, we have good narrative tools to write tragedy, and to transition focus to the characters who live on, commemorate, remember. Accept a lower quality of life in a story means losing, giving up, surrendering, all the things we want our brave and plucky characters to never do, and then having to live with every day being that much worse forever. It’s neither a happy ending nor a tragic ending, it’s a discouraging ending, and we rarely tell those stories.
I vividly remember the first story like that I ever met, it was a James Harriot All Creatures Great and Small story, about a man whose family had been coal miners, who really wanted to farm, and bought a farm, and worked tirelessly to do a good job, and was a really nice person and always kind and earnest (unlike a lot of the characters in the stories), but then his cows got sick and James tried everything he could to cure them but it didn’t work, and then the farmer came to tell him, with a calm demeanor, that he was selling the farm and had always promised his father he’d go back to coal mining if “things didn’t work out” (coal mining which in the 1920s-30s meant a much shortened life expectancy as well.) James realizing how huge this was (accept a lower quality of life) despite so many efforts said, “I don’t know what to say,” and the farmer answered, “There’s nothing to say, James. Some you win.” I still tear up just thinking of that scene, the cruel unspoken and some you lose applied to a whole long life-still-to-come, every day of which would be worse, and there was no other way. A big part of modern advancement is about avoiding there being no other way–offering insurance, social safety nets, appropriate grants–but it’s also an important type of story to tell sometimes, and one I really needed some examples of. Why? Because those stories, those phrases in my memory (some you win, and, accept a lower quality of life) are not where I think I am now, I’m still working hard on treatments and therapy etc., but I needed to have them in my palette of expectations of things that could be the case, to help me plan. I needed those at the start of term to get out of the, “But surely it’ll get better in a couple weeks if I work hard,” mindset to the better attitude of, “The doctors don’t know how long this will last, I’d better plan in case it lasts a long time.”
If the only outcomes in our expectations are (A) powering through and it gets better, or (B) death/villainy/helplessness-forever, none of those archetypes will give us the sensible advice that it’s wise to plan long-term just in case there is a long-term thing that impacts quality of life. Because today a lot of those can be addressed with adapting tech/stuff/habits. I put off buying a lap desk for 2.5 months this summer, struggling to work lying down, since I didn’t want to waste the money if I was about to get better. But having a lap desk and turning out not to need it is much better than needing one and grinding on without. I also put off adapting the area around my bed to optimize for work, put off getting the new screen which finally today (Oct 7, I started wanting this in July!) got installed so I can have multiple monitors while lying down. I put off realizing that instead of watching chores pile up expecting to catch up when I got better, the household needed to discuss and make changes to reduce the total load of chores (simpler meals, paper plates, self-watering planters, planning! Also: thank you so much Patreon supporters, you made my new lying-down desk and canes and such possible!!).
The some you win stories are extremely sad and shouldn’t become our dominant narrative, but they need to be in the mix, one color in the color wheel, to help people who do face disability to weigh the odds better, and not think well, in 90% of stories I know the person gets better so probably I’ll get better and this [desk/ screen/ cane/ adaptation] is likely to be a waste of money. Because you now what’s a good thing even if the end of one’s real life story is accept a lower quality of life? Accepting a quality of life that’s only 5% lower instead of 20% lower because you’ve adapted your home/ routine/ desk/ fridge/ breakfast routine etc. to mitigate as much of the negative impact as you can. So here I am in what is probably the best possible lying-down desk, writing and producing more than nothing, but I sure would’ve produced more over the last few months if I’d done this sooner. And I also would’ve been a lot more willing to say “You’re right I should take medical leave,” if I had believed my odds of recovering quickly were, say, 50/50, instead of, as narrative tells me, expecting that if I tried hard it was certain that I’d quickly power through (and that if I didn’t recover quickly that heralded either moral weakness, helplessness, or death, three things our minds work very hard to resist). A broader mix of disability narratives whispering in the back of my unconscious mind, telling me there might be many outcomes and I should plan for many outcomes not just for the best, would have done so much good–that’s why we need variety.
As a coda to this discussion, chatting about it with Jo Walton, she pointed out that both my examples of accept a lower quality of life stories are nonfiction (Herriott’s fictionalized from real life, the other just real life), and that after she and I first discussed the Herriott story she tried hunting for examples of that kind of story far and wide but basically never found them, that she often found it as “a Caradhras, a mountain you can’t get over so you go under, never the end.” But recently she found several examples in the work of the extremely obscure and neglected Victorian writer Charlotte M. Yonge; it’s great to find one, but also to have confirmation from a voracious reader about how rare such narratives genuinely are.
Now, my other reflection is on academia not disability things.
When I finally decided on taking leave I joked to myself, “For academics, ‘vacation’ means when you do the work you really want to do, and ‘medical leave’ is when you actually vacation.” But the reality is that even medical leave I’ve been finding myself doing minimum four hours of academic work a day, sometimes much more. It has been an interesting chance to see, both which specific parts of academic work absolutely can’t be cancelled or handed off to others, and just on the sheer volume of time that academics are required to give to things which are neither teaching nor research. Letters of recommendation wait for no man, ditto letters for other scholars’ tenure files, and mentoring meetings with Ph.D. students about their urgent deadlines; it’s one thing to set aside one’s own agenda but another to neglect things that other people really depend on. So here I was on full disability leave, with all teaching and research obligations on hold, something my university was quickly able to give, and yet I found myself working intensively from waking until dinnertime and still falling farther and farther behind even when the only work I did was letters of recommendation and inescapable paperwork. In other words, at least when rec letter season is upon us, the paperwork and mentoring parts of academia are pretty close to a full 9-to-5 job even without teaching or research! And that is for someone tenured at U Chicago, one of the most privileged teaching positions in the world, with a light load at a very supportive university.
As one friend put it, “I’m not a teacher, I’m a full-time e-mail answerer,” another, “I teach for free, it’s the grading and admin they pay me for,” another, “You can either produce research or keep up with email, but you can’t do both.” We need to factor this in as we think about how academia functions and what reforms to push for, and into how we teach Ph.D. students since things like email skills and time-management skills are absolutely essential to teaching and research when they need to be balanced basically like hobby activities squeezed into the corners of time we can scrape out around the full-time job of admin. It doesn’t have to be this bad. Possibly the problem is best summarized when I was talking to people about a high-level search committee (i.e. hiring at tenured full professor level instead of junior level) and they said they weren’t going to ask for letters of recommendation until they got to the short list of finalists and would only ask for letters for those few, not everyone, “Because we want to respect the time of the important people writing the letters.” Subtext: we don’t respect the time of the less-high-status people writing the many hundreds more letters needed for junior hires. I genuinely think every academic field would produce another 80+ books per year if we just switched to only requesting rec letters for finalists instead of all applicants, and that’s just one example of a small change. In sum, anyone near academia needs to acknowledge that the real pie chart of academic work is depicted below, that we need to plan for that and remember that small changes to self-care or workflow (just studying up on gmail tag and shortcut things for example) can make a huge difference to reducing the unreasonable load and avoiding burnout, and above all that we should always remember that phrase–respect the time of the people doingX–when we plan how to organize things (syllabus, meetings, forms, applications, committees, etc.). And I’m sure a lot of this applies far beyond the academic world as well.
Meanwhile, between recommendation letters I can’t get out of writing, plus disability paperwork, doctor’s appointments, and working on getting my home adapted so my quality of life is diminished by a little instead of a lot in my present state, I’m definitely working-rather-than-resting more than 40 hours a week, and that’s a pretty typical illness experience. It’s good to know that going in, accept it, plan for it, carve out time for the inescapable tasks and to think of adapting the home as time-consuming (or something we should ask for help with!) Otherwise it’s very easy for a week, or month, or three months of ‘rest’ to be not at all restful, and the hoped-for ‘recovery’ to remain elusive. I still have three months of leave before me and I’m definitely leveling up at how to make my leave actually be leave (delegating, adapting things, finding others to write letters when possible) but learning how to make leave actually be leave, and rest actually be rest, is definitely a skill one must level up at, and I think if we understand that it’s a skill (and perhaps tell stories about it?) we’ll be better at realizing we need to actively work to learn it when we (or loved ones) need that skill.
So, for now, I’ll be focusing on rest, and doctor’s appointments, and home adaptation, and things to keep my morale up, and writing (keeps morale up!), and getting ready for the release of Perhaps the Stars (!!!!!!!) but I hope these reflections are helpful, and many thanks to everyone who’s been supportive & helpful throughout. I’ll see you soon when I’m either (A) better or (B) fully adapted to a partly-but-minimally-lower quality of life.
First, I’m excited to announce that you can now pre-order the first segmentof the new cast recording audiobook of Terra Ignota that’s being done by Graphic Audio. I’m really excited about this new audio set, which is doing the whole series broken down into volumes so this first release is the first half of book 1. I’ll write about it at greater length soon, but what I love is how the different voices with their different accents make you so much more aware of the global/international nature of the characters and setting, and the amazing director Alejandro Ruiz worked with me on some really exciting experiments with gender and casting, casting a lot of roles against what one might expect, so that the voices and physical descriptions and pronouns are all mismatched, enhancing the way Mycroft’s strange use of gender in the narration disrupts the reader’s perception of character gender, inviting the reader reflect on how perceived gender affects our feelings toward characters. We also got to do some really great representation in the casting, including not only race and nationality, but also a fantastic nonbinary performer doing Sniper and a brilliant trans woman doing Carlyle. I’ll reflect more later on but I’ve stayed up irresponsibly late more than once being unable to stop listening to the audio files, so if you enjoy the books and enjoy audiobooks I think you’ll love them! (Though the best of all possible Terra Ignota experiences definitely also involves listening to the Derek Jacobi audiobook of the Fagles Iliad right before you read book 4…)
Meanwhile… Why I Care About Barbie’s Career of the Year:
This topic is very far from my usual bailiwick but important in its odd way, expanding on a Twitter thread from 2020. I am not, nor have I never been, a Barbie collector, but I find the Career of the Year series fascinating as a metric of public attitudes toward feminism. In the broad spectrum of feminist discourse, the fringes and harsh or stinging voices are often the loudest (the progressive left & conservative right), making the Mattel Barbie team an informative contrast.
Generally Mattel’s team wants to present Barbie as a feminist trendsetter but in a centrist way, a model of forward-thinking but non-controversial feminism, and it’s fascinating to watch that metric evolve.
Mattel knows that what it includes or excludes in the Barbie line gets attention and has a political impact, and knows it’s doing and who it’s offending/pleasing, in terms of both profit-seeking and messaging, when it creates things like its new trans-friendly nonbinary Creatable World doll kits which make gender-mixing easy:
But while producing something like that means Mattel is taking a stand in one sense, it’s notably not in the main Barbie line. The Barbie line itself tends to be a bit more cautious, especially with Barbie herself.
Since the Barbie Career of the Year doll is designed to be the most discussed, and aims to make an impact in its claims about the attributes of an ideal female role-model, it is a fascinating reflection of what a group of decision-makers who are almost all women feel is the right focus for their annual feminist-yet-centrist message about what girls and women should aspire to. The bodily diversity of the Barbie line has been growing steadily, as shown in the image below of the 2019 line with its range of body types, racial characteristics, and disability representation, but the politics of careers is fascinating separately.
The 2020 discussion (slightly tidied w/ 2021 addition at the end):
Barbie’s 2020 Career of the Year is (for the first time) not a single Barbie but a team, a Political Campaign team featuring 4 dolls: Candidate, Campaign Manager, Fundraiser, and Voter, with diverse race and body types. Interesting to compare to past Career of the Year Barbies.
I’ll give the list of past ones first, then some analysis. Barbie had already had many earlier careers, including astronaut, president, business woman, and others, but the formal Career of the Year series launched in 2010:
2010 = Computer Engineer
2011 = Architect
2012 = Fashion Designer
2013 = Mars Explorer
2014 = Entrepreneur
2015 = Film Director
2016 = Game Developer
2017 = no Career of the Year doll for 2017 that I can find
Barbie has had a lot of careers over time, including earlier iterations of astronaut and president, & her 60th Anniversary Career set has astronaut, firefighter, soccer player, airline pilot, news anchor, and “political candidate.”
2016 was the 6th presidency-focused Barbie, making her political career thus:
1992 = President
2000 = Presidential Candidate
2004 = Presidential Candidate
2008 = President
2012 = “Barbie for President”
2016 = President & Vice President candidate pair
2019 = Barbie 60th Anniversary Career series repackage of the 2016 presidential candidate doll in a different box w/o her VP running mate & with lighter skin & straighter hair. Fascinating.
2020 = Campaign TEAM and NOT SPECIFICALLY PRESIDENT
Note that 1992, 2000, 2004, 2016 & 2020 all candidates, while 2008 & 2012 are Presidents, i.e. already victorious rather than running, an interesting choice for the window right after Hillary lost to Obama in 08 primary.
The 2016 and 2020 political Barbies have variety in skin tone and hair color, and 2020’s has variety of body type as well, in line with Mattel’s recent changes, whereas 1992-2012 are all distinctly the original blonde Barbie moving into a political career.
But no earlier president or presidential candidate Barbie was in the Career of the Year series, which is the most visibly political moment of Mattel’s year, the Barbie choice they expect to get the most discussion and spark the most newspaper coverage etc.
Career of the Year started as something of a disaster in 2010 when the well-meaning Computer Engineer Barbie, winner of a voting contest to pick the first Career of the Year, was launched w/ a badly-thought-through accompanying book which focused on her repeatedly messing up and needing male programmers to fix her machine (even to get rid of a simple virus!) and to turn her concept into a game.
When Architect was next (2011) I remember thinking about the fact that (at the time) when you looked at lists of college majors by expected salary, architect was usually listed as the highest-paid major for women.
In fact the story is really cool, this great article discusses the campaign for architect Barbie, effort to convey power via her glasses & hardhat (which she never wears in the photos), & the experiments presenting her to girls to make sure the doll’s professional skill went unquestioned, unlike her computer engineer predecessor.
2013 Mars Explorer was the 1st mission-specific space Barbie though there had been several astronauts.
She was pinker than average though less sparkly than average, and accompanied by many science facts. Below are three earlier astronaut Barbies, for contrast:
2014’s “Entrepeneur” is strangely vague, forgettable, and was much mocked at its release with headlines like “Entrepreneur Barbie will Inspire Girls to Be Vaguely Ambitious.” It was very well researched underneath, made in consultation with some major global feminist leaders like Reshma Saujani, the founder of Girls Who Code, but it struggled to get a clear concept across.
The vagueness of Entrepeneur Barbie for me is an exposure of the strained path-to-wealth archetype in our society, since it’s so much about networking, pitching, acquiring companies, buying out rivals, moving money around rather than making things; hard to describe on a box. The fact that there’s no comprehensible clear thing an entrepreneur Barbie would do or make, other than have money and move money around to make more money, is an example of how hard it is to communicate to kids how power and money really work (and how nonsensical it often is).
2015 Film Director was a clear response to discussion happening at the time about how few female film directors there were in Hollywood. She did come with several types of hair and skin tones (version options on some of these are hard to trace).
2016 Game Developer was a direct effort to redo and recover from the mistakes of the 2010 computer engineer. She (left) has one of the least pink least feminine outfits ever on a barbie, a silver (not pink) laptop and much more technical info on the box. 2010s on the right.
Here you see game developer Barbie next to her unsuccessful software predecessor:
One Casey Feisler on Flickr did this great compilation comparing the packaging for both plus for the 2018 Robotics Engineer career of the year Barbie. Note the new focus on precisely what she does herself not leaving it to colleagues.
The robotics engineer doll was very similar, still glasses and a laptop, but notable for the black variant being used in publicity images a lot more, almost 50/50 with the white version.
I’m still trying to figure out why there appears to have been no Career of the Year Barbie for 2017. I’ve looked and looked at this strange gap between the two very similar engineering dolls of 2016 and 2018. I’d love for someone to solve the mystery. It’s worth remembering that Mattel had very excitedly made their all-female presidential ticket President & Vice President pair in 2016 and seemed really invested in Hilary’s campaign to be the first female president – was it a morale thing that slowed them down for 2017?
Regardless, after 2018’s robotics engineer, Mattel exploded into the super political with 2019’s Judge Barbie, a clear and extremely not-neutral reference to the activities of RBG on the Supreme Court, and Republican-led stuffing of the courts. And as you can see in the image below, the 2019 judge Barbie, like the presidential set from 2016 and the two engineering Barbies, actively spotlighted its increased diversity in hair color and skin tone in its media:
Which gets us to 2020’s Career of the Year team, the first team and, so far as I can tell, the first politically active Barbie that isn’t focused on the presidency specifically, but could be running for congress, senate, local office, anything.
The distribution of race and body type was clearly carefully calculated, with an African American candidate, a medium-skin-toned POC-looking voter (could be Latina, First Nations, many things), and Mattel’s new heavier body type for the blonde in the role of fundraiser. It’s about teamwork, both the idea that a successful campaign requires many people beyond the candidate, & about the importance of many kinds of races, continuing Judge Barbie’s turn toward branches of government beyond the Presidency.
So many Barbie careers are about celebrity (actress, singer, rapper, princess) & the Presidency is a celebrity position (more so under Trump) so the break with celebrity & focus on non-famous staffers & voters & less spot-lit races is a bigger change for Barbie than it may seem. Mattel’s goal is clear, their contribution to the turn-out-the-vote movement, but I think the attention to teamwork and the importance of non-celebrity people, of the people who aren’t the center of attention, has a potential power beyond the political.
Careers of the Year have always been the one in the spotlight: the director, the architect, the designer, the one who steps on Mars, the president, with little discussion of being on the team, or the fact that movies, buildings, Mars missions are teamwork. So after six presidents or presidential candidates (and one VP) and many other Barbies-in-the-spotlight I hope this teamwork focus will help girls feel like they’re powerful even if they aren’t on the stage, in the spotlight, or in charge. A good message.
My 2020 summary thought: Keep it up, Mattel! This year’s team is great, let’s see more Career of the Year teams! Design teams, surgical teams, the Mission Control team, crisis intervention teams, pharmaceutical development teams, publishing teams (author, editor, publisher, publicist)!
My 2021 addendum: Barbie’s 2021 Career of the Year, music producer, is less remarkable than the last two in many ways, another iteration of Barbie with a pink laptop and headphones, which seems to be Mattel’s signature for Barbie-in-tech, though this one also has the music levels slider board:
Yet there are some interesting elements. Her range of unnatural hair colors is not the first in the career line, and is something Mattel appears to associate with tech as well as with music, but I find her ripped-knee jeans is notable since no earlier career Barbie wore anything quite so casual, except for the “voter” in the team set. Since this doll was certainly in development in 2020, that likely reflects the advance of casual-is-okay -for-work ideas in fashion in the age of work-from-home. That 2021’s doll is not a team does make sense in a world where work from home separated us so much, but it will be interesting to see if the solo Barbie continues to be a pattern. It is neat, though, seeing them once again showcase a job which is part of the fact that media is teamwork, i.e. the producer not the rockstar, similar to when Barbie the Film Director in 2015 directed girls’ attention to a different type of power in Hollywood from the many movie star Barbies of earlier Barbie decades. In a sense it’s a job which showcases teamwork even while alone, and thus very apt for 2020/2021, and perhaps a good sign for Mattel continuing to think about teamwork and plural agency even in their solo dolls. And the fact that it got much less media attention than judge or political team may mean that the forces of capitalism step in to encourage Mattel to try something bolder next year–we must never forget the $$ side of commercial political messaging.
So, what does the Career of the Year sequence show us about Barbie as a mark of centrist feminism? A few things. One is that women-in-tech is definitely a thing, far more in the minds of the organizers than women-in-STEM, since we haven’t seen biologist Barbie or epidemiologist Barbie showcased, only several iterations of tech Barbies, including software and hardware. It also shows through things like entrepreneur Barbie and architect Barbie that sometimes they look a lot at research, especially about income and what are high-paying careers, and think it’s important that Barbie encourage girls to go into high-paid professions not just exciting ones (beloved-yet-underpaid careers like teacher and nurse have been frequent Barbie careers but not showcase ones). They also sometimes run into challenges in communication, i.e. ‘entrepeneur’ is a very important concept but very difficult to communicate in a doll via clothing and accessories, as is true of many careers.
Several of the career Barbies–notably game designer and music producer–have been major steps in more casual clothing, which is a not insignificant message when we think of the target market largely including middle-class suburban mothers (parents buy the toys more often than kids, after all) who are thus expected to consider ripped knees and wild hair a respectable image for girls to aspire to. The increase in tightly-fitted-yet-somewhat-ungendered clothing, which reached its peak with the carefully-planned game designer doll, is also notable. Recalling how much fashion pressures linger in business, how many employers still expect makeup and highly feminine dress for all women, the dolls’ statement that sneakers, jeans, and a shirt and jacket whose only feminine coding lies in the tightness of their fitting and the small amount of pink on the shirt is a genuinely significant change. That 2021’s is a step more feminine in coding even as it is a step more casual is interesting when put in dialog with gender and transgender issues becoming such a hot topic in the past few years.
And, of course, we saw with judge Barbie and the political candidate team Barbie that a lot of people who consider themselves politically fairly neutral/centrist, including Mattel, felt that the wake of Trump’s election and the midst of the authoritarian surge of 2016-2020 was an important moment to step forward, become more active, and, for the first time in Barbie’s history, to take a semi-overt political stance, since celebrating judge Barbie in the midst of so much focus on Ruth Bader Ginsburg, is not explicitly pro-Democratic-party but it’s extremely clear the way it leaned. Many organizations that strive for party neutrality, from Mattel to the ACLU to the science journal Nature, felt that Trump’s second run was the moment to use that history of neutrality for an important end, since breaking a multi-decade string of never endorsing one party over the other makes the moment when one does speak out that much more powerful. That 2021’s doll is far less political, except for being pro women-in-tech, raises the question whether we should view the renewed projection of party neutrality as a happy return to normal, or as a scary sign that the wave of sudden political engagement sparked in 2016-20 is fading again, and that voter turnout may wane with it.
In sum, since news and social media both tend to magnify radical voices on both sides, things like Mattel’s carefully-calculated political stances can be a valuable window on the often-quieter middle, though whether it really is the middle or just attempting to claim “this should be the middle!” as the real middle moves left and right is another question. And the fact that the fashion-focused “Fashionistas” line and new sets like the glamorous bond-movie style “Spy Squad” Barbie set persist alongside our career Barbies also shows that the extremely gendered Hollywood femme fantasy side of Barbie is still just as strong in the moderate center of this particular feminine ideal as all the politically-progressive versions are (if not stronger since the fashion focused Barbie lines are usually much larger than the career sets). Of the nine dolls in the 2016 splash add below, one-third are narrative-free fashion-consumers, one-third Hollywood fanatsy babes, and one-third career role models, a telling microcosm of the imagery proportions kids are pelted with. Ongoing food for thought.
Partnering with my good friend and fellow author & history lover Jo Walton (more on her below), we interview fellow writers, historians, researchers, editors, and other friends, talking about the craft of writing, history, food, gelato, and other nifty topics, with some episodes of just me and Jo having the kinds of intense writing or history discussions we enjoy. You can listen for free on Libsyn, on Apple Podcasts, on Spotify, and on YouTube. Those whosupport me on Patreon get new episodes early (and new ExUrbe posts early too.)
Sample Episode: Speculative Resistance with Malka Older
The episodes in this first season are modeled on the kinds of panel discussions one has at science fiction conventions, and are long (an hour plus), and since our interviewees are all so interesting! Episodes of this season will come out monthly, with occasional bonus episodes, those are the ones with just me and Jo.
For those who aren’t familiar, Jo Walton is a voracious reader in a huge number of genres with an encyclopedic knowledge of the history of genre literature, as well as the Hugo and Nebula award-winning author of more than a dozen novels including Among Others, and an F&SF critic, author of What Makes This Book So Great and An Informal History of the Hugos. Jo and I travel a lot together when I go to Europe for research, and we’ve had such wonderful conversations over the years connecting dots between our shared interests in history and the writer’s craft that we wanted to share such discussions for more people to enjoy.
Interviewees in the first season (to give a sense of the range) include Malka Older, political scientist and author of Infomocracy, Jonathan Sneed, a Mars astrogeologist & astrobiologist, Ruthanna Emrys, a city/state planning & politics expert and author of the Innsmouth Legacy series, Mary Anne Mohanraj a wonderful writer friend and creator of Sri Lankan cookbooks, Max Gladstone, author of The Craft Sequence and a favorite friend to discuss the craft of writing with, David M. Perry, journalist, activist, and Medieval historian, Emily Cambias, game writer & editor/writer for Cricket, the children’s magazine company, and another writer friend Naomi Kritzer, author of Cat Fishing on CatNet.
Second, I’m Teaching an Open-to-All Online History Course This Fall!
I’ve long wanted to find a way to open up my teaching beyond the university, so through U Chicago’s Graham School continuing education program, and taking advantage of the Zoom skills we’ve all developed this year, I’m teaching an online course this fall on Saturdays, 10 AM to 12:30 PM Central Time, called FFAC10100 Monks to Voltaire: European Intellectual Transformations 1200-1750. It’s a version of a course I’ve taught for undergrads which starts with late Medieval thought and looks at four successive major revolutions in European ideas, scholasticism, then Renaissance “humanism,” then the 17th century’s “new philosophy” or “scientific revolution”, then the Enlightenment, presenting them in continuity and showing how they didn’t replace each other (as summaries often make it seem), but rather joined each other, continuing to thrive side-by-side. I’m aiming at a variant on a “flipped” model of a course, in which I will share the lectures as text transcripts people can read, and then the class sessions can be entirely Q&A digging in more intensively. If you’re interested, anyone can register for it, and you can learn more at the discussion I’m going to have about it with the Graham School staff on August 24th, which you can register for here: Conversations @Graham, August 24 | UChicago Graham
Third, My Introductions to Gene Wolfe’s Book of the New Sun
Tor invited me to write introductions for the new Tor Essentials editions of Gene Wolfe’s four book Book of the New Sun, collected into two volumes, Shadow & Clawand Sword & Citadel. It’s hard to express how formative these books were for me, staggeringly brilliant and ambitious SF which showed me how high I could aim, how deep world building can reach, and how complex a narrator can be. I haven’t felt nearly so nervous and impostor syndromy about a project in a long time as sitting down to write about these books, so seminal both for the history of science fiction and for me, but I’m really happy with the resulting essays, so if you’d like to read or reread (these are books designed for rereading!) some incredible SF with a little bit of my guidance, I can’t recommend them enough, especially to anyone who enjoys Terra Ignota.
Speaking of which…
Fourth, a new Terra Ignota audiobook series is coming from Graphic Audio
I’m extremely excited for this project, now up for pre-order. I’m planning to do a blog post about them soon, but while the Recorded Books audiobooks have a single actor, these are a cast recording, with many different performers playing the different roles, and it’s amazing how different that is in terms of things it can achieve. At my suggestion we’re trying a somewhat radical experiment, so the recording begins a note from Gordian saying the performances have been made in line with Gordian’s recommended genderblind casting practices, and then the casting of the parts is largely unrelated to the gender of the performers, so voices of all kinds are playing characters of all kinds, letting performers who never usually get to do a booming-voiced old man or a delicate child exercise those parts of their ranges, and adding an amazing additional layer to the book’s complexly-worked gender confusion, layering on top of how Mycroft’s use of pronouns often doesn’t match physical descriptions of bodies, and now it won’t match voices either, further encouraging the listener to question all Mycroft’s gendered language and to examine even more how perceived gender affects the way we judge or react to different characters. I’m also especially excited that, against this backdrop of intended ambiguity, the amazing casting director Alejandro Ruiz met my requests to be careful about representation, and found brilliant trans woman Kay Eluvian to play Carlyle Foster, and a nonbinary performer, Taylor Coan, to do Sniper.
Alejandro and I are also both excited about how diverse the cast is in terms of race and nationality, even with a performer from Mumbai to play Bryar Kosala, and we’re doing some double-casting, giving multiple roles to the same performers to encourage the listener to think about and compare them (Ganymede & Danae for example), creating intertextual links between different characters, modeled on the way the inestimable Jane Howell did it in her direction of her Henry VI sequence for the BBC Shakespeare project, my very favorite work on film. These recordings will be slightly abridged, as Graphic Audio usually does, adding some music and special effects and cutting things like “he said” “she said” or some of the descriptions designed to remind readers of who characters are or where they’re from since hearing a Mumbai accent will by itself achieve the same information reminder. It’s been an absolute thrill working on the productions, and I couldn’t be more excited for the new layers they’re adding to what the books are already aiming at in creating a truly global-feeling cast of characters, and stimulating questioning and introspection about gender.
Fifth and last, the publication of Perhaps the Stars is finally close!
The fourth and final volume of Terra Ignota comes out October 19th, and it’s really for sure this time, it has a cover, and the final most finalest final page proofs are done, and all the Latin and Greek and other special characters are taken care of, everything! It’s up for pre-order on Bookshop.org and Amazon and Barnes & Noble and at all sorts of local indie bookstores (please support them if you can!). It may not feel like news that a book which has been planned for months to come out in October is actually coming out in October, but it’s hard to articulate how many invisible steps there are on the back end, including a somewhat-COVID-related continent-wide shortage of printing press time which is making book printers everywhere struggle for time spots to actually get the physical book made at the factory, pushing a lot of things back to 2022…. but not this thing! I’ll definitely be blogging more about book 4 in the coming months, but short version, there are only 2 chapters in the whole of book 4 which, from a craftsmanship point of view, weren’t harder than the hardest chapter in any of the earlier three books, and I can’t wait to share it with everyone!
Hello, readers! The past few weeks have been very intense for me with the 2021 run of my Papal Election Simulation, but I wanted to post some links and announcements about a couple of free online talks, two recent and two upcoming.
One is today (May 13th) at 5:30 PM Central time on “The Apocalyptic Renaissance” for the U Chicago Smart Museum of Art’s fabulous new exhibit “Lust, Love, and Loss in Renaissance Europe.” I’ll be presenting some material from my book in progress “Why We Keep Telling the Myth of the Renaissance.” Sign up for free at this link.
The second is a talk on Saturday at 1 PM Central time for the Chicago Women’s Alliance (also free and open to anyone) on“Who Has the Power to Change History?” in which I’m going to talk about my teaching and how I use historical reenactments and role-playing to teach better ways of thinking about power, what really controls change (individuals, great forces, both?), and thus encourage feelings of empowerment and activism. Several former students will be joining me to talk about how the role-playing elements of my teaching changed the way they think about history and power and how they apply that in their activities. I’m really excited to discuss the question with my students there too. Sign up for free at this link.
And two other recent things I did which are now online:
I did a fun interview for History Hacks podcast, about The Inquisition(s) and the history of censorship (drawing on my research).
And I did a video lecture for the Paideia Institute about Recovering a Lost Classic in the Renaissance, with webcam footage of some real 16th century books and manuscript samples from my book history teaching collection. So excited to have a webcam I can do that with; hoping to do more like that this summer!
NOTE: An unfinished draft of this post was accidentally published for a little while on March 2nd-3rd, but it wasn’t actually ready yet then, but here’s the finished version:
Hello, wonderful readers! What I have to share today is not a polished essay, but the transcript, slightly cleaned up but mostly as given, of a talk I gave recently at a science fiction convention, Capricon 2021, whose theme this year was “Making the Future We Want”—a great topic for reflection. In the talk I look at our ideas about who has power to shape the future, stringing together short precis of several different articles and such that I’ve been working on lately. Each little precis is less polished and evidence-packed than the long versions (links & citations provided where I can), but I think the combination, though compressed, has a useful flow and brings together some points that I hope will help people reflect on how our narratives about history shape the power (and powerlessness) we believe we have. I hope you enjoy!
As another treat, here is a wonderful video made by my friends at the Paideia Institute which recently invited me to give a talk on the process of recovering a classical text in the Renaissance, with live examples of me showing Renaissance era printed books and manuscripts thanks to the miracle of webcam. I’m hopeful I’ll be able to use the same webcam system to do more rare books demos in future!
I was struck by how Capricon 2021 (Capricon is a fabulous F&SF con! you should all go to it!) had a theme this year—“Making the future we want”—which overlaps some of the history work that I’ve been doing, so I thought people might enjoy a bit of a serious talk on a very interesting question. Some of you may have read my blog post “On Progress and Historical Change” which gives a history of the concept of progress, and this talk will overlap that a bit. But what I want to talk about here is the question of how we imagine how society and history change, who we imagine has control over that, how much control we imagine we have, and how that has changed over time. And our feelings about how much power we have, or how much power we feel other people have over change in real history, is often very different from what the historical record suggests. So I’ll be talking about some of my work as a historian and what it shows about how culture changes, vs the concepts that we usually tend to have about that.
I’m what’s called an intellectual historian, which means I focus on what we think is true. This is related to history of ideas, so I study concepts like the concept of progress, the concept of atoms, the concept of rights or equality, the concept of atheism, not just what atheist ideas existed but also what people at different points in history who didn’t consider themselves atheists thought an atheist should or would be like in terms of ethics, the personality the imaginary atheist whom Thomas Aquinas is arguing with in some of his writings, for example—these are examples of things an intellectual historians studies. But intellectual historians also look a lot at worldview. If a material culture historian is working on reconstructing the clothing of another time period, and a food historian is working on the diet and recipes of another time period, and an art historian on the architecture, and a historian of science on the technology or weapons of another time period, the intellectual historian is trying to get at the mindset of that time period. What world do those people live in, from their own perception, from what they believe is true? What is the potential of their world, what do they believe is true about how it works or how it changes, how does it differ from the world we believe we live in? Someone from a culture which believes that disease is caused by astrological influences instead of by germs makes decisions about medicine and health as if living in a different world from the one we believe we live in.
So, for example, we in the present have a very particular expectation that every generation’s experience will in most fundamental ways be different from the experience of the generation before, that there is a constant process of change, progress is one of several names for this, or a name for one element of this process, but we expect, for example, that in two generations while some things will be similar, many things will be very different. For example, very few of us expect that our grandchildren or great-grandchildren will still live in the same house we live in, and use the same teapot we’re using. In contrast with, let’s say, a medieval European figure, who is very likely to have the expectation that their descendants will live in the same house for a number of generations, and who generally has the expectation that change may come if there’s a war, change may come if there’s a great king, change may come if there’s a bad king, change may come if God curses the land, but the change isn’t inevitable, the way we think of it being inevitable in our own period.
So, in different moments in history people have had different ideas about how constant change is, what causes change in the human condition over time, what aspects of the human condition change over time, and which people, if any, have power over the way the human condition changes over time.
I’m going to discuss briefly at first the origin of the concept of anthropogenic progress, and I’ll come back to that term in a moment. Then I’m going to zoom in to a very very microcosm example within my own field, in my own period of specialization, which is Renaissance Italy, which shows some of the problems generated by the disconnect between the way we imagine the world changing from the way it really does, and then from that microcosm zoom out again to the larger question.
(If you’ve read the longer version of this in On Progress and Historical Change pretty recently, you may want to skip the next couple paragraphs down from here to the italicized note, by the picture of a pretentious Roman orator on the rostra, or you may prefer to keep reading so the content is fresh)
So, as I often say when I’m beginning this discussion, in about 1620 Francis Bacon invented progress. What I mean by that is that Francis Bacon, a British intellectual and statesman, published the Novum Organum, and some other works in which he argues, pretty much for the first time that not only is it possible for human beings to change the human condition but that if they do so intentionally and systematically through science—this is the birth point of the modern scientific method, which is to say collaboration among groups of people sharing knowledge to work as a team to gradually expand human knowledge—and remember, Bacon is the origin of the saying “knowledge is power,” by which he means power over Nature, not individual power, he means human beings collectively having more power when we understand more of what he calls “the secret motions of things” or what we might call how diseases work, how physics works, how electricity works, etc. By “knowledge is power” he means our power to command electricity, our power to cure diseases, these things that in 1620 he hoped science might someday achieve. And Bacon (and I’ll get to why in a moment, and also what similar concepts existed before in a moment) articulates for the first time the suggestion that if human beings collaborate as a team to observe nature, to do scientific experiments, to double check (we would say peer review) each other’s experiments, to publish this knowledge and share it, and to collectively try to expand humanity’s scientific understanding, then every generation thereafter will be a little bit more powerful in terms of how many diseases we can cure, how well we can preserve food, how well we can grow food, and thus that every generation’s experience will be a little better than the generation before.
Bacon characterized this in Christian terms as an act of charity, that to be a scientist is the ultimate act of charity because it gives the gift of a happier life to every generation and every human that will be born after you—a very interesting root for the scientific method, which many people are used to seeing presented as an enemy of religion as opposed to as mandated by Christianity which is the way he presents it. And he suggests that by conducting research as a team and sharing it, it can intentionally come to pass that every generation’s experience is different.
Now, we would say that every generation’s experience has been different since the very beginning of humanity, that progress, while it might accelerate later on, as technological advances accelerate has been constant, and the experience of somebody in 1500 is different from somebody in 1400. And indeed, we would argue that human action, discoveries, development of states, gradual processes such as centralization of government or of finding agriculture were always causing progress. People were inventing the moldboard plough, preservation techniques, gradually breeding corn to be a slightly better and slightly better crop until it became the strange super crop that it already was before modern scientific genetic meddling, that progress was always there, than anthropogenic progress, meaning human caused progress, progress that is the result of human beings taking action.
But nobody had described this as a phenomenon before Bacon, and Bacon and his peers in fact believe that there isn’t anthropogenic progress. What kinds of progress or change do they think there is? One dominant idea in Europe at the time, and indeed for many centuries before, is that the primary changes you see experienced by humans are a plan scripted out by God, are Providence. Sometimes God, or Fate, decides that it is time for an Empire to rise, because God has a plan for it. Now it’s time for the Empire to fall, because God has a different plan for it. God decides to send a great king onto the stage of life, or God decides to send a bad king to teach people the moral lessons that tyrants teach. That is the dominant model of what people think causes change in this period, that it is an external decision made by a divinity or plan whose intentions are largely didactic, largely educational, that the purpose of sending a good king or a bad king is to send people moral examples for people’s personal moral education, to increase the likelihood of human souls doing to heaven.
If we move earlier than that, to Antiquity, there are ideas of what I would not quite call progress or anthropogenic progress, but there are ideas of development that the Epicureans discuss for example that just as an organism might develop from a juvenile state to a mature state, so similarly planets develop over time and human civilizations develop over time. In Lucretius’s De Rerum Natura, our longest surviving classical Epicurean text, he describes the idea that Earth undergoes a process similar to the lifespan of an organism, that, for example, only in the earliest days, when Earth was young and fertile, did new species come into existence, and that only those species that were suited to their environment survived until the present day. It sounds very close to Darwin. Lucretius goes on to say that no new species are created any more because Earth is old and has undergone menopause and no longer has giant placentas growing out of the ground everywhere which is where animals came from in the first place. So before you give him too much credit for the survival of the correctly adapted idea, which he does have, and is a very sophisticated idea, there are plenty of others that we chuckle at. But he describes the idea that early on humans lived peacefully in nature, and they ate the fruits that they found and they slept under trees. You can recognize this as one of the roots of pastoral man or of a golden age. But then humans gradually discovered luxuries, gold, and treasures. And when they did that, and there were limited supplies of luxuries, then people needed to defend them, and they developed weapons, and they developed armies and laws and social structures. Then, from that, developed war, and the descent from a golden age to a silver age to an iron age. Lucretius is not the only classical Mediterranean source of this, but he is an example. So there is an idea of development, and even development where people are causing it, but it isn’t an idea of constant development, and it is an idea of negative development and not positive.
That idea exists in Antiquity, is much less prevalent in the Middle Ages, when many of those texts are not available, is rediscovered in the early 1400s, and disseminates again. So it’s one of several influences on Bacon. But what’s innovative in Bacon is his idea that there’s a strong intentionality, that the team of humans who are scientists work together intentionally to create the next generation’s experience, and indeed that if they don’t, the next generation’s experience will not be different from this generation’s experience, that unless progress is intentionally caused, progress will not occur. This is very different from our sense in which progress is constant and inescapable and positive, but also negative. We’re very used to thinking about progress giving us better technologies and better medicine, but also giving us the negative sides. As Freud puts it in his “Civilization and Its Discontents” writing just after World War 1, “it is indeed miraculous that my daughter who is across the ocean on another continent and I can speak over these new electric telephone wires, that is an amazing fruit of science but,” he says “If we didn’t have science she wouldn’t be on the other continent, she would still be here, because we wouldn’t have made the ship that carried her.” Civilization and its development creates problems in addition to creating positive innovations.
That is an idea that we’re very familiar with in the Twenty-First and Twentieth Centuries and that really develops largely in the Nineteenth Century as a result of the Eighteenth Century’s romantic interest in the idea of the pastoral and the influence of Jean-Jacques Rousseau, but also in the aftermath of the French Revolution. In 1620 Francis Bacon said “If we do science we can have progress,” and people said “Great! We can use science to evaluate everything and make more rational medicine, and more rational engineering, and more rational architecture, and more rational farming, and more rational laws, and more rational religion, and then they had an enormous war in which huge numbers of people died, and the scale of war got an order of magnitude bigger than anyone was used to, and it created a lot of anxiety and fear that possibly we don’t have as much control over this progress thing as we thought. So it’s then in the Nineteenth Century that you get the image of being a cog in the machine being trapped within progress, progress being inescapable.
Within this review of concepts of progress the question then comes: who has power over determining what happens? Is the answer “No one”? That progress happens in so many different arenas and so many different directions that pretty much it’s just chaos? Is the answer “Governments or rulers”? that a great king or a great virtuous leader is going to be what determines whether our country prospers. Or do we believe that certain individuals exercise influence?
And here is where I’m going to zoom in to the Renaissance, which is to say, to my own period, which is where we see ideas of who has power over progress, and especially who has power over mindsets and the clash between the assumptions we often make about that vs the historical evidence that we have.
So, Italian Renaissance, or Renaissance in general, we’re used to thinking of this as a moment of great change when things accelerate, when the Middle Ages, which we tend to think of as stagnant, and perhaps backwards and without dynamism and without a lot of progress or change—all of these things are false, but they are the assumption most people have—suddenly shifts and we get a lot more. We get faster innovations, and faster development of technologies, and a change in mindset. Now all of these things are true, we do get an acceleration, but it’s a wedge, it’s not a flat line and then a spike.
The core figures that we know from the Renaissance, and the way you often know about the Renaissance when you’re not a specialist, are famous geniuses. Leonardo da Vinci is the best known Renaissance figure, followed by Machiavelli and Michelangelo and Raphael, and we think of this as an age of genius, an age when excellent brilliant people created magnificent art and magnificent architecture, and there is a focus on those geniuses as being the core of change, on those geniuses as being fundamentally what makes the progress happen. There are in fact a lot of cultural reasons, and this is not the moment for me to give you my list of what the causes of the Renaissance are, though there is a blog post about that if you’d like to look for it in my discussion of COVID-19 and why we have the Medieval/Renaissance distinction and the problems with it, but the key data point that I want to start with here is when I’m at a dinner party and I meet somebody’s spouse and they ask me what I study and I say I study the Renaissance, and they ask what I study in the Renaissance, and say well, I study Renaissance radical thought, I study censorship and the Inquisition, and I study atheism, and I study heretics and heresy trials and the dissemination of radical thought. It’s very common for there to be a pause, and an excited face, and then the question “Oh! So isn’t it true that x famous Renaissance genius was an atheist?” It’s incredible how often that’s the question. And the reason for that is a narrative that many of us are familiar with that the Renaissance is an era of secularization or secular thought, that it’s the moment when the shackles of religiosity are being challenged, when Humanism as a centering on humans comes in (that’s not actually what Humanism was in the Renaissance, but that’s what it is now and that’s what people think it was) and that the core of the difference between the Middle Ages and the Renaissance was the breaking free of faith and the beginning of an age of Reason.
That is an expectation largely based on ideas about the Renaissance that were published largely by German and British historians in the Nineteenth Century. This is very, very different from the way the Renaissance presents itself, and also very different from what we find when we look at things. But the important detail here is everyone’s excitement to have me confirm at the dinner party that yes, Leonardo was an atheist, or yes Machiavelli was an atheist, or yes, Michelangelo was an atheist. Why is that? Mainly it’s because the dominant narrative about the Renaissance is that there were a bunch of geniuses who had a pseudo-modern mindset. Meaning more secular, not necessarily atheistic, some people are simply excited for it to be modern, scientific, etc., and that those geniuses saw lying before them the possibility of a more secular more rational age that would be better, it would be modern, it would be free of all the limitations that dragged the Renaissance down. And that these geniuses seeing that in their future, seeing it, then worked to make it so. I see this narrative in a lot of places. I see it in academic articles or books that assume that it’s true. But if I zoom out further, you’ll see it in op eds, when people are writing an op ed about the Black Death and COVID and they’ve researched this for 48 hours, that’ll be what they say, or it’ll be in the intro of an economics book.
If you zoom out even farther, you get it in the Cosmos TV series, the new one with Neil de Grasse Tyson which chose to begin its first episode with this glimpse of the life of Giordano Bruno (analyzed in the article above), it presents Bruno as a kind of martyr for science, and whom it describes as a very modern person, that he saw that there could be this other world of reason and he tried to buck against the system and fight against it, and the Inquisitional Thought Police, and he uses the phrase “Thought Police” which is a very important detail, chased him down and suppressed him because they didn’t want these ideas to be spread. But he knew that one day it would, and other people who were like-minded with him worked really hard, and their underground efforts eventually made it so.
And if you zoom out even further again, you get it in pop culture stuff. The plot of the extremely mediocre old David Warner fantasy movie Quest of the Delta Knights. The plot of that is that there has been an ancient scientific and secularizing rationalist secret society founded by Archimedes that has since Antiquity been keeping the light of reason secretly alive through the Middle Ages, and eventually will break through and reveal all the lost documents and science from Atlantis and usher in the modern age, and the hero helps make that happen, to defeat the Middle Ages and make the modern age happen because this cabal of people who fundamentally think like modern people has existed since Antiquity and preserved the light and now it’s going to triumph. It’s also the plot of the Assassin’s Creed video games, that there is a secret society of rationalist people who fight the pope—literally, by punching him!—and they advance science, and Machiavelli is in charge of it.
And these are hilarious and fun but they are echoes of the idea that the people who are in charge of intellectual change, the people responsible for getting us from Medieval to Enlightenment thought like us, that they thought ahead of their time, that they saw the future potential, that they had a plan and they secretly implemented it—in other words that the people who have power to determine what the future is, is a few particular geniuses, who in their brilliance see what the future is going to be. That is not true. First of all, none of these people predicted any of what the future is like, and none of them in fact tried to undermine the church—I’m talking about Leonardo, Machiavelli, etc.—and none of them articulated things that would read like modern when you really sit down and read them. They are presented as proto-moderns by historians of the Nineteenth Century who want to claim a descent from them, because we respect these names, they are exciting, and so if you can claim “I am carrying on in the tradition of X!” you can make your own regime seem powerful. This is a tool that the rising nationalist movements of the Nineteenth Century, in Germany, in Britain, in France, in Italy, used to compete with each other by trying to claim “Oxford is a truer descendant of the ideas of XYZ than the universities in Italy, look at how we’re carrying on the rationalist secret underground messages that were in Machiavelli and were in Leonardo” which are not there.
There are radical messages in these things. Here is the key. Renaissance radicals were really radical! Machiavelli was really radical, Leonardo was really radical. People like Giovanni Pico della Mirandola, who is a little less famous but very frequently pointed at as the Renaissance genius, mastered fourteen languages by the time he was a teenager, he was brilliant, he was rich, everyone loved him, he wrote a giant nine hundred thesis synthesis of all world religions, which he proposed to defend in front of the pope. And he has this famous text known by its Nineteenth Century title The Oration on the Dignity of Man. I’ve got a 682 page book on my shelf by Brian Copenhaver about how The Oration on the Dignity of Man is not an oration and it’s not about the dignity of Man—that is a Nineteenth Century reading of it which still colors forward. It was actually a manual on how to turn yourself into an angel, by hybridizing Islam and Zoroastrianism with Kabala, channeling those through Plato and making a new version of Christianity which would synthesize all world religions. And by the way, Pico and his friends believed that if they just explained this to the Ottoman Muslims they would immediately understand that Islam and Christianity were exactly the same and then there would be world peace. That was the plan, that everyone would real Plato and then they’d realize that all religions were the same and then there would be world peace. These guys are really radical! Their versions of Christianity have reincarnation and soul projection and did you know that if you study enough of the Chaldean Oracles you can project your soul out of your body and use it to spy on Padua? This is what Latin class and Greek class should be for. Renaissance radicals were incredibly radical. But their radicalism did not resemble our modern mindsets. Their radicalisms were all over the place. The more I study them the more I love them, there are dozens of worldviews in there more alien than any alien in Star Trek ever made up, they’re gorgeous, but they don’t resemble the way we thought in the Twentieth Century, or in the Nineteenth Century, or now. Those people in the Renaissance didn’t see a future and then intentionally make it. Those people in the Renaissance had radical ideas, proposed them, debated them, mixed stuff all over, had different weird inconsistent influences, and the cauldron of all of it, of the people you’ve heard of and the many dozens of people you haven’t and the many hundreds of people who read them and debated them and the many thousands of people who read those books in schoolrooms and disseminate those weird ideas, those are what actually makes the change happen.
Now I’m going to zoom in to my most micro example. So. The least interesting set of texts that I can think of, and I’ve been trying to look into this, were editors’ prefaces at the beginnings of copies of Epictetus (aka. my article “Humanist Lives of Classical Philosophers and the Idea of Renaissance Secularization” in Renaissance Quarterly Vol. 70, No. 3 (2017), pp. 935-976). Epictetus’s manual is a set of Ancient Roman moral maxims about how to be a virtuous person, and it’s short, so people liked to teach it at school, and a lot of it sort of lines up with Christian virtues, so it’s very compatible with teaching in a Christian context, so it was a super popular textbook. It was a super popular schoolbook, the equivalent of making kids read Dickens or making kids read a Shakespeare play as part of High School English class. So there are many, many dozens of editions printed pretty much as soon as printing is invented—from the 1490s, and all the way through the 1500s, and the 1600s and the 1700s there are editions of Epictetus, and every one has a preface from the editor explaining why Epictetus is a great text to read, and usually trying to seem more awesome than the previous edition of Epictetus, so that you’ll buy that edition for school instead of the other edition, or so that the teacher will recommend that edition instead of the other edition. Publishing, as we all know, is very competitive, and you’re always trying to have the blurb that makes it attractive — if someone walks into the bookstore and there are four books with dragons on the cover, something is going to make the difference to which one they pick up, it might be the copy on the back, this is the equivalent of that.
So when you look at the prefaces, which are being written by scholars who aren’t very important and aren’t very famous and nobody has heard of or cares about any more, but they have a job and the printing house has got them to edit this thing because this is their job. Those guys who are doing that, that you’ve never heard of, are reading the guys you have heard of. They’re reading Pico, or they’re reading people who are talking about Pico. They’re reading all these dozens of bizarre strange ideas that are going on. Most of them definitely don’t agree with it, or if they agree with it, they only agree with one thing, because there are dozens of different ones, but there’s the milieu. And they notice the arguments that are big and all over the place, and one of them is, Renaissance people were very interested in the fact that the sayings of Epictetus were remarkably similar to some of the ethical teachings of Christianity, especially in the letters of St. Paul. In fact in the Middle Ages there was a rumor that Seneca and St. Paul knew each other and wrote letters and Seneca was associated with Epictetus, and Epictetus was supposed to have secretly converted. None of this is true, and in the Renaissance they had pretty much figured out that it wasn’t true. But it was very interesting to note that Epictetus’s moral maxims were similar to St. Paul’s. So if you were writing a preface to Epictetus, you would write a preface that said “He was a pagan, but he was almost as good as St. Paul!” And then ten years later when there was a new edition and someone wanted to make that edition sound better, and I have all of these editions in Latin and I have the article version of this where you can read the translations of all of these, they’ll say they want to make a stronger claim, so they’ll say “Epictetus’s moral maxims are barely less good than St. Paul’s.” And the next one will say “Epictetus’s moral maxims are just as good as St. Paul’s!” And the next one will say “Epictetus’s moral maxims, even though he was a pagan, were even better than St. Paul’s, because they are simpler and clearer and more effective at teaching ethics.” And by the time you get to the 1700s there’s an edition, which you can tell is copied from these earlier editions, they even plagiarize sentences from each other’s prefaces, it’s direct evolution, that says “Epictetus by the light of Reason alone and without the necessity of scripture or revelation arrived at better ethics than St. Paul.” And that is the kind of book that Voltaire owns when he is a kid.
So, who is transmitting this radical idea? Gradually of the idea that Scripture and Christian revelation are necessary for ethics? Who is transmitting that? Is it the big famous people? No. Because there are thousands upon thousands more copies of these classroom Epictetus volumes than there were of Machiavelli in this period. Machiavelli is banned in most places. You can’t even get it without hunting hard, you can only get digests of it, and it’s not printed nearly as often. People have it but it’s extra work, like how there are people around who use Linux instead of Windows or Mac but it’s extra work and you have to work at it and not a lot of people do it. But Epictetus? He’s as ubiquitous and default-accessable as Windows. So for every one person who’s actually reading Machiavelli in that era, a hundred people at least are reading this Epictetus preface that says you don’t need Christian Scripture to arrive at a good system of ethics. Who spread the radical idea? Thousands of people you haven’t heard of, dozens of editors who wrote these editions, most of them not intending to make anything radical happen but just intending to sell a copy of a book, in fact most of them genuinely believing that Epictetus was an author who would advance Christianity and make people believe in it more, they, nonetheless, unknowingly and unintentionally transmitted the radical ideas that turned into Deism, and that turned into that secular turn that we associate with the Enlightenment, and that we falsely associate with the Renaissance. So you see it is the Renaissance seeds that lead to it, but it some genius list of special radical people who thought the way we do that made it happen, it’s thousands of people who had dozens of different worldviews, some very orthodox, others very radical in ways that don’t resemble us, but all these ideas discussed and wrote and published and debated, and those debates influence textbooks written by nobodies who get left out of the kinds of history that focus on big names, but it’s those small names that have so much deeper broader reach than the treatises of the people who are the most famous today, largely because our canon of who’s famous was cemented by nineteenth-century people who were looking into the past and cherry-picking people to celebrate whose ideas they thought resembled their own, in order to legitimize themselves (and prop up their belief that they had a right and duty to dominate the world and ‘elevate’ ‘lesser’ cultures with their ‘right’ ideas and ‘right’ path of progress).
So, what this example and other similar studies shows us is that we overestimate how much intentionality we think individual special people have. So every time I see a cover of a tech magazine that has a new tech start-up billionaire that says “Will this man be the first man to live for two hundred years?” or “die on Mars?” or whatever, and in the article it claims “This person has a vision for the future and knows what the future will be like, and it’s this, and he’s working on building it!” (or occasionally “she”, but usually “he”) and “the future is in the hands of these geniuses.” That idea, that the future is in the hands of people whose mindsets somehow already match the future, that makes us feel powerless, makes us feel like our job is to sit back and wait for those geniuses who can already see the future to bring it about the way we imagine Machiavelli did. I remember when I first realized this it was watching the first Iron Man movie, where Tony Stark, after inventing pollution-free green energy, and saving stuff with the iron man suit gives this press conference statement “I have successfully privatized world peace,” and that is the line that gets pushed, that the solution to our strife, or climate change, is waiting for the genius to appear like a superhero with the special power to reshape the world, while everyone else sits back and gets rescued. Why do we get drawn into the rhetoric that claims this? Because our textbooks tell us that’s how we got to where we are now, that the present came about because historical figures who were geniuses and could foresee the wonderful, liberated, rational modern era that was coming then sat down and intentionally tried to make that era come. Well, they didn’t. This world is nothing like anything that people in the Renaissance tried to make. They tried to change what their world was like, but the things they were aiming for were not what actually resulted. It was the mixing together of all sorts of different and incompatible ideas, radical and orthodox, mixing in the contributions and debates of enormous numbers of forgotten, un-famous people, that was key to what changed mindsets overtime. Sometimes particular key works were extra prominent in that discourse, be it Hobbes or Epictetus, but that doesn’t mean that the effect of Hobbes publishing Leviathan was to engineer a future Hobbes envisioned, it was to stimulate thousands of people to have new conversations, many of them hostile, which in turn generated new ideas.
So, who has the power to shape the future? The answer the Renaissance shows us is not genius, not “a few people who have the right vision” it’s lots and lots of people, but a lot of the time the result of their efforts isn’t going to be what they meant. So for example, the first Renaissance scholars who tried to get people to read Epictetus, Petrarch and such, their idea was that fractious and war-torn Renaissance Italy—which Shakespeare depicts somewhat accurately in Romeo and Juliet, where rival noble houses are having feuds and goons and killing each other in the streets, and people are very willing to cause a civil war to advance their family) Petrarch and other Renaissance intellectuals thought that this was a problem and that the best way to address it was to improve the moral education of these ruling families. The reason that they wanted to build libraries and advance the new reading of the classics and so on was that they thought if their leaders had better ethics they would stop having feuds, civil wars, that if they could get the leaders of their society to read the books that young Cicero, or young Seneca, read when they were growing up, that you could then rear a new generation of leaders who would put the good of the country before the good of their family, who would like the Roman Brutus (not the one who killed Caesar, the earlier one) be willing to execute his own sons if they were committing treason, in contrast to the Renaissance norm which was to get your goons to assault the jail and liberate your son if he’s been arrested. These Renaissance people wanted people to read Epictetus, and similar works, to make them more Christian, and more ethical. We have reams and reams of this writing where they say it “We think that reading these Roman texts will make you a better Christian.” They say it over and over and over. And yet the inadvertent effect of that was disseminating a lot of radical ancient thought, which made a whole lot of different radical stuff happen, which caused changes within Christianity, which contributed to the Reformation as well as deism and atheism, which none of these early Renaissance figures would ever have wanted to have had happen, but which was the real effect of their project.
Important sideline because people always ask: so, lots of Renaissance figures who spread radical ideas but also claimed to be good Christians get pointed to by modern people who want to say “They’re secretly an atheist, they’re just self-censoring!” but the thing is that tons of these figures publicly endorse ideas that are way more dangerous, way higher on the Inquisition’s list. You don’t meticulously self-censor that you’re secretly an atheist or secularizer and then publish a demon-summoning manual with your name on it, or a book that endorses Martin Luther or espouses your belief in reincarnation. Someone who recklessly expresses ideas much more dangerous than atheism would not bother to be so careful with their atheism that we can’t find traces even in their private writings. Atheism isn’t as risky in the period as the other stuff which these people are espousing. I’ve got an article in which I go through every single person we know in the Renaissance who read Lucretius and got in trouble with the Inquisition and what they got in trouble for and it was never Lucretius. One of them goes to the stake for sola fide, Luther’s belief that faith alone, rather than faith and good works is how you get to Heaven. No atheist would choose to go to the stake for sola fide, they would say “I don’t care” and walk away. Giordano Bruno, pointed to in the Cosmos TV series which claims that he went to the stake for believing in atoms and the existence of alternate worlds, when you actually read the trial record, he doesn’t espouse that stuff in the trial, nor do they ask him about it. The only time Lucretius comes up in his entire trial is a time in which Bruno brings him up to call him stupid and disagree with him. They’re worried about Bruno’s use of Aristotle, and they’re worried about Bruno’s ideas about the division of the soul and the way the soul connects to the body and stuff that has nothing to do with atheism, atomism, materialism, or anything else that we consider modern. That’s not what gets Bruno in trouble. So we think of it as being atheism, but nobody would bother to self-censor atheism and then avow this other stuff that they all know is more dangerous. Most of them avow incredibly dangerous stuff and get away with it because they’re friends with a Duke, or a king, and this is a period in which you can say almost anything if you have the protection of royalty, and that’s how we get lots and lots and lots of radical works, because people are under their protection. So the claim that there was a radical underground of secret atheists meticulously hiding their traces so carefully that we can barely detect them even in their own works while they were running around declaring their support for much more dangerous ideas, and getting in big trouble sometimes for doing so, just doesn’t make any sense.
Were there some atheists in the period? Totally! See fabulous treatments by Matytsin, Hunter/Wootton, Kors, Sheppard, Popkin, D. C. Allen) and I think I’ve personally confirmed about three definite ones, but the three I’ve looked at didn’t hide it that much, we know about it because all their friends knew and commented on it, it wasn’t that secret. And—the more important part—it wasn’t those three who had monopolized the power and influence and created modernity through their genius vision, they were fairly minor people—one was an illiterate fishmonger, an awesome one, but not a person in power in any traditional. The key is that these awesome atheist radicals were only one kind of radical amid dozens and dozens of other equally awesome Renaissance radicalisms that don’t resemble modernity at all, but it all mixed and churned among the many more orthodox ideas, and percolated into things like Epictetus prefaces, and that’s what caused the change, not the few people who most resemble later things, the mix of everyone.
To briefly quote my polished version of talking about this (from my chapter in the Hardie/Prosperi/Zucca volume) “We do these courageous freethinkers a disservice if we dismiss the diverse and original ideas expressed in their Christian Epicurean and anti-Epicurean works as mere veils over a comfortably proto-modern rationalism. Rather than taking a step forward on a triumphant path leading inevitably toward modernity, Lucretius’s radical Renaissance readers took many steps in many directions, breaking new and fruitful ground… As we seek the agents who forged modernity, we need to stop looking for people who look like heroes, for people who look like villains, and above all for people who look like us. The characteristic ideas and values of modernity were not birthed by people we would have agreed with. They were birthed by a vibrant and diverse range of pre-modern minds alien to our own, advancing plural projects which all moved and shaped each other, plural particles in constant motion all with dynamism and momentum, not passive and inert until struck by a single modernizing genius who contained the swerve.”
So, zooming back to the present, a lot of people have a sense of powerlessness, as if we’re supposed to wait for the geniuses who clearly see the future to make it happen, and we don’t resemble those geniuses because, as history presents them, the genius figures who shaped modernity always had a clear plan, they never have vague self-doubt, or maybe they have like one dramatic turning point doubt crisis and then come out of it as their mature perfect genius selves, they’re perfect, like the protagonists of novels, and they never do laundry, and they never run out of socks, or worry about paying rent, and if the historical record shows them worrying about money then they’re somehow morally compromised and not true intellectuals, which isn’t true! I have letters where Machiavelli writes home complaining that his salary hasn’t arrived and he’s run out of socks, and has holes in his shirt and he’s worried about looking scruffy in front of the pope, and then there’s a letter back “We gave your salary to Michelangelo and we hope he can bring it to you,” and then an angry letter “Michelangelo didn’t bring me my salary!” And “Sorry, Michelangelo had trouble on the road and turned around and had to come back, there were bandits.” Real famous people in history had lots of setbacks and problems and laundry like anyone else. They resemble us more than we think. (This is why I teach Machiavelli’s letters so much). And such people also had less power over change than we ascribe to them.
Real changes in what a society thinks, in what a culture values, come from thousands of people debating something. It comes from that classroom where people are talking about Epictetus. And the modern equivalent of that classroom where people are talking about Epictetus then is this talk, this convention, blogs and social media spaces, even Twitter, anywhere where people are talking about books and events and thoughts. What’s going to shape the future? It’s people online debating about which actions are ethical or unethical in Game of Thrones. That’s exactly like these classroom discussions of Epictetus, which turn into introductions to Epictetus, which turn into the education of Voltaire, which turn into the pen mightier than all swords. Random conversation is where it happens, not one genius, thousands of people exchanging ideas. And it doesn’t result in the world those people envision. Renaissance people did try to intentionally re-engineer the world, and they did sort-of have a shared plan, they wanted to make a world that was more ethical, that had a lot of moral education of its elites in those values which both Ancient Rome and Christianity shared, and this would result in an era of peace.
The actual social engineering project that the Renaissance undertook, the world that they wanted to make, is not the world that resulted. The world that resulted was more different, more strange, and more awesome than anything they ever envisioned. Because you know what came out of the Renaissance? Francis Bacon inventing progress and the scientific method and the beginnings of the Enlightenment, which are fruits far more innovative, far more exciting, and far more powerful than anything any Renaissance genius sat down to do. So when you ask yourself “The work I’m doing to try to make a better world, is it helping? Is it going to make the world I envision?” The answer is: it’s not going to make the world you envision, but it is helping, and it’s going to combine with the efforts of thousands of other people that happen in every conversation, in every convention, every workplace, every school, and media post where you’re debating or disseminating an idea or even sharing a concept, it all contributes. But the world that we end up making is not going to be he we envision, it’s going to be—like Francis Bacon’s world—stranger, more different, and more awesome, than those who created it could imagine, just as the Enlightenment was relative to the Renaissance.
So I hope what you take away from this is some point of encouragement and hope, and the understanding that we will not make exactly the world we imagine, but the world we’re going to make is going to be an amazing world, and that we are all contributing to making it, not just elite geniuses, but every one of us, every day.
Hello! It’s been a while since I posted since, as usual, many projects press, so it’s rare for me to have the time to write the kinds of polished essays I like sharing here. But I’ve been hoping to share more things, since a lot of the history work I’ve been doing lately has helped me with reflecting on current events, and I want to share that faster than the slow grind of book-length work and academic journals will allow. So I’m going to start posting a few things here that are a little rougher. I hope to still post formal essays a few times a year as before, but I’m going to start also sharing things like transcripts of lectures or talks I’ve given, excerpts from my teaching notes, or assemblages from Twitter threads which took meatier turns. I hope you’ll enjoy them, but I’ll also try to always make clear what kind of content each post is, so you know which are the polished essays you’re used to.
I’m also launching a Patreon, so if you’ve enjoyed my posts, books, music etc. please consider supporting me.
I’ve felt torn about Patreon for a long time since, unlike so many wonderful scholars and authors I know, I have a steady living wage from my university and don’t struggle to get by. But, as my Patreon page explains, what I don’t have enough of is the means to hire help. As someone trying to create a lot (and as a chronic pain sufferer who often has fewer than 7 days in my week) it makes an enormous difference to how much I can do if I can pay for help: pay a music editing service to turn polish vocal tracks into completed albums without spending hours on it myself, to pay my part-time assistant Denise who helps with my calendar and paperwork and fire-hose of email which so easily eat up whole days, to hire a sound editor to finally make it possible to launch a podcast with my good friend Jo Walton talking about books, and craft of writing, and history, and science, and Florence, and gelato, and interviewing awesome friends. Even the little post below was made possible by having help, and wouldn’t exist otherwise. And supporters will get updates on what I’ve been up to, and early access to blog posts and podcast episodes, and snippets of outtakes and works in progress. So if you’d like to help me hire the help I need to turn more ideas into reality, and to have more time to write, please have a look at the Patreon page for details, and thank you very much!
Why I Teach Machiavelli Through His Letters
(excerpt from a lecture transcript, so this is how I explain this to students too)
Teaching Machiavelli through his letters is a separate thing from being an historian accessing Machiavelli through his letters. One of the reasons that I love teaching Machiavelli through his letters is that you get a very different view of the person from letters. You get unimportant details. You get the things that the person cared about that week, as opposed to the things that the person wanted to be discussed by many people in the context of that person’s name for a long time. You do get the serious political thought, but you get it mixed with “Where is my salary?” “Hello my friend,” “Here’s the party I was at,” “I have a cold,” all of these very human elements that don’t come to us when we just read a thesis.
Thanks to interdisciplinarity, both at University of Chicago and elsewhere, I move from department to department a lot–I spend some of my time with historians, and some with classicists, political science people, Italian literature or English literature people, and with philosophy people. Each of these disciplines has a different way of approaching text, but many of them approach the text perhaps not with the formal philosophical attitude of “death of the author, we care only about the text,” but all the same with the effective attitude of “we try to learn about this author only through the text,” and only through the formal polished text, the treatise.
When I’m trying to unpack not only Machiavelli but history in general to my students, it’s very easy for the history to seem like a sequence of marble busts on pedestals who handed us great books. It’s much harder to get at the fact that those people are also people who are like us: people who messed up, people who ran out of money, people who had anxieties, people who failed in things that they undertook. People who had friends, people who were nervous without their friends, and lonely. And that isn’t a version of history that we get shown very often. We get shown heroes, we get shown villains, and we get shown geniuses, as if there isn’t a person present as well. Machiavelli is a very valuable example, because we have such a great corpus of letters, but he’s also such a name. If you want to make a shortlist of people who are a marble bust on a pedestal in the way that they’re presented as we talk about the history of thought, Plato, Aristotle, Socrates, Cicero, Machiavelli, are major major figures in that way. So the letters humanize them and make them real.
I feel it’s important not to approach these works as if these people are somehow superhumanly excellent, as if these people are somehow perfect in what they undertake. I’ll often be at a conference where someone will talk about a passage in a work in isolation. I was recently at the Renaissance Society of America conference, and there was an interesting discussion of a passage in which Ficino had a really weird interpretation of this one passage of Lucretius. And there was a very nasty fight between two scholars over the interpretation, in which one of the scholars insisted he’s making this complicated subtle three-part reading of a thing that relates to another thing, diagram diagram diagram. The other person said “I think he translated the passage wrong. Because the passage was really hard. And his copy didn’t have a very clear script. And I think he didn’t read the sentence the way we read the sentence.” And the first person was adamant that it is inappropriate to question whether someone like Ficino might have had trouble reading a piece of Latin, that of course his Latin is immaculately better than our Latin. And his Latin was better than our Latin, because he spent more of his life doing it and I do believe he’s better than most classicists at this — but most classicists really struggle with that line. And when you read the commentaries on it there’s lots of ambiguity even now about what it means, and we have dictionaries, which he did not.
It was very interesting to me to see that battle between thinking of the figure as human, in which the question “Did he mess up?” is a valid question, as opposed to thinking of the person as someone who could never mess up. And a lot of the ways we approach historical figures, whether it’s Machiavelli, or Aristotle, or anyone, involve the idea that all of their works are fully intended, that they’re somehow in an a-temporal vacuum, that we should look at them all in sequence, that no one is ever going to change his mind about a thing unless the person themselves made changing their mind about a thing be a big deal. We create this idea of these geniuses where everything they wrote even from early on is exactly what they meant, which then all gets incorporated into material.
I want my students to come away from my courses not thinking about historical figures like that, but remembering that every historical figure had to pay for socks, or had to deal with laundry, or have a servant who dealt with laundry for them and then they had to deal with the servant. But they all had everyday practical existences, and they all mess up. Machiavelli’s letters give you access to somebody who feels like a real human being. Some of the things he’s doing are really weird. Some of the things he’s doing involve bizarre sexuality. Some of the things he’s doing involve uncomfortable politics. Some of the things he’s doing involve very astute politics. Some of them involve very terrifying moments like his wife saying: “I’m so glad you’re alive, we heard that Cesare Borgia massacred all of his people, I’m so glad you’re alive!” And others are very much “We’re trying to get my brother a job and no one will give him a job because it was corruptly given to the other person and we have to figure out how to get my brother a job,” which is not the sort of thing we imagine such people giving their hours to.
When you read Michelangelo’s autobiography there’s an interesting point in it where he stops talking about art for a while and starts talking about the lawsuit that went on between him and people associated with Giuliano della Rovere because he was contracted to build Giuliano della Rovere’s tomb, but then for a variety of complicated reasons the tomb did not materialise as it was supposed to have, largely because the plan for the tomb was the most insane ridiculous over-the-top impossible tomb that you could ever possibly conceive of. That was obviously never going to happen. But also there were lots of fights between him and della Rovere over who had to pay for the marble and whether the marble was delivered and he said the marble was delivered and Della Rovere said the marble wasn’t delivered and there was a crack in it… and all these lawsuits went back and forth, and also Guiliano della Rovere was starting a giant war and invading Ferrara. At one point Michelangelo ran away from Rome saying “I’m not going to work on this stupid tomb any more” and went to Florence, and then Giuliano della Rovere moved an army over to besiege Florence and started threatening them “Florence! I will besiege you and burn you down unless you give me back Michelangelo!” We have these great documents where Michelangelo is begging Signoria “Please don’t make me go back to Della Rovere! I hate him and he just torments me. I’ll build you really good defensive walls! Look at my engineering ideas for how to improve the walls!” and they had to say “No, I’m sorry Michelangelo, we’re not going to war with the Battle Pope just for you, go back to Rome, build the stupid thing.” And he did go back to Rome, and then Della Rovere made him paint the Sistine Chapel ceiling knowing Michelangelo hated painting, basically as punishment for trying to run away. I’m not exaggerating. And that’s why there are lots of angry figures in the Sistine Chapel ceiling. But the wonderful horrible flirtatious strange antagonism between Michelangelo and Giuliano della Rovere is magnificent.
And in his autobiography he’s talking about this lawsuit that arose because of the della Rovere tomb project, in great detail, and then there’s a line that says Michelangelo realized that, while dealing with a bunch of lawsuits and Pope Adrian and such, he’d been so stressed he hadn’t picked up a chisel in four years. Because he spent the entire time just dealing with the lawsuit. (Anyone feeling guilty about being overwhelmed by stress this year, you’re not alone!) And we have four years worth of lost Michelangelo production, because he didn’t do any art that entire time, because he was just dealing with a stupid lawsuit. And that’s not the sort of thing that fits into our usual way of thinking about these great historical figures. We imagine Michelangelo in his studio with a chisel. We do not imagine him in a room with a bunch of lawyers being curmudgeonly and bickering and trapped in contract hell.
Those sorts of things are important, I think, to reintroduce into the way we imagine historical figures. That they have an everyday mundanity that we imagine that they don’t. And I think that’s a big part of why when we compare ourselves to them we feel as if we can’t live up to that greatness. Because we tell edited versions of the lives of great men and great women, in which we edit out the things that feel like us. So of course we feel as if our everyday lives full of mundanity can’t rise to those levels, because we’re not comparing ourselves to the real people, we’re comparing ourselves to the edited version in which we take out the mundanity! So Machiavelli’s letters give us that. And they give us a person with problems, and a person with mistakes, and a person with a sense of humor, and a person with sexuality, all of these elements we erase from our marble busts on pedestals. And so that’s a big part of why I use the letters while teaching, and when my students read them I want them to put together “Here is a real person who is like us,” as well as “Here is the everyday on the ground experience of what it’s like to live in this crisis.”
We need that, when we live in a real crisis ourselves, and it makes us feel so often like we’re powerless and weak compared to these impressive people in the past–but they felt that way too.
Short version: I’m posting today to share two files I made for my university. One is a Healthy Work Habits and Self-Care Guide for the Pandemic Crisis, which many people have said they found helpful and refreshingly different from others that are going around (details below). Some bits are academia specific but most of it is applicable broadly, and I’ve left in the links to university services because it’s likely you can find such services . The other is teaching-specific, a guide to Adapting a Syllabus for the Crisis, not focused on remote teaching, but on the fact that everyone on Earth is already on the verge of breaking down, so it’s vital to do all we can to structure courses and assignments to have a little more leeway, and paths to recover when students (and instructors) do break down.
I want these to help as many people as possible, so please download them, share them, excerpt them, adapt them, make a version for your school or business, use them any way you wish; while credit would be nice my only firm request is that, as you pass them on, you also pass on the wish to pass them on. (If you want .docx files they’re here for self-care and here for syllabus design.)
Early this spring, as the COVID-19 crisis set in, someone on #DisabilityTwitter asked (I wish I could find the tweet) if others too had found that the self-care skills needed for chronic pain are the same as the ones needed to cope with the pandemic.
I was among many who answered “Yes!”, and soon a small thread was describing our experiences, that it felt almost like a superpower, already understanding the slow, invisible toll of constant daily trauma, the exhaustion that sets in, how to self-monitor, how to spot when you can’t do it and should switch to a different task (rest is a task), and how to fight back against that self-accusing voice inside that insists you should keep pushing through, and is plain wrong. Worldwide health organizations have recognized this as a World Mental Health Epidemic, as well as a viral epidemic. Chronic exposure to fear and anxiety (which are forms of pain), have a real, measurable traumatic effect on the brain, dealing neurological damage which is worse because it repeats every day, and which is affecting every human on Earth right now. Symptoms include fatigue, difficulty concentrating, difficulty sleeping, short temper, diminished higher cognitive functions like writing, reading, creativity, second languages and other skills seeming to vanish away. Does that sound familiar? It’s very familiar to me, the feeling of waking up one morning to find I just can’t do it, words aren’t flowing, my eyes keep straying from the screen, the Latin or Italian which made sense yesterday is just a wall. This is not to say that the crisis hasn’t been bad for me and others with chronic pain, it’s like having the condition twice at once; not like, it is having two chronic-traumatic conditions at once, the usual plus 2020 (and with quarantine many aspects of treatment and care are harder, higher risk, or just not possible). But having the skill set there was still invaluable.
So I joined my university’s committee to adapt teaching for COVID in the fall, and made these guides.
A lot of it is stuff that’s always helpful, but polished specifically for the current crisis. Here’s a little bit of the logic behind the things I tried to focus on, while the downloadable files themselves are more the focused methods than the abstract principles:
First, I tried to make it non-proscriptive. A lot of Pandemic productivity advice says “60% of people are more productive when they [wear a suit or whatever] so you must do it!” So the other 40% of people should, what, get bent? So I tried to focus on learning about yourself: try this work tip, then try the opposite, try different things to learn what helps you personally produce at your best.
Second, I talk about self-care as a work task and a duty. Culture pressures us to skip it, that when corners need to be cut we cut rest, play, and sleep. We shouldn’t. When we cut those, we start producing less (in quality and quantity) in those hours when we do work. It hurts our productivity as much if not more than cutting work tasks, and also makes us miserable. I remember a dreadful article a couple years ago with the thesis If you want to be successful you have to work 60 hours a week like these rich CEOs, but when you looked at the breakdown of what they called “work” they counted in those 60 hours all their commuting time, the gym & shower, power naps, tennis with a colleague, lunch meetings. So, to be clear, if you worked 9-5 with a 30 min commute on each end they were counting that as 10 hours’ work for the CEO but only 8 for employees, and if there was a trip to the gym and dinner or drinks with colleagues after work the CEO got to count those too, increasing 10 to 12, but the employee didn’t. If you counted only the activities that regular employees get to count as “work” the CEOs were working barely 35 max but the article was calling it 60 to advance this horrible false argument that only work-a-holics get ahead (a claim so many corporations want us to believe). You know what that article showed helps people get ahead? Having time in the day for rest, and exercise, and breaks, and games, and leisurely lunches, and spending time outside, and counting that self-care as vital to your work. Most jobs won’t count that as work (and the offices that do have nap rooms & massage chairs & lego rooms often do that to entice employees to also ridiculously late and never leave), but we can at least count self-care as work in our minds, and tell our guilt reflexes that this is not where corners should be cut.
(And if you’re a Terra Ignota fan, remember how the Utopian Oath requires you to promise to take that minimum of rest and play that’s necessary for your productivity? And remember that in their society twenty hours a week is the default work week? Utopia’s standards of rest and play are very high, and skipping the self-care that keeps you at your best is oathbreaking just as much as skimping on work. Also, Unusual Frequency has awesome new Utopian travel mugs if you want to reconfirm defeating death and attaining the stars! And Cousin flags if you want to affirm doing so kindly while taking care of yourself and others.)
Third, I talk about the brain like an organ. A body part. Which it is, one we push to its max a lot in daily life. So we should monitor it like one. Tennis elbow affects 40% of serious players, so it’s common sense for any tennis player to learn about tennis elbow and how to watch for it to set in. The latest studies I’ve seen show at least mild depression affecting 33% of undergrads, 41% of Ph.D. students, and in this pandemic it’s affecting a huge swath of the entire human race, so we should all learn about it, and watch for it, and train ourselves on how to mitigate it as much as possible while the symptoms are still mild, just like tennis players learn about tennis elbow. The only reason we don’t is that our culture stigmatizes problems with the brain totally differently from other organs, treats them as a failure of character or failure of will; they aren’t. “Push through the pain” is the wrong advice for that tendon in a tennis player’s elbow, and it’s the wrong advice for brain things too.
Fourth, I tried to stress in both documents that, in this situation, breaking down is normal. Lower productivity is normal. Grief, exhaustion, short tempers, snapping at friends, regretting it, they’re all normal. We have to plan for that, expect it, brace for it, recognize that in a team or in a household these months are going to be everybody taking turns having small breakdowns – if we prepare for that we can help each other prevent the *bigger* breakdowns that are the real problem. Voices inside will tell us that the days we wake up in the morning and sit down to work and just… can’t… deal, are bad, our fault, our weakness, failure, but all the neuroscience we have says it’s not our fault, it’s natural, it’s what brains do pushed past their limit, and our brains are past their limit. So on the mornings when you sit down to work and just can’t deal, and the self-doubt voice inside looms up to say weak! failure! push back. That voice is common too. I hear it. I still hear it after years of chronic pain and every saying that the pain is real, that I should take it easy, and all my friends being supportive, and my family, but something in our culture still makes us blame ourselves inside, weak! failure! So if you hear that voice on mornings when you just can’t deal, try to summon up another voice to shout back at it: everyone on Earth is breaking down. Today my job is not this task–today my job is to take care of myself, and protect the work I’ll be able to get done tomorrow.
A closing thought: Early in the pandemic the anecdote went viral that Isaac Newton came up with his theory of gravity while he was quarantining in the country from a plague, and many people (not jokingly enough) used it to say we should have high standards for what we produce in a pandemic, or that if we don’t set high standards it means we’re not geniuses like him. The true fact (historian here, this is my period!) is that Newton did theorize gravity while quarantining, but didn’t have library access, and while he was testing the theory he didn’t have some of the constants he needed (sizes, masses), so he tried to work from memory, got one wrong, did all the math, and concluded that he was wrong and the gravity + ellipses thing didn’t work. He stuck it in a drawer. It was only years later when a friend asked him about Kepler’s ellipses that he pulled the old notes back out of the drawer to show the friend, and the friend spotted the error, they redid the math, and then developed the theory of gravity. Together, with full library access, when things were normal after the pandemic. During the pandemic nobody could work properly, including him. So if anyone pushes the claim that we should all be writing brilliant books during this internationally recognized global health epidemic, just tell them Newton too might have developed gravity years earlier if not for his pandemic. And for a better historical model to use for how productive we should be in 2020, remember 1522-3, when Michelangelo was being hounded by lawsuits, and there was a political takeover crisis in his homeland, and he was so stressed he wrote later that he couldn’t touch a chisel the whole time, he couldn’t concentrate on any kind of art, too stressed and scared. Even Michelangelo, whom everyone agrees to call “genius.” Breaking down is normal for everyone, there are no special geniuses immune somehow to the slings and arrows of outrageous 2020. So next time you find a project taking longer than your planned, and your attention straying, and your ability to cope fading away, remember that if you’re getting anything accomplished in these months you’re already doing better than Michelangelo. And then do some self-care.
“If the Black Death caused the Renaissance, will COVID also create a golden age?”
Versions of this question have been going around as people, trying to understand the present crisis, reach for history’s most famous pandemic. Using history to understand our present is a great impulse, but it means some of the false myths we tell about the Black Death and Renaissance are doing new damage, one of the most problematic in my view being the idea that sitting back and letting COVID kill will somehow by itself naturally make the economy turn around and enter a period of growth and rising wages.
Brilliant Medievalists have been posting Black Death pieces correcting misconceptions and flailing as one does when an error refuted 50 times returns the 51st(The Middle Ages weren’t dark and bad compared to the Renaissance!!!). As a Renaissance historian, I feel it’s my job to shoulder the other half of the load by talking about what the Renaissance was like, confirming that our Medievalists are right, it wasn’t a better time to live than the Middle Ages, and to talk about where the error comes from, why we think of the Renaissance as a golden age, and where we got the myth of the bad Middle Ages.
Only half of this is a story about the Renaissance. The other half is later: Victorian Britain, Italy’s unification, World Wars I and II, the Cold War, ages in which the myth of the golden Renaissance was appropriated and retold. And yes, looking at the Black Death and Renaissance is helpful for understanding COVID-19’s likely impact, but in addition to looking at 1348 we need to look at its long aftermath, at the impact Yersinia Pestis had on 1400, and 1500, and 1600, and 1700. So:
This post is for you if you’ve been wondering whether Black Death => Renaissance means COVID => Golden Age, and you want a more robust answer than, “No no no no no!”
This post is for you if you’re tired of screaming The Middle Ages weren’t dark and bad! and want somewhere to link people to, to show them how the myth began.
This post is for you if you want to understand how an age whose relics make it look golden in retrospect can also be a terrible age to live in.
And this post is for you if want to ask what history can tell us about 2020 and come away with hope. Because comparing 2020 to the Renaissance does give me hope, but it’s not the hope of sitting back expecting the gears of history to grind on toward prosperity, and it’s not the hope for something like the Renaissance—it’s hope for something much, much better, but a thing we have to work for, all of us, and hard.
I started writing this post a few weeks ago but rapidly discovered that a thorough answer will be book-length (the book’s now nearly done in fact). What I’m sharing now is just a precis, the parts I think you’ll find most useful now. So sometimes I’ll make a claim without examples, or move quickly over important things, just linking to a book instead of explaining, because my explanation is approaching 100,000 words. That book will come, and soon, but meanwhile please trust me as I give you just urgent parts, and I promise more will follow.
Now, to begin, the phrase “golden age” really invokes two different unrelated things:
(1) an era that achieved great things, art, science, innovation, literature, an era whose wondrous achievements later eras marvel at,
(2) a good era to live, prosperous, thriving, stable, reasonably safe, with chances for growth, social ascent, days when hard work pays off, in short an era which—if you had to be stranded in some other epoch of history—you’d be likely to choose.
The Renaissance fits the first—we line up to see its wonders in museums—but it absolutely positively no-way-no-how fit the second, and that’s a big part of where our understandings of Renaissance vs. Medieval go wrong. So, our outline for today:
Renaissance Life was Worse than the Middle Ages (super-compressed version)
Where did the myth come from in the first place? (a Renaissance story)
Why is the myth of a golden Renaissance retold so much? (a post-Renaissance story)
Conclusion: We Should Aim for Something Better than the Renaissance
It’s also important to begin this knowing that I love the Renaissance, I wouldn’t have dedicated my life to studying it if I didn’t, it’s an amazing era. I disagree 100% with people who follow “The Middle Ages weren’t really a Dark Age!” with “The Renaissance sucks, no one should care about it!” The Renaissance was amazing, equally amazing as the Middle Ages, or antiquity, or now. I don’t love the Renaissance for being perfect. I love it because it was terrible yet still achieved so much. I love it because, when I read a letter where a woman talks of a nearby city burning, and armies approaching, and a friend who just died of the plague, and letter also talks about ideas for how to remedy these evils, and Xenophon’s advice for times of war, and how Plato and Seneca differ in their advice on patience, and the marvelous new fresco that’s been finished in the city hall. To find these voices of people who faced all that yet still came through it brimming with ideas and making art, that makes me love the human species all the more. And gives me hope.
In Florence, there are little kiosks near the David where you can buy replicas of it, and alongside the plain ones they have copies dipped in glitter paint, so the details of Michelangelo’s design are all obscured with globs of sparkling goo. That’s what the golden age myth does to the Renaissance. So when I say the Renaissance was grim and horrible, I’m not saying we shouldn’t study it it, I just want you to scrape off the glitter paint and see the details underneath: damaged, imperfect, a strange mix of ancient and new, doing its best to compensate for flaws in the material and mistakes made early on when teamwork failed, and violent too—David is, after all, about to kill an enemy, a celebration of a conquest, not a peace. Glitter drowns all that out, and this is why, while the myth of the golden Renaissance does terrible damage to how we understand the Middle Ages, it does just as much damage to how we understand the Renaissance. So let’s take a quick peek beneath the glitter, and then, more important, let’s talk about where that suffocating glitter comes from in the first place.
Part 1: Renaissance Life was Worse than the Middle Ages (super-condensed version)
The Renaissance was like Voldemort, terrible, but great.
On February 25th 1506, Ercole Bentivoglio, commander of Florence’s armies, wrote to Machiavelli. He had just read Machiavelli’s Deccenale primo, a history in verse of the events of the last decade. Bentivoglio urged Machiavelli to continue and expand the history, not for them, but for future generations, so that:
“knowing our wretched fortune in these times, they should not blame us for being bad defenders of Italic honor, and so they can weep with us over our and their misfortune, knowing from what a happy state we fell within brief time into such disaster. For if they did not see this history, they would not believe what prosperity Italy had before, since it would seem impossible that in so few days our affairs could fall to such great ruin.”
Of these days of precipitous ruin, Burkhardt, founder of modern Renaissance studies, wrote in 1869:
“The first decades of the sixteenth century, the years when the Renaissance attained its fullest bloom, were not favorable to a revival of patriotism; the enjoyment of intellectual and artistic pleasures, the comforts and elegancies of life, and the supreme interests of self-development, destroyed or hampered love of country.” (The Civilization of the Renaissance in Italy, end of Part 1)
Burkhardt seems to be describing a different universe from Bentivoglio, so desperate to prove to posterity that he tried his failing best to defend his homeland’s honor. Yet this was the decade that produced Leonardo’s Mona Lisa, Michelangelo’s David, Raphael’s Marriage of the Virgin, Bramante’s design for the new St. Peter’s Basilica, Josquin des Prez’s El grillo (the Cricket), the first chapters of Ariosto’s epic Orlando Furioso, and Castiglione’s first courtly works at the court of Urbino, soon to be immortalized in the Courtier as the supreme portrait of Renaissance culture. These masterworks do indeed seem to project a world of enjoyment and artistic pleasure in utter disconnect with Bentivoglio’s despair. Can this be the same Renaissance?
This double vision is authentic to the sources. If we read treatises, orations, dedicatory prefaces, writings on art or courtly conduct, and especially if we read works written about this period a few decades later—like Vasari’s Lives of the Artists which will be the first to call this age a rinascita—we see what Jacob Burkhardt described, and what popular understandings of the Renaissance focus on: a self-conscious golden age bursting with culture, art, discovery, and vying with the ancients for the title of Europe’s most glorious age. Burkhardt’s assessment was correct, if we look only at the sources he was looking at. If instead we read the private letters which flew back and forth between Machiavelli and his correspondents we see terror, invasion, plague deaths, a desperate man scrambling to even keep track of the ever-moving threats which hem his fragile homeland in from every side, as friends and family beg for frequent letters, since every patch of silence makes them fear the loved one might be dead.
Machiavelli’s correspondent, Ercole Bentivoglio, typifies the tangled political web which shaped these years. His father had been Sante Bentivoglio, who began as a blacksmith’s son and common laborer but was identified as an illegitimate member of the Bentivoglio family that dominated Bologna (remember Gendry in Game of Thrones?), so Sante was called to rule Bologna for a while when the only other adult Bentivoglio was murdered in an ambush, and young Ercole grew up in a quasi-princely court with all the grandeur we now visit in museums. Ercole’s mother was Ginevra Sforza, an illegitimate niece of Francesco Sforza who had recently conquered Milan, replacing the earlier Visconti dukes who had in turn seized the throne by treachery fifty-five years before. Renaissance politics isn’t turtles all the way down, it’s murders and betrayals all the way down.
Why was life in the Renaissance so bad? This is going to be a tiny compressed version of what in the book will be 100 pages, but for now I’ll focus on why the Renaissance was not a golden age to actually live in, even if it was a golden age in terms of what it left behind.
Let’s look at life expectancy: In Italy, average life expectancies in the solidly Medieval 1200s were 35-40, while by the year 1500 (definitely Renaissance) life expectancy in Italian city states had dropped to 18.
It’s striking how consistently, when I use these numbers live, the shocked and mournful silence is followed by a guy objecting: those numbers are deceptive, you’re including infant mortality—voiced as if this observation should discredit them. Yes, the average of 18 does include infant mortality, but the Medieval average of 35 includes it too, so the drop is just as real. If you want we can exclude those who die before age 12, and we do get a smaller total drop then, average age of death 54 in the 1200s dropping to 45-48 in 1500, so only a 12-16% drop instead of 48%, but the more we zoom the grimmer the Renaissance half proves. Infant mortality (within 12 months) averaged 28% both before and after 1348, so the big drop from Medieval to Renaissance Italy is actually kids who made it past the first year, only to die in years 2-12 from new diseases. We also think of the dangers of childbirth as lowering women’s lifespans, but death from childbirth stayed steady from Medieval to Renaissance at (for Tuscany) 1 death per 40 births, while the increase in war and violence made adult male mortality far higher than female even with the childbirth threat. If we look at the 20% of people who lived longest in Renaissance Italy it’s almost entirely widows and nuns, plus a few diehards like Titian, and poor exiled Cardinal da Costa of Portugal languishing in Rome to the age of 102, with everyone he’d known in the first 2/3rds of his life long gone. Kids died more in the Renaissance, adults died more, men died more, we have the numbers, but I find it telling how often people who hear these numbers try to discredit them, search for a loophole, because these facts rub against our expectations. We didn’t want a wretched golden age. (Demographics are, of course, an average, and different bits of Europe varied, but I’m using the numbers for the big Italian city-states precisely because they’re the bit of Europe we most associate with the golden Renaissance, so if it’s true there, it’s true of the Renaissance you were imagining.)
Why did life expectancy drop? Counter-intuitively the answer is, largely, progress.
War got worse, for one. Over several centuries, innovations in statecraft and policy (which would continue gradually for centuries more) had increased the centralization of power in the hands of kings and governments, especially their ability to gather funds, which meant they could raise larger armies and have larger, bloodier wars. Innovations in metallurgy, chemistry, and engineering also made soldiers deadlier, with more artillery, more lethal weapons, more ability to knock a town’s walls down and kill everyone inside, new daggers designed to leave wounds that would fester, or anti-personnel artillery designed to slice a line of men in half. Thus, while both the Middle Ages and Renaissance had lots of wars, Renaissance wars were larger and deadlier, involving more troops and claiming more lives, military and civilian—this wasn’t a sudden change, it was a gradual one, but it made a difference.
Economic growth also made the life expectancy go down. Europe was becoming more interconnected, trade increasing. This was partly due to innovations in banking (which had started in the 1100s), and partly, yes, the aftermath of the Black Death which caused a lot of economic change—not growth but change—some sectors growing, others shrinking, people moving around, people trying to stop people from moving around, markets shifting. There were also innovations in insurance, for example insuring your cargo ship so if it sinks you don’t go bankrupt like our Merchant of Venice. This meant more multi-region trade. For example, weaving wool into fine-quality non-itchy thread required a lot of oil, without which you could only make coarse, itchy thread. England produced lots of wool but no oil (except walnuts), so, in the Renaissance, entrepreneurs from England, instead of spinning low-profit itchy wool, started exporting their wool to Italy where abundant olive oil made it cheap to produce high-quality cloth and re-export it to England and elsewhere. This let merchants grow rich, prosperity for some, but when people move around more, diseases move more too. Cities were also growing denser, more manufacturing jobs and urban employment drawing people to crowd inside tight city walls, and urban spaces always have higher mortality rates than rural. Malaria, typhoid, dysentery, deadly influenza, measles, the classic pox, these old constants of Medieval life grew fiercer in the Renaissance, with more frequent outbreaks claiming more lives.
The Black Death contributed too—in school they talk as if the plague swept through in 1348 then went away, but the bubonic plague did not go away, it remained endemic, like influenza or chickenpox today, a fact of life. I have never read a full set of Renaissance letters which didn’t mention plague outbreaks and plague deaths, and Renaissance letters from mothers to their traveling sons regularly include, along with advice on etiquette and eating enough fennel, a list of which towns to avoid this season because there’s plague there. Carlo Cipolla (in the fascinating yet tediously titled Before the Industrial Revolution) collected great data for the two centuries after 1348, in which Venice had major plague bursts 7% of years, Florence 14% of years, Paris 9% of years, Barcelona 13% of years, and England (usually London) 22% in the earlier period spiking to 50% in the later 1500s, when England saw plague in 26 out of 50 years between 1543 and 1593. Excluding tiny villages with little traffic, losing a friend or sibling to plague was a universal experience from 1348 clear to the 1720s, when plague finally diminished in Europe, not because of any advance in medicine, but because fourteen generations of exposure gave natural selection time to work, those who survived to reproduce passing on a heightened immune response, a defensive adaptation bought over centuries by millions of deaths.
Today thousands of cases of Y. pestis (the plague bacterium) still occur each year, largely in sub-Saharan Africa and East Asia where it was not endemic so immunities didn’t develop. And if geneticist Mihai Netea is correct that the immune mutation which helps those of European descent resist Y. pestis also causes our greater rate of autoimmune disorders, then the Black Death is still constantly claiming lives through the changes it worked into European DNA over 400 years (and literally causing me pain as I type this, as my own autoimmune condition flares). While the 1348 pandemic was Medieval, most of the Middle Ages did not have the plague—it’s the Renaissance which has the plague every single day as an apocalyptic lived reality.
Economic growth also made non-military violence worse. Feuds (think Montagues and Capulets) were a Medieval constant, but the body count of a feud depends a lot on how wealthy the head families are, since the greater their wealth and the larger their patronage network, the larger the crowd of goons on stage in the opening scene of Romeo & Juliet when partisans of the two factions are biting their thumbs at each other, and the larger the number of unnamed men who also get killed in the background while Romeo fights Tybalt. In Italy especially, new avenues for economic growth (banking and mercenary work) quickly made families grow wealthy enough to raise forces far larger than the governments of their little city states, which made states powerless to stop the violence, and vulnerable to frequent, bloody coups. The Bentivoglios of Bologna and Sforza of Milan (whose marriage alliance produced Ercole who wrote that letter to Machiavelli) had risen by force, ruled by force, and were in turn overthrown by force, several times each, in fact, as rulers were killed, then avenged by returning sons or nephews, and cities flip-flopped between rival dynasties every few years:
In the 1400s most cities in Italy saw at least four violent regime changes, some of them as many as ten or twelve, commixed with bloody civil wars and factional massacres, until all Italy’s ruling houses were so new that the Knights Hospitaller—who normally required knights to have been noble four generations to join—let Italians in with only two generations because otherwise there would have been no one. Petrarch talked about this in his poem Italia Mia, which we think was written by 1347 (i.e. before the Black Death); he described Italy’s flesh covered with mortal wounds, caused by “cruel wars for light causes, and hearts, hardened and closed/ by proud, fierce Mars,” and his poor poem begging Italy’s proud, hard-hearted people for, “Peace, peace, peace.” It sounds just like what Ercole described to Machiavelli, doesn’t it? Well, Petrarch’s poem is as far from Machiavelli’s history as Napoleon’s rise from Yuri Gagarin’s space flight, a long time during which the wars grew worse, armies bigger, cities richer, plagues more frequent, steady escalation of the same things Petrarch feared would wipe out Italy 150 years before.
Important: none of this was new in the Renaissance! These were all gradual developments: banking, trade, centralization, the cultural produce of the Renaissance too (paintings, cathedrals, music, epics), these had all been gradually ramping up for centuries, changing the character of Europe decade by decade. Banking innovations started in the 1100s, insurance innovations in the 1300s, economic shifts before as well as after 1348, political shifts accumulated centuries, it’s all incremental. Thus, when I try to articulate the real difference between Renaissance and Medieval, I find myself thinking of the humorous story “Ever-So-Much-More-So” from Centerburg Tales (1951). A traveling peddler comes to town selling a powder called Ever-So-Much-More-So. If you sprinkle it on something, it enhances all its qualities good and bad. Sprinkle it on a comfy mattress and you get mattress paradise, but if it had a squeaky spring you’ll never sleep again for the noise. Sprinkle it on a radio and you’ll get better reception, but agonizing squeals when signal flares. Sprinkle it on the Middle Ages and you get the Renaissance. All key qualities were already there, good things as well as bad, poetry, art, currents of trade, thought, finance, law, and statecraft changing year by year, but add some Ever-So-Much-More-So and the intensity increases, birthing an era great and terrible. Many different changes reinforced each other, all in continuity with what came before, just higher magnitude, the fat end of a wedge of cheese, but it’s the same cheese on the thin end too. The line we draw—our slice across the cheese—we started drawing because people living in the Renaissance started to draw it, felt it was different, claimed it was different, and their claims reordered the way we think about history.
Some more quick un-fun facets of Renaissance life: while the Medieval Inquisition started in 1184, it didn’t ramp up its book burnings, censorship, and executions to a massive scale until the Spanish Inquisition in the 1470s and then the printing press and Martin Luther in the 1500s (Renaissance); similarly witchcraft persecution surges to scales unseen in the Middle Ages after the publication of the Malleus Maleficarum in 1486 (Renaissance); and the variety of ingenious tortures being used in prisons increased, rather than decreasing, over time. Rule of thumb: most of the scary practices we think of as “Medieval” were either equally true of the Renaissance, worse in the Renaissance, or only started in the Renaissance. If you want corrupt popes, they too can be more terrible as they get richer. And pre-sanitation, the more luxury goods traveled, the more people grew wealthy, the wider the variety of food people ate, and with more kinds of foods came more different kinds of parasites living in your intestines eating your from the inside out, hooray! Even in the Middle Ages we can tell your social class from the variety of parasite eggs in your preserved feces (the more you know!), but in the Renaissance the total could go up, and the frequency and intensity of chronic pain with it (not to mention a wider variety of horrible toxic things doctors would try to feed you as a cure; before sanitation more doctors = bad, not good).
In sum, if you’re a time traveler and you’re being banished, don’t pick the Renaissance.
As for how an age so terrible to live through produced the masterpieces and innovations we still hold in awe, my ultrashort answer is that Renaissance art and culture was also a gradual ramp-up from ever-changing Medieval art and culture, and that the leaps we seem to see in the later period are the desperate measures of a desperate time.
Legitimacy is a key concept here. The secret we all know is that governments, countries, laws, they’re all just a bunch of stuff we made up. They exist only as long as we all keep agreeing they exist, and act accordingly. Far more than Tinkerbell, regimes and governments need us to believe in them or they die. Sometimes this death takes the form of people just ignoring old structures, like in the Hellenic age when a remote Greek colony might hear from the founding city so infrequently that it starts ignoring the empire and just makes its own government. A more common consequence when people stop believing in governments is that some rival will take advantage of that lack of confidence, and rise up to claim power instead, whether through an electoral primary challenge or a bloody civil war.
For this reason, regimes to work hard to gain legitimacy, that is to acquire any and all things that make people agree the regime is real, and has the right to rule. When a usurper murders the old king but marries his widow, sister, or daughter, that’s an attempt to secure legitimacy in a world where people are used to government going with blood right. When no local royal-blood bride is available, the usurper might instead marry a princess from a famous distant kingdom, and fill his court with expensive, exotic treasures and other indications that he’s connected to foreign powers and money—this is another bid at legitimacy since it implies the new ruler has strong allies and the means to bring prosperity and trade. There are lots of other ways to project legitimacy: getting trusted local elites to work for you, getting religious leaders to bless you, publishing your pedigree (fake or real) with mighty ancestors, cracking down on crime and having showy trials, paying an astrologer to circulate your horoscope with great predictions, mounting a big parade, building an equestrian statue of yourself in the square that everyone walks past, receiving ambassadors in a showy way so everyone sees how much foreign powers honor you, repairing bridges and caring for orphans so people talk about your generosity and virtue, even a modern city funding a zoo and orchestra and art museum is that city projecting legitimacy with the trappings we associated with cultured power. When a regime has lots of sources of legitimacy, it makes people more willing to go along with that regime continuing. Some sources of legitimacy tie into a culture’s traditional ideas about what makes power lawful (religion, heredity, virtue, particular values), while other sources of legitimacy, like a collection of exotic animals or a fancy palace, just impress people, and make them feel that life under this regime will probably be good, and that overthrowing it would probably be difficult if it has money to throw away on palaces and elephants.
Thus the radical oversimplification is that, when times get desperate, those in power pour money into art, architecture, grandeur, even science, because such things can provide legitimacy and thus aid stability. Intimidating palaces, grand oratory, epics about the great deeds of a conqueror, expensive tutors so the prince and princess have rare skills like Greek and music, even a chemical treatise whose dedication praises the Duke of Such-and-such, these were all investments in legitimacy, not fruits of peace but symptoms of a desperate time. In an era when a book cost as much as a house (it really did!), and Florence’s Laurenziana library cost more per GDP than the Moon Landing, you don’t get that level of investment unless elites think they’re going to get something out of it. Just as today giant corporations fund charities or space tech because they get something out of it, publicity raising their stock prices, so a mighty merchant family might repair a church or build a grand public square and put their coat of arms on it, drawing investment and intimidating rivals.
Culture is a form of political competition—if war is politics by other means, culture is too, but lower risk. This too happened throughout the Middle Ages, but the Renaissance was ever-so-much-more-so in comparison, and whenever you get a combination of (A) increasing wealth and (B) increasing instability, that’s a recipe for (C) increasing art and innovation, not because people are at peace and have the leisure to do art, but because they’re desperate after three consecutive civil wars and hope they can avoid a fourth one if they can shore up the regime with a display of cultural grandeur. The fruits fill our museums and libraries, but they aren’t relics of an age of prosperous peace, they’re relics of a lived experience which was, as I said, terrible but great.
All this I’ll explore further in the book, but if you want more info in the meantime you can get an excellent overview of the period in Guido Ruggiero’s The Renaissance in Italy, and a look at how this fed philosophical innovation and birthed Renaissance humanism in James Hankins’s Virtue Politics. For today, though, our goal isn’t to look deeply at the David, it’s to look at the glitter we just scraped off it, and to understand where that glitter comes from.
Part 2: Where did the Myth Come From in the First Place? (A Renaissance Story)
Whenever I’m with Medievalists and the subject turns to one of the bad things people say about the Middle Ages (dark age, backwards, superstitious, stagnant, oppressive, enemy of progress, all homogenous), I make a point of speaking up and saying, “Yeah, that’s my guys’s fault. Sorry.” It was a joke the first time, and it’s still half a joke, but I keep doing it because there’s this special smile under the resulting chuckle, this pause, warming, affirming, on the Medievalist’s face that says: I’ve always felt I deserved an apology from the Renaissance! Thank you!
Because the beginning of the problem was the Renaissance’s fault.
Pretty-much every culture, when it tells its history, divides it into parts somehow (reigns, eras, dynasties). These labels may not seem like a big deal, but they have a huge effect on how we imagine things. Think of how the discourse about boomers vs. Gen-X vs. millennials affects people’s self-identities, who associates with whom, and the kinds of discourse we can have with those terms that we couldn’t have with different ones. The lines and labels in our history are powerful. In my Terra Ignota science fiction novels I mention that the people in my 25th century society debate whether World War I ended in 1945 or 1989, and it always blows readers’ minds for a few seconds, and then follows the reflection: yeah, I could see WWI and WWII being considered one thing, like the Wars of the Roses. My first exposure to the way this makes your brain go *whfoooo* was as a kid and hearing Eugen Weber provocatively call WWI and WWII “The Second Thirty Years War”. Feels weird, right? Weird-powerful.
People living in the European and Mediterranean Middle Ages generally (oversimplification) divided history into two parts, BC and AD, before the birth of Jesus and after. For finer grain, you used reigns of emperors or kings, or special era names from your own region, i.e. before or after a particular event, rise, reign, or fall. There was also a range of traditions subdividing further, such as Augustine’s six ages of the world which divided up biblical eras (Adam to Noah, Noah to Abraham, etc.), though most of those subdivisions are pre-historical, without further subdivision post Christ’s Incarnation. The Middle Ages also had a sense of the Roman Empire as a phase in history, but it was tied in with the BC-to-AD tradition, and with ideas of Providence and a divine Plan. Rome had not only Christianized the Mediterranean and Europe through the conversion of Constantine c. 312 CE, but authors like Dante stressed how the Empire had been the legal authority which executed Christ, God’s tool in enacting the Plan, as vital to humanity’s salvation as the nails or the cross. Additionally, many Medieval interpreters viewed history itself as a didactic tool, designed by God for human moral education (not the discipline of history, the actual events). In this interpretation of history, God determined everything that happens, as the author of a story determines what happens. The events of the past and life were like the edifying pageant plays one saw at festivals: God the Scriptwriter introduces characters in turn—a king, a fool, a villain, a saint—and as we see their fates we learn valuable lessons about fickle Fortune, hypocrisy, the retribution that awaits the wicked, and the rewards beyond the trials and sufferings of the good. The Roman Empire had been sent onto the world’s stage just the same, a tool to teach humanity about power, authority, imperial majesty, law, justice, peace, offering a model of supreme power which people could use to imagine God’s power, and many other details excitedly explored by numerous Medieval interpreters. (Many Renaissance interpreters still view history this way, and the first who really doesn’t do it at all is Machiavelli.)
The two people most directly responsible for inventing the Middle Ages are two men from Tuscany: Petrarch (Francesco Petrarca, 1304-1374), and Leonardo Bruni (1370-1444).
Petrarch was the first person to talk about the era after the Roman empire as a separate, bad period of shadow, misery, darkness, and decay. Petrarch gained his fame with his Italian poetry, and popularized the sonnet (though we have a long time still to wait for Shakespeare), but later in his life he was part of a circle of Italian scholars who loved, loved, loved, loved Cicero, and read his political works intensively as they thought about questions of republicanism and statecraft. Petrarch described himself as having been born in exile. He was born in exile in space quite literally, while his parents were in banishment, and he grew up in Avignon in the period the papacy was there in French control. But he also considered himself an exile in time, exiled from that community of antiquity which was the true home of his spirit. I already quoted his lament Italia Mia, and his sense of the degeneration of his era was enhanced by the feeling that France’s control of the papacy had ravaged and spoiled Rome and Italy. He also lived through the Black Death, and lost almost all his scholar-friends in it. Two surviving friends wrote to him after the main wave had passed to plan a precious reunion—they were attacked by bandits on the way, and one murdered, while the other escaped but was missing for many months. You can understand why Petrarch, reading of the Pax Romana when the ancient texts claim you could walk in safety from one end of Rome’s empire to the other, might see his age as one of ash and shadow. He projected that ash and shadow back on everything since Rome, lumping together for the first time the long sequence we now refer to as the Middle Ages.
Petrarch, importantly, did not claim his era was already a golden age, nor did he use the word Renaissance; he claimed his era needed tohave a transformation, that desperate times called for desperate measures, and that if Italy was to have any hope of healing it must look to its ancient past, to Rome, the Pax Romana, that dream age when there were no bandits on the road or pirates in the sea. The lost arts that nurtured the age of Emperors were languishing in ancient tomes waiting to be restored if only people reached for them. We know the Renaissance as the era that revived a lot of lost Roman technologies, geometry, engineering, large-scale bronze work, and those were important, but what Petrarch really thought would change things were people, intellectual technologies, not science or engineering tools.Petrarch wanted the library that educated Cicero, and Seneca, and Caesar. When we today look at ancient Rome we’re often struck most by the wicked Emperors, Caligula, Nero, the anecdotes of decadent corruption, but Petrarch instead saw the republican Brutus, who executed his own sons when they conspired to take over the state—in a world where city after city was falling to monarchal coups, and Lord Montague was used to using his great influence to make the Prince let Romeo get away with murder, the thought of Brutus putting Rome before his family felt like a miracle. (Unhelpfully, Petrarch didn’t write a single clear treatise where he spelled this out, but if you want a sample try his letters and invectives, or for the mega-thorough scholarly version see James Hankins’ Virtue Politics).
Important: even using antiquity wasn’t new in the Renaissance. Medieval people had been reading Seneca, and Cicero, and Virgil the whole time, and imitating and reusing ancient stuff, they just used the classics differently from how Petrarch did, just as the classics are also used differently in the 17th century, and the 19th century, and today. There were some major innovations in Renaissance engagement with the classics (several stages of innovation in fact), that differentiate them from Medieval, but those are complexities for another day.
Leonardo Bruni was the next step. He was child when Petrarch died, and grew up in the era of heady excitement of trying to use classical education to create the golden age Petrarch proposed. Bruni studied Latin with a focus (as Petrarch encouraged) on imitating ancient Latin instead of Medieval Latin whose grammar and vocabulary had evolved (as any language does) over the centuries. Bruni served as Chancellor of Florence, and imitated ancient Roman historians in writing his History of the Florentine People, which for the first time formally divided history into three parts: ancient, middle, and modern, which we now call Renaissance. He also filled his history with analysis and deep interpretation, which many Renaissance scholars will tell you was the first modern history, the first history of a post-classical time/place, and the first truly analytic history written since antiquity, and then Medievalists will scream at them and pile up examples of Medieval chronicles full of framing and moral analysis, which absolutely are doing sophisticated interpretive work, and vary enormously from each other, but Bruni’s is recognizably as different. Why? Largely because Bruni actively wanted his history to seem innovative and different, and wrote with that as a goal, in a new kind of Latin, with new structure, setting out to make something everyone would look at and say: Wow, it’s like what the Romans did!
With Bruni we had three periods—ancient, medieval, and the new age. That new age wasn’t called rinascita until Vasari’s Lives of the Artists in 1550 (more than a century after Bruni) and renaissance proper was coined by Jules Michelet in 1855, but Bruni’s idea of three periods, and that this new one could be a golden age, caught on quickly because of its potential for… (da da da daaa!) …legitimacy! Back then, as now, claiming that you’re the start of a new golden age is an ideal way to make your (teetering, illegitimate) regime seem exciting, full of momentum, glorious. History-writing modeled on Bruni quickly became all the rage, and you could awe people with a history of how great your city/people/family is, get them excited about a golden age, make yourself seem legitimate. And Bruni’s history writing had another power too.
One set of events Bruni described in his Florentine History were the conquests of Gian Galeazzo Visconti the “Viper of Milan” (1351-1402), a man who lived up in every way to his badass family crest of a serpent swallowing a helpless little dude. After ambushing and supplanting his uncle, the Viper seized Milan (bribing appropriate powers to make him duke), then took Verona, Vicenza, Padua, and tried for all of northern Italy including Bologna and Florence, securing a great victory at the Battle of Casalecchio in 1402. But then (according to Bruni) brilliant Florentine cunning arranged the would-be conqueror’s defeat and downfall. When Bruni’s history circulated in 1444, the Viper’s grandson Duke Filippo Maria Visconti did a spit take: “What the?! We didn’t lose that war! Granddad dropped dead of a f*ing fever and the troops had to go home! The Florentines never beat us in a single battle! They can’t say won the war!” They can. They did.
It turns out history isn’t written by the winners; history is written by the people who write histories.
So, what are you going to do about it, grandson of the Viper of Milan? There’s only one thing to do: hire one of these new classically-educated humanisty types to write a history of your city and your family framed your way, and replacing the murdered-his-uncle bribed-the-king totally-illegitimate conquest-by-force narrative with a glorious lineage that constantly kicked Florence’s ass!! That’s what he did—that’s what everybody did, Milan, Venice, France, England, Hungary, Naples; everybody had to have a history, and all the histories claimed there had been a bad middle age, that it was over, and that we were now in the glorious classical-revival-powered new age which had the potential to surpass it thanks to the virtues and glories of [Insert Prince Here]. This is why, up in England, baby King Henry VI’s uncle Duke Humphrey of Gloucester tried to hire Leonardo Bruni to come to England and work for him, and write a history that would shore up the tenuous Lancastrian claim to the throne (we’re entering the Wars of the Roses here). And this is why, while Bruni stayed in Florence, another major Florentine figure Poggio Bracciolini actually was lured by the high pay to go to England and work for Humphrey’s rival Cardinal Beaufort. And all these histories pick and choose details to make the current regime/ruler look great and legitimate, at the expense of making the newly-invented middle age look bad.
This is why all Medievalists, deep down inside, know they deserve an apology from the Renaissance.
One attempt at a solution is dropping the term Renaissance, but that doesn’t actually solve the problem, since it leaves us with antiquity and a period from then to… what? Is the dividing line the Enlightenment? Industrialization? Colonialism? The Industrial Revolution? The Agricultural Revolution? The French Revolution? WWI? No matter how late you push the line, any of these divisions is still accepting Bruni’s ancient-middle-modern division, and involves making a claim about what begins the modern. Normal parlance in history now is “early modern” which begins with [insert-scholarly-squabble-here] and ends roughly with the French Revolution, which is generally agreed to kick off “modern” proper. While “early modern” does avoid accepting claim that the Middle Ages were bad and needed a rebirth, and I use it myself, I also think it’s a dreadful term, since (A) it’s confusing (“early modern” sounds like the Crystal Palace, not Shakespeare’s Globe), and (B) the term actively worsens the degree to which your selected start date is a judgment call about what makes us modern. Because the real problem with the myth of the bad Middle Ages versus golden Renaissance is not what Petrarch and Bruni created within the Renaissance itself—it’s what happened later to entangle both terms with an equally problematic third term: modern.
Part 3: Why is the Myth of a Renaissance Golden Age Retold so Much? (a post-Renaissance story)
The thing about golden ages—and this is precisely what Petrarch and Bruni tapped into—is that they’re incredibly useful to later regimes and peoples who want to make glorifying claims about themselves. If you present yourself, your movement, your epoch, as similar to a golden age, as the return of a golden age, as the successor to a golden age, those claims are immensely effective in making you seem important, powerful, trustworthy. Legitimate.
In sum, one of the most powerful tools for legitimacy is invoking a past golden age. Under my rule we will be great like X was great! Whether it’s a giant golden age (Rooooome!) or a tiny golden age (the US 1950s!), if you can claim to be bringing it back, you can make a very clear, appealing case for why you should have power. This claim can be made by a king, a duke, a ruling council, a political party, an individual, or a whole movement. It can be made explicitly in rhetoric (I am the new Napoleon!) or implicitly by borrowing the decorative motifs, vocabulary, and trappings of an era. An investment banking service that uses a Roman coin profile as its logo, names its different mutual funds after Roman legions, and has a pediment and columns on its corporate headquarters is trying to project legitimacy from the idea of antiquity as a golden age of power and stability.
The newborn United States of America when it decided to make the Washington Monument be a giant obelisk, that was another bid at legitimacy and projecting power by invoking the golden ages of ancient Egypt and conquering Rome, combined in the Washington Monument’s case with other things like, instead of the traditional gold tip on top, using high-tech more-expensive-than-gold aluminum, mixing golden age with power claims about wealth and science.
…because the Renaissance had called itself a golden age, by the 17th century it had joined the list of epochs that you can invoke to gain legitimacy, and has been invoked that way many times. This is why 18th and especially 19th and earlier 20th century governments and elites raced to buy up Italian Renaissance art treasures and display them in their homes and museums. This is why Mussolini, while he mostly invoked imperial Rome, used the Renaissance too, and even made special arrangements to meet Hitler inside the Vasari Corridor in Florence to show off the art treasures of the Uffizi. And this is why the US Library of Congress building is painted all over inside with imitations of Renaissance classicizing frescos and allegorical figures in Renaissance style even though the quotations they include and values they celebrate are largely not Renaissance.
One consequence of golden ages being so powerful is that powers squabble over them: “I’m the true successor of [XXX]!” “No, I’m the true successor!” You see this in the fascinating modern day dispute over the name Macedonia in which both Greece and the country now called North Macedonia both want to be seen as the land of Alexander the Great, and argued over the name tooth and nail, dragging in both the UN and NATO. Since golden ages are mythical constructions (the events are real but the golden age-ness is mythmaking) they’re easy to redefine to serve claims of true successor status—all you have to do is claim that the true heart that made the golden age great was X, and the true spirit of X flourishes most in you. Any place (past or present) that calls itself a new Jerusalem, new Rome, or new Athens is doing this, usually accompanied by a narrative about how the original has been ruined by something: “Greece today is stifled by [insert flaw here: conquest, superstition, socialism, lack of socialism, a backwards Church, whatever], but the true spirit of Plato, Socrates and the Examined Life flourish in [Whateverplace]!”
Ancient Rome is particularly easy to use this way because Rome had several phases (republic, empire, Christian Rome) so if some rival has done a great job declaring itself the New Roman Empire you can follow up by saying the Empire was the corrupt decadent period and the Roman Republic was the true Rome! Simply quote Cicero and talk about wicked emperors and you can appropriate the good Rome and characterize your rivals as the bad Rome. If republic, empire, and Christian Rome are all claimed, you can do something more creative like the 19th century romantic movement which claimed the archaic pastoral Rome of Virgil’s Georgics, replacing pediments and legionary eagles with garlands and shepherds and claiming a mythic golden age no one had been using lately.
The same is true of claiming Renaissance. If you can make a claim about what made the Renaissance a golden age, and claim that you are the true successor of that feature of the Renaissance, then you can claim the Renaissance as a whole. This is made easier by the fact that “the Renaissance” is incredibly vague. When did it start? 1400? 1350? 1500? 1250? 1550? 1348? When did it end? 1600? 1650? 1700? You can find all these dates if you dig through books about “the Renaissance” written in different countries and different fields (art history, literary history). I pointed out that Petrarch’s Italia Mia is as far from Ercole’s Bentivoglio’s letter to Machiavelli as Napoleon’s rise from Yuri Gagarin’s space flight, but even at Machiavelli we’re still only half-way through the large, vague period that different people label Renaissance. On my own university campus, if I drop by different departments and ask colleagues when Renaissance begins, I get 1200 or 1250 from the Italian lit department (some of whom say Machiavelli is already “modern”), but in the English building I might get 1450 or even 1500. I think drawing a line after Black Death makes sense for Italy at least, or maybe at 1400, but there are plenty of counter-arguments, and people on campus who identify as Medievalists who study things later than some things I work on. I think it’s great for Medieval and Renaissance to overlap, since I—looking mainly forward—ask different questions about someone like Petrarch from the questions my Medievalist colleagues ask. The only “wrong” answer to where the line falls, in my opinion, is to believe there is a clear line.
And if we zoom into this long, vague period, when was the “high Renaissance” i.e. the best part, the most characteristic part? If you ask a political scientist it’s usually the very early 1400s, when Bruni and other innovative political thinkers were writing; if you ask an art historian it’s the decades right after 1500 when ¾ of the Ninja Turtles overlapped; if you ask a theater scholar it’s Shakespeare who was born fully 200 years later than Bruni and his peers discussing politics. It all depends on what you think defines the Renaissance, so if you have a different focus then different dates feel like periphery or core.
So, just as when we invoke Rome we can pick republican Rome, imperial Rome, pastoral Rome, Christian Rome, the conquering Rome of Julius Caesar or the peaceful Rome of the Pax Romana, similarly there are a huge range of Renaissances one can invoke: Bruni’s, Raphael’s, Machiavelli’s, Luther’s, Shakespeare’s. But choosing your Renaissance is an especially potent question because of… (drumroll please)… the X-Factor.
Okay, deep breath.
After the Renaissance, in the period vaguely from 1700 to 1850, everyone in Europe agreed the Renaissance had been a golden age of art, music, and literature specifically. Any nation that wanted to be seen as powerful had to have a national gallery showing off Renaissance (mainly Italian) art treasures, and capital buildings with Renaissance neoclassical motifs, while an individual with elite ambitions had to know classicizing Latin, and a bit of Greek, and have opinions about Raphael, Titian, Petrarch, and the polyphonic motets of Lassus. Seriously: in the original Doyle Holmes stories, so 1850-1910, after having Watson establish Holmes’s “Knowledge of literature—nil. Philosophy—nil.” still has Holmes carry a pocket Petrarch and write a monograph on the polyphonic motets of Lassus, because that’s what a smart, impressive person did in 1850. This also meant that Renaissance art treasures were protected and preserved more than Medieval ones—if you’re valorizing the Renaissance you’re usually criticizing the Middle Ages in contrast, so these generations learned to think of Renaissance art as good taste and the periods on both sides (Medieval and baroque) as bad taste, and a lot of great Medieval art was left to gather dust, or rot, or was even actively destroyed, since nothing invokes the Renaissance like sweeping away the “bad” medieval. As a result, the Renaissance became a self-fulfilling source base: go to a museum today and you see much more splendid Renaissance art than Medieval, leading to the natural conclusion that the Renaissance produced more art in general, but Middle Ages did make splendid art, it’s just that later centuries didn’t preserve it as carefully, so less survives, and what survives is more likely to be in storage than in the main gallery.
The transition from people being excited about Renaissance art and culture to being excited about the Renaissance as an era came in the mid-1800s, primarily with the work of Swiss historian Jacob Burkhardt, and his 1869 The Civilization of the Renaissance in Italy. It’s a gorgeous read, unskimmably rich prose, and Burkhardt’s work was a major breakthrough moment for the practice of history as a whole, because he showed how you could write a history, not of a country or a person, but of a culture, discussing the practices and ideas of an era, examining art and artists side-by-side with authors, soldiers, and statesmen as examples of people of a period and the way they thought, acted, and lived. The book pioneered cultural history, the practice of trying to study societies and their characteristics, acknowledging the interrelationship of politics with art and culture instead of examining them separately. Cultural history remains a major field, and one where some of the best work on once-neglected topics like women, pop culture, and non-elites has flourished. But…
Burkhardt was also the main figure who popularized the terms “modernity” and “modern.” He argued that the Renaissance was the birth of “modern man,” and that modern man was defined by a powerful sense of human excellence and human potential. According to Burkhardt, the core of this change—the spirit of the Renaissance which sparked the triumphant path of progress toward modernity—was the rise of individualism. As he says in the beginning of Part II:
In the Middle Ages both sides of human consciousness—that which was turned within as that which was turned without—lay dreaming or half awake beneath a common veil. The veil was woven of faith, illusion, and childish prepossession, through which the world and history were seen clad in strange hues. Man was conscious of himself only as member of a race, people, party, family, or corporation—only through some general category. In Italy this veil first melted into air; an objective treatment and consideration of the state and of all the things of this world became possible. The subjective side at the same time asserted itself with corresponding emphasis; man became a spiritual individual, and recognized himself as such.
The Medievalists reading this are gnashing their teeth, and yes, this moment is core to the persistence of the myth of the bad, backward, stagnant, sleepy middle ages, and equally core to the myth of the Renaissance Man: awake, ambitious, aware of his own power, rational, ripping through the cobwebs of superstition, desirous of remaking the world but also of intentionally fashioning him or herself into something splendid and excellent. A human being who realizes human beings can be their own masterpieces.
In the mid-19th-century, when Burkhardt wrote, Europe was very enamored of individualism, of new democratic ideas of government, of nationalism and ideas of individual consciousness and national consciousness, and of the notions of genius, both genius individuals and the geniuses of peoples. Thus, Burkhardt’s claim that the Renaissance was born from individualism gave all sorts of 19th century movements the ability to claim the Renaissance golden age as an ancestor. Germany, Britain, the young United States, despite having little to do with Italy, they could all claim to be the true inheritors of Renaissance greatness if they could claim that individualism and the opportunity to be a self-made man prospered more truly among their peoples than in Italy.
But there was more: by claiming that the Renaissance—and all its glittering art and innovation—was caused by individualism, Burkhardt was really advancing a claim about the nature of modernity. Individualism was an X-Factor which had appeared and made a slumbering world begin to move, sparking the step-by-step advance that led humanity from stagnant Medieval mud huts to towers of glass and iron—and by implication it would also define our path forward to an even more glorious future. In other words, the X-Factor that sparked the Renaissance was the defining spirit of modernity. If individualism was responsible, not only for the Renaissance, but for the wonders of modernity, then logically those regimes of Burkhardt’s day which most facilitated the expression of individualism could claim to be the heart of human progress and to hold the keys to the future; those nations which did not advance individualism (where socialism prospered, for example, or “collectivism” which was how 19th century Europe characterized most non-Western societies) were still the slumbering Middle Ages, in need of being awakened to their true potential by those nations which did possess the X-Factor of human progress.
I hope you winced a few times in the previous paragraph, recognizing toxic 19th century problems (eurocentrism, orientalism, “White Man’s Burden” thinking), as well as basic historical errors (spoiler: you can find plenty of individualism in Medieval texts, and lots of things that are absolutely not individualism in Renaissance ones). But those specifics aren’t the big problem. The big problem was how entrancing the idea of an X-Factor was, the notion that there is one true innovative spirit which defines both Renaissance and modern, and advances in a grand and exponential curve from Petrarch through Leonardo and Machiavelli on to [insert modern hero here]. Thus Burkhardt birthed what I call thequest for the Renaissance X-Factor. Because when the first scholars disagreed with Burkhardt, they didn’t objcet to the idea that the Renaissance was caused by a great defining X-Factor, they loved that idea, they simply argued about what exactly the X-Factor was.
Thanks to Burkhardt, the Renaissance came to be defined as the period after Medieval but before Enlightenment when something changed and pushed things toward modernity—the moment that the defining spirit of modernity appeared. From that point on, claiming you were the successor to the Renaissance didn’t just mean claiming a golden age like Rome, it let you also claim that modernity itself was somehow especially yours. If you could argue that the reason the Renaissance was great was that it did the thing you do, then you are the heart of modernity and progress, even of the future, while those who don’t celebrate that spirit are the enemies of progress. Thus every time someone proposed a new X-Factor, a different explanation for what made Renaissance different from Medieval, that made it possible to make new claims about the nature of modernity, and which nations or movements have it right. This model even lets one claim the future: the X-Factor was born in the Renaissance, grew in the Enlightenment and in modernity, and is the key to unlocking the next glorious age of human history as it unlocked both Renaissance and modern. This lets you advance teleological arguments about the inevitable triumph of [democracy, nationalism, atheism, capitalism, whatever]. It’s a version of history that’s not only legitimizing but comforting, since it lets you feel you know where history is headed, what will happen, who will win.
To give specific examples, if we’re in the middle of the Cold War, and an influential historian publishes a book arguing that the X-Factor that sparked the Renaissance was double-entry bookkeeping, i.e. the rise of banking and the merchant class, America can say: “The Renaissance X-Factor was the birth of capitalism! The fact that it was a golden age proves capitalism will make a golden age too, and the true successor of this golden age is our alliance of modern capitalist regimes!” If, on the other hand, we’re in a nationalist wave, say in 1848 or 1920, and someone argues that the X-Factor that sparked the Renaissance was the call for national unity articulated in Petrarch’s Italia Mia or Savonarola’s sermons (this is Pasquale Villari), and that what ended the Renaissance golden age was when Italy was conquered and divvied up among the Bourbons and Hapsburgs, then the Renaissance can be claimed as a predecessor by the Italian unification movement, the German unification movement, any nationalist movement anywhere can claim that uniting peoples into nations is what drives modernity. If we claim the Renaissance was birthed by the rise of secular thought, that Renaissance geniuses were the first people to break through the bonds of superstition, and that Leonardo and Machiavelli were secret atheists (this is Auguste Comte), then we can claim that secularization and the secular state is the heart of human progress and modernity. And if someone claims the X-factor was republican proto-democratic thought, the political writings and discourse of civic participation unique to the Italian city republics, Florence, Venice (this is Hans Baron), then we can claim that republican democracy is the key to human progress, that modern democracies are the heart of modernity, and everything else is backwards, outside, Medieval, bad, and needs to be replaced.
To this day, every time someone proposes a new X-Factor for the Renaissance—even if it’s a well-researched and plausible suggestion—it immediately gets appropriated by a wave of people & powers who want to claim they are the torch-bearers of that great light that makes the human spirit modern. And every time someone invokes a Renaissance X-Factor, the corresponding myth of the bad Middle Ages becomes newly useful as a way to smear rivals and enemies. As a result, for 160 years and counting, an endless stream of people, kingdoms, political parties, art movements, tech firms, banks, all sorts of powers have gained legitimacy by retelling the myth of the bad Middle Ages and golden Renaissance, with their preferred X-Factor glittering at its heart.
We scholars do our best to battle this, to introduce a complex and un-modern Renaissance, but the very usefulness of the myth guarantees that it will be repeated much more broadly than our no-fun efforts to correct it. A lot of Renaissance historians today reject the idea of a single X-Factor and try instead to talk about combinations of mixing factors. Many of us also try to argue that the Renaissance was not fundamentally modern, that it was its own distinctly un-modern thing. But it’s a hard sell, because the narrative of a special spirit launching us from Petrarch to the Moon Landing is enchanting, and because a complicated, messy, un-modern Renaissance snatches away the golden Renaissances most people meet first. Nobody in this century has read about the French Invasion of 1494, or even about the Guelphs and Ghibellines, before meeting the genius cults of Leonardo and Michelangelo.
Scraping the glitter off to reveal the imperfect and violent David underneath is an assault on our understandings of our past and present, on what it means to be ourselves, even on our sense of where the future is heading. People find that unsettling. And people who look to Renaissance celebrities as role models and intellectual ancestors don’t like to hear about their rough un-modern sides. So people get hostile, or unsettled, they keep telling the myths, and use cherry picked sources to glob the glitter-paint back on. It’s not always done in bad faith—if from early childhood you’ve always learned the Renaissance was sparkling and golden, and you see a bare patch where the glitter has come off, of course you’ll think that bare patch is the error, that the still-sparkly parts are the real thing. You treat the oddball patch as damage, and keep believing what that documentary or museum label told you years ago when you saw your first Renaissance masterpiece and fell in love. So the myth persists, and for every attempt to correct it we’re up against a dozen tour guide scripts, and TV specials, and corporate statements, and outdated textbooks, and new books (fiction and nonfiction alike) that glob the glitter on. So you can understand why, from time to time, Renaissance and Medieval specialists alike just have to stop and scream like Sisyphus.
Conclusion: We Should Aim for Something Better than the Renaissance
This, in not-very-brief, is why we keep telling the myth of the golden Renaissance, and bad Middle Ages.
Now, let’s look again at our other starting question: “If the Black Death caused the Renaissance will the COVID pandemic cause a golden age?” You see the problems with the question now: the Black Death didn’t cause the Renaissance, not by itself, and the Renaissance was not a golden age, at least not the kind that you would want to live in, or to see your children live in. But I do think that both Black Death and Renaissance are useful for us to look at now, not as a window on what will happen if we sit back and let the gears of history grind, but as a window on how vital action is.
The Black Death first: it didn’t cause the Renaissance, no one thing caused the Renaissance, it was a conjunction of many gradual and complicated changes accumulating over centuries (banking, legal reform, centralization of power, urbanization, technology, trade) which came together to make an age like the Medieval but ever-so-much-more-so. The idea that the Black Death caused a prosperity boom comes from old studies which showed that wages went way up after the Black Death, creating new possibilities for laborers to gain in wealth and rise in status (like the golden 1950s). But those were small studies from a few places (mainly bits of England), and we have newer studies now that show that wages only rose in a few places, that in other places wages didn’t rise, or actually went down, or that they started to rise but elites cracked down with new laws to control labor, creating (among other things) the first workhouses, laws limiting freedom of movement, and other new forms of unfreedom and control. What the Black Death really caused was change. It caused regime changes, instability letting some monarchies or oligarchies rise, or fall. It caused policy and legal changes, some oppressive, some liberating. And it caused economic changes, some regions or markets collapsing, and others growing.
If you really want to know what COVID will do, I think the place to look is not Renaissance Italy, but the Viking settlements in Greenland, which vanished around 1410. Did they all die of the plague? No. We’re pretty sure they never got the plague, they were too isolated. But the Greenland settlements’ economy had long depended on the walrus trade: they hunted walruses and sold the ivory and skins, and ships would come from Norway or Iceland to trade for walrus, bringing goods one couldn’t make in Greenland, like iron, or fine fabric, or wheat. But after 1348 the bottom dropped out of the walrus market, and the trading ships stopped coming. By 1400 no ships had visited Greenland for years except the few that were blown off-course by storm. And meanwhile there were labor shortages and vacant farms on the once-crowded mainland. So we think the Greenland Vikings emigrated, asked those stray ships to take them with them back to Europe, as many as could fit, abandoning one life to start another. That’s what we’ll see with COVID: collapse and growth, busts for one industry, booms for another, sudden wealth collecting in some hands, while elsewhere whole communities collapse, like Flint Michigan, and Viking Greenland, and the many disasters in human history which made survivors abandon homes and villages, and move elsewhere. A lot of families and communities will lose their livelihoods, their homes, their everythings, and face the devastating need to start again. And as that happens, we’ll see different places enact different laws and policies to deal with it, just like after the Black Death. Some places/regimes/policies will increase wealth and freedom, while others will reduce it, and the complicated world will go on being complicated.
That’s why I say we should aim to do better than the Renaissance.
Because we can. We have so much they didn’t. We know so much.
For one thing, we know how pandemics work. We know about germs, viruses, contagion, hand-washing, sanitation, lowering the curve. We can make plans, take action that does something. Forget 1348, even in 1918 we didn’t understand how to treat influenza, how it moved, and hand washing was still controversial. 1918 was a US election year but we didn’t discuss delaying or changing the election, there was nothing we could do to make it safer, we didn’t know about six-feet-apart, or sanitizing voting booths, or have the infrastructure to consider vote-by-mail, all we could do was let men (women still had two more years to wait) vote and die. We’ve come a long way.
This year, 2020, this is the first time in the history of this planet that any species has faced a pandemic knowing what it is, and how to take effective action. We aren’t taking perfect action, and we absolutely should be criticizing and condemning the many flaws—some small, some huge—in how it’s being dealt with, but there is real, efficacious action we can take. As an historian, not just of the plague of 1348, but of the plagues of 1435, and 1485, and 1494, and 1503, and 1596, and 1630, and 1656, what I see is those many generations who not only had to live through this over and over, but who had no hope that their children would ever be free of it. We know about vaccines, and that we’ll make one—it’ll take a while, and we’ll mess up various ways along the way, but none of us is afraid our grandchildren will grow up spending one year in ten locked up in their homes like this as COVID-19 spreads; we will solve it. We know we’ll solve it, and any other age in history would treasure that confidence like miracle. Because all Petrarch could say after losing his world in 1348 was that, the next time plague comes back, we should console ourselves by thinking of it as dying with much good company.
We know about mental health now too. We’re talking about the mental health crisis of COVID, the mental health costs of fear, poverty, racial injustice—in 1918 we were still excited by electroshock, and debating the radical new idea that outpatient psych treatment might be a thing, instead of doing only institutionalization. We have the language to talk about the mental cost of crisis, and that language alone opens so many possibilities for helping, acting, aiding that previous eras never had. Without the concept, we couldn’t start to try to treat it—now we can.
And we have more language: social safety net, social welfare, social services, concepts for thinking how state and society can put structures in place to relieve human suffering. We have economics now, not the kind of economics that’s trying to prognosticate the stock market, the basic kind with terms like GDP, and unemployment rate, and wealth gap, and retirement age, and inflation. There were economies in 1348, and even social services, hospitals, orphanages, city grain supplies, but we didn’t have a science for discussing it, vast banks of data comparing how different systems work, or help, or harm. After the Black Death when different places tried different policies for their recovery, they didn’t have comparisons, examples—we do. We won’t be guessing in the dark when each nation decides its recovery plan for this pandemic—we won’t be omniscient, but even partial knowledge makes us powerful. That raises the stakes.
Because, like after 1348, there is about to be big change. There are many options before us, different things that states can do post-COVID, some of which will help with poverty, empower labor, lend a helping hand to those exhausted Greenland Vikings as they start again, and there are other things states can do that will instead widen the gaps, entrench elites, help the rich get richer and see the disempowered locked more inescapably into modern versions of workhouses. Different places will make different choices. Some places will see regime changes, others just policy shifts, but there aren’t vast wheels of history that lock a pandemic into automatically yielding a boom or bust. There is no automatic outcome. Rather, all nations in the world are about to make a set of choices which will have a far larger, deeper impact on the next decades, on lives, rights, options, everything, than the normal choices states make in a normal year. The stakes are higher. Unlike in 1348 we have a lot of knowledge, answers, options, concepts we could try like safety nets, or UBI, or radical free markets, many very different things. Which means that acting now, demanding now, voting, pushing, proposing change, we’re shaping policies that will affect our big historical trajectory more than normal—a great chance to address and finally change systemic inequalities, or to let them entrench. There is no predetermined outcome of pandemic; pandemic is a hallway leading to a room where something big is going to be decided—human beings decide.
I love space exploration. I’ve written novels about it, and a song that makes everyone cry, I make myself tear up thinking about it all the time, especially civilian spaceflight and the hope that this chapter of history might be advanced by curiosity, teamwork, and human hope, not war or competition. But after looking forward to it for so long, the recent SpaceX launch was the first I’ve watched in a long time without tearing up. Because watching a space ship launch while looters are smashing shops outside my window (and cops ignoring them in favor of harassing peaceful protestors & giving carte blanche to the gunwielding vigilante on the corner) feels a lot like Leonardo painting the Mona Lisa while cities around were literally burning (and rich merchants’ private goons guarding their wealth & allies as faction dictated). This year, this specific year, 2020, with the world shut down by plague, and civil strife, and fire in the streets, and teetering distrust in governments, this is the first time our present has reminded me of the Renaissance. But we aren’t the Renaissance—we have social science, and efficacious medicine, and the Enlightenment under our belts, when we learned we can analyze our laws and institutions, and step by step replace them with better ones. We aim for better.
At the Renaissance Society of America Conference some years ago, two scholar friends got into a debate about whether Machiavelli’s world was fundamentally pre-modern, different from our own, or whether fundamentally it faced the same problems we do. Responding to the claim that the Renaissance was far more violent than our present, the advocate of Renaissance-as-modern quoted the statistic that modern Chicago had as many murders every year as Renaissance Florence. The rebuttal that surged in my mind was that the population of Florence was less than 100K, so Chicago’s millions have far fewer murders per capita, but the other speaker had a far better answer. We’re working to change that murder rate. We study it, understand it, plan interventions, act. We believe it’s a problem we can solve, should solve, that citizen and state should act, and if the state will not the state should change. We have policy studies, plans, alternatives.
Petrarch wanted to end the cruel wars for light causes that were wounding Italy, but had no plan beyond sending his poem out into the world, and urging elites to have their kids read Cicero. Machiavelli also wanted to end the cruel wars for light causes, and seeing that reading Cicero had failed he proposed a new way of evaluating history, collecting examples of what worked and didn’t in the past, basing our statecraft and actions on them so the next time we try things we’ll choose more wisely. It was the birth of social science. It took us a long time for us to get good at it, to turn the observations in The Prince into big databases and systematic studies, just as it took a long time for medicine to get from the four humors to our confidence that we can make a vaccine, but we can make one. We can make good social policy. Will we do it perfectly? No. Many bad policies will be advanced, just as vaccines and treatments will be distributed unfairly and slowed down by bigotry and selfishness. But we can do it, we have tools, as real in our hands and libraries as the knowledge of vaccines is real—tools Machiavelli and Petrarch would have given anything to have. We can aim for better than another Renaissance.
Below you’ll find some news, links to excerpts from Terra Ignota book 4, and a discussion of point-of-view in the craft of writing, but I’m posting today mainly to announce that I have an essay about censorship and its relationship with genre fiction in this month’s Uncanny Magazine, which is now free to read online. I’ve been researching censorship for a couple years now, in collaboration with Cory Doctorow, and it was delightful being invited to share my thoughts with the genre fiction writer/reader world because genre fiction has a lot of power to affect our ability to resist censorship, more power than I think most people realize.
It feels strange having an essay on an unrelated political topic come out in the midst of this Black Lives Matter & COVID double-crisis, but they are related if we look more carefully. Every time we see coverage of an injustice, news of brutality or cruelty, news of heroism; every time we talk about the challenges and importance of disseminating medical news; every time we read op-eds and cheer, or disagree, or criticize, share, or condemn; every time someone exposes a fake, a lie; in all these cases we’re depending on free speech, on journalism, analysis professional and amateur, on speech. Networks of digital communication are disrupting everything right now, by amplifying voices that have been silenced. In this, the digital revolution is following the pattern of past information revolutions (printing press, radio, etc.). During any information revolution, there is a cost of switching to the new medium: the cost of new equipment, the cost of setting up distribution and building an audience, the cost in time and effort of learning a new method; whether it’s buying a printing press or learning a new app, there is a cost. Consequently, the first adopters of a new information technology are always those who haven’t been able to communicate with the previous technology, the voices that have been silenced, the communities that have been unable to find each other, or have been actively silenced. First adopters of the printing press included scholars eager to spread rare Cicero to peers spread across Europe, but also religious and political radicals whose demands for change, banned on pulpits and in town squares, could circulate in pamphlets which authorities at first had no way to control. My wonderful colleague Kathleen Belew has done work on the ditto machine, the first cheap way to print small-scale at home, and how (much like the early internet) it empowered many communities to speak and find each other: LBGT+ groups, tropical fish hobbyists, civil rights activists, science fiction fans, linguistic minority communities, and also the KKK, which saw a surge of membership and activity via ditto-printed newsletters, just as it’s seen on digital platforms (Kathleen’s book Bringing the War Home: The White Power Movement and Paramilitary America is invaluable reading right now).
We are living in the crisis of uncontrolled expression that comes with any information revolution, the equivalent of 1517; most of the denunciations of Church corruption in Luther’s 95 Theses had been voiced by others many times before, but in 1517 pamphlet distribution was so fast-yet-uncontrolled that when he released the 95 Theses they were in print in London 17 days after he made them public, enabling a pan-European movement and making it impossible for those in power to silence the calls for change they had silenced before. Black Lives Matter has momentum now around the world, a call for change that can’t be silenced; the hate it battles also has momentum, and amid their clash another wave is gaining momentum, as it does in every information revolution: the wave of those in power (politicians, corporations, alarmed elites) wanting to silence the uncomfortable voices empowered by the new medium. We need to fight this battle too, a battle to find a balance between protecting the new ability of radical voices to speak while also protecting against hate speech, misinformation, and other forms of communication toxic to peace and democracy. As I explain in my essay, genre fiction, we who read it, we who write it, have a lot of power to affect the battle over censorship. These days are hard; as someone both disabled and immunocompromised I can’t go join the protests in the streets, not without both endangering fellow protesters by getting in their way, and the risk of this one moment of resistance destroying my ability to be here helping with the next one, and the next. But I can help on the home front as it were, working to protect the tools of free expression which those out on the streets depend on every minute, every protest, every video exposing cruel realities. Everything we do to strengthen speech and battle censorship protects our best tool, not just for this resistance, but for the next one, and the next. Change needs to speak, hope needs to speak, and one thing we absolutely have the power to do–especially we within the F&SF community–is to defend the instincts that defend free speech. So that’s what my Uncanny essay is about, and I hope it makes you feel a little bit more powerful, with the responsibility that brings. (And if you enjoy it, you can see more discussion of the question in this video, and please support Uncanny Magazine!)
Meanwhile, a few other things:
Two excerpts from Perhaps the Stars, the now-finished 4th book of Terra Ignota,were recently released as part of the Decameron Project, an awesome project which is raising money to protect refugees from COVID by posting a new story every day of the pandemic, in the spirit of Boccaccio’s Decameron. Lots of amazing authors have shared stories or novel excerpts, tons there to enjoy, and these two excerpts from book 4 are philosophical ruminations on the role of distance in the human condition, a bit topical in our current lockdowns and crisis.
I also recently did a giant 2.5 hour marathon podcast episode on the Singularity Podcast, where we talked about progress, technology, history, pandemics, teleology, the singularity, the history of knowledge, the crisis of the late 16th century when there was “too much to know”, what science fiction is for, how learning is a form of joy, and all sorts of things. I think it’s one of the best interview discussions I’ve ever done, and a lot of people have said they were wowed and comforted by it, so it’s highly recommended (when you have the time).
In happy personal news, the French edition of Too Like the Lightning (Trop semblable à l’éclair) has won the Grand Prix de l’Imaginaire 2020 award for Best Translation, for the heroic work Michelle Charrier did reflecting all the complicated things I do with gender, and archaic language, with my thee’s and thou’s and he’s and she’s. Michelle absolutely deserves it, and it makes me especially happy too. A few years ago I heard the French translation of Gene Wolfe’s Shadow of the Torturer had won France’s best translation award–that book was a huge influence on me and Terra Ignota, and I remember wishing at the time that I could someday write something complex and subtle enough that the French translation would require the same skill and mastery to translate. So I’m very happy for Michelle, but it’s also a little wish of mine come true.
Finally, at Balticon 54 last week Jo Walton and I did a craft-of-writing panel focused on point of view, and afterward we had some great discussions in text chat which I thought others might enjoy – I’ve condensed and saved a best-of here (with consent from those who asked the questions). The panel was also filmed and will be online once Balticon does the captioning, and when I figure out how to post a link to it I will do so here but right now I don’t know. Meanwhile, enjoy!
On Writing And Point of View
Question: What I don’t get is why they tell new writers to not have multiple POVs in a novel. I mean, if the story calls for it, and you’re clear on the change, why not?
Jo Walton: Minimizing POVs is good discipline because it’s very easy to get sloppy. So it’s one of those things that’s good advice when you’re starting out, but not a law.
Ada Palmer: I agree that minimizing POVs is often wise. Whenever I find myself wanting a scene to be in a different POV I think really hard about it. Sometimes it’s the right answer, but the fail condition is that you have too many POVs and the reader expects each of them to have follow-through and they don’t
Jo Walton: You need as many as you need, but no more. It’s good discipline as an exercise to try doing without the switch
Ada Palmer: In particular I think one challenge with multiple POVs, especially adding one part-way through, is making clear to the narrator whether we’ll keep seeing this POV again or not. Since once you switch to a new POV this could be a one-time thing, or it could be that we’re now going to constantly take turns, and if you’re doing the one-time thing but a reader thinks you’re doing the many-times thing then the reader can be distracted waiting for narrator #2 to come back and wondering how long it will be. In Too Like the Lightning we do switch narrators a couple times but I’m always careful to make very clear to the reader that it’ll be the exception not the rule. I do it using chapter titles, but it can be done many ways, but you want to make sure you think hard about whether the reader (A) expects this new POV to now be a frequent companion, (B) expects the new POV to be one-time or rare, or (C) doesn’t know. Any of these can be what you want (sometimes C is what you want) but you always want to decide clearly what you want the reader to be thinking, and communicate that in some way. I have a chapter called “Sniper’s Chapter” narrated suddenly by a character called Sniper, and it makes it very clear that it’ll be unique. But another time I might switch and intentionally give no clue how long it’ll be until we see the original narrator again, because I want the reader to not know. I want the reader to have that suspense.
Ada Palmer: A good way to summarize it is that you want to think carefully about whether you want the reader to be in suspense about what the narration will do next, and if so what you want that suspense to be. Do you want there to be very regular alternation of characters? Or irregular so it’s uncertain? Let’s imagine we were doing Romeo & Juliet but jumping POVs. If we regularly alternate Romeo, Juliet, Romeo, Juliet then the reader isn’t in suspense, but if we are mostly Juliet with only occasional Romeo chapters then during any given Juliet chapter the narrator will be wondering how long it’ll be until we see what Romeo thinks of things. Or if we were rotating Romeo, Juliet, Mercutio, Romeo, Juliet, Mercutio, and then we do the death of Mercutio during a Romeo chapter, during the next Juliet chapter the reader will be in suspense wondering what will happen when we would get to the next Mercutio chapter, and guessing (will it be a new character? will it be a ghost? will it be nothing?). That suspense may be something you can use effectively, or it may be something you don’t want that would distract the reader — that’s why you want to always think it through.
Jo Walton: I once did a chapter from the POV of a dead character in the rotation like that
Ada Palmer: We usually think of plot as the main source of suspense, but narration can be a big source of it
Jo Walton: Oh yes
Ada Palmer: One challenge that often makes one resort to a POV is when there’s information you want to get across that your existing narrator(s) wouldn’t have access to. In book 3 of Terra Ignota I had a real struggle at one point where we really needed to know what was going on with one faction (the Mitsubishi) who at that time would not be sharing their info with our narrator (or indeed with any of the narrators we’d had so far). It was a real puzzle for a long time and I was considering using another POV but there wasn’t any new POV that I felt would really justify itself since there wasn’t any character whose interiority I really wanted to share. In the end I realized I could have the narrator overhear a phone conversation in which the faction was talking to a different person that they did have reason to share info with, and that was a much simpler solution to the problem which didn’t require substantially changing the characterization of whoever I would’ve had be the interim narrator. At another point I use a newspaper clipping to achieve the same. Different source of info but no new POV – documents can be great that way, if what you need is to get info across that no narrator would see but you don’t want to gain another narrator.
Question: Would titling chapters by character (in addition to place/date-time) smooth out the potential bumps that make a POV change difficult for the reader?
Ada Palmer: It can, but it doesn’t smooth the first bump, only the later ones. It can be a tool to clearly signal when you’re changing characters (as can style, or tone, or subtitles, etc.) but it doesn’t get over the initial challenge of asking the reader to settle into and get to know a new voice. It’s hard the way your class being moved to a new classroom is hard, it requires adjustment. It’s disruptive. it needs to have payoff: ah, way better chairs! (worth-it) but if it isn’t worth-it, if there isn’t a clear gain (ah, new information!) it can be frustrating.
Question: Would also it help if the new POV has already been present, and to some degree characterized, before the switch?
Ada Palmer: Yes and no. It makes the switch easier, and for people who like that character it’ll make that character be excited. On the other hand, the more new and unexpected the POV the more new and unexpected the information and perspective and expansion of experience that can come with it. In our Romeo & Juliet, if when Mercutio dies our new POV is Friar Lawrence, that’s pretty easy, we know him already, we get to learn more about why he thinks his stupid plan is a good plan, it all works. If the POV is instead Lord Capulet it’s more unexpected, he’s been unsympathetic so far, we don’t know from the play what he’s doing during much of the action, we aren’t excited to enter his head but we’re learning more new stuff and our perspective is more radically different. If our new POV jumps to Cesare Borgia who is camped with his army having just sacked Modena and is thinking of trying to conquer Verona, and his scouts have just come to tell him that Verona is weak thanks to this feud with the Montagues and Capulets and that if Borgia allies with one of them and promises to fight the other it could aid his conquest of the city, suddenly our mind is blown and the story is totally different–but the author really has to follow up on this, and can’t just have it be a one-off! It has to become what the rest of the book is about! Jumping instead to Friar Lawrence or Lord Capulet could be a one-off and still work.
Jo Walton: I want Cesare conquers Verona! In Daniel Abraham’s Dagger & Coin series, he uses absolutely standard fantasy different POVs by chapter, but he uses kinds of characters we don’t expect as POVs. Like Lord Capulet would be. There is in fact an elderly conservative lord who wants people to get off his lawn. You never see POVs like that, so it’s refreshing.
Question: So zooming way out, is the PoV character the one where you know some of their thoughts? How about a scene were they’re not present, but other characters are, but you don’ thear anyone’s thought, just conversation?
Ada Palmer: There are many ways to do that.
Jo Walton: Everything is from some POV
Ada Palmer: The phone conversation example I used is one. Having there be a transcript or in-world document is another. Switching to another POV is another. In general if you want to see a scene but your narrator isn’t there, you want to avoid just narrating it in generic 3rd person because your readers will wonder who is narrating if they’re used to there being a narrator. So you either have to make the voice very different to make clear that it’s absolutely not just another character narrating now, or you need to do something else to signal the difference, like having it be a transcript, or an overheard phone call, or in a different tense, or something clear. Just leaving your narrator w/o changing the narration will be awkward and confusing.
Jo Walton: It’s a really good exercise to filter everything through one POV and then rewrite the same scene through another — not for final story, but to learn. And doing unlikely people is also a good way to learn a lot about how to write. And doing 1st and then 3rd on the same scene.
Question: Do we have to know and care a lot about a new narrator before we switch to them? I feel like they have to be really interesting to make the reader willing.
Ada Palmer: Sometimes, but switching into the POV of someone who’s hidden a lot about themselves can be an amazing moment to suddenly learn the truth about them. What if in our Romeo & Juliet, after Mercutio dies, we suddenly switch into the POV of the Prince of Verona, who’s basically been doing nothing throughout the play and about whom we know very little, but suddenly it turns out he’s actually aware of what’s happening and scheming to get Romeo & Juliet to kill themselves because he wants the Montague and Capulet houses to die out so he can confiscate their goods! Shock! Unexpected! Or what if we instead switch into that random servant who accompanies Romeo to the tomb, whom we know nothing about, but he suddenly turns out to be a prophet who’s used alchemy to predict what’s going to happen and is here to try to prevent the double-suicide, and then we see the tragic fascinating story of what intervenes to make him still fail to save them even though he knows what’s going to happen! Both of those suddenly make an unexplored character’s very unexploredness into something fascinating
Ada Palmer: Another thing that can be powerful is paying attention to when your reader will be excited to get to particular POVs. Sort-of like being excited to see who’ll write the next Mercutio chapter after he dies, you can, for example, set up a pattern where the reader realizes a certain POV must be coming and gets excited
Jo Walton: Martin does that with Jaime Lannister. And with the long absence of Theon. You can see the pattern of antici…..pation
Ada Palmer: Let’s imagine a story where our main POV is the captain of a prison who, and the pattern is that we always have two chapters from the captain’s POV, and then we see one from the POV of one of the prisoners in the cells. And the cells are numbered and we’re counting down from cell 8, then 7, then 6, visiting each prisoner only once. But as we go we keep hearing/learning strange things about the prisoner in cell 2: that prisoner is given strange food, everyone’s scared, everyone just calls it “Number Two” and never uses a name or even a pronoun, we learn it’s been there a looong time, we get into great suspense and excitement as we get toward cell 2. But wait… what’s in cell 1 which we haven’t heard anything about and will come after the super-mega-foreshadowed cell 2? We get meta-suspense. Now, you can achieve most of that suspense w/o switching POV if, instead of being in each prisoner’s head, we instead simply had every third chapter be the captain interacting with that particular prisoner. Doing it while staying in the captain’s head is easier for the reader. But on the other hand switching into their heads could gain a lot. It depends on which the story needs more – smoothness staying with one narrator, or the extra breadth and complexity of getting all those POVs, which would let us plunge quickly into their opinions, experience, backgrounds, knowledge, telling a lot of new info more quickly than the same POV could. Those POVs could let us quickly explore a big giant detailed world build in a way one POV couldn’t, or they could be too complex and more than the story will really make have payoff. And in the sequence, he buildup needs to be worth-it. Whatever is in cell 1 has to stand up even after whatever was in cell 2.
Here the discussion wound down, though of course one could discuss POV forever!!!! But for the meantime here is a short list of works with interesting uses of POV that are useful to read and think about as you want to learn more about unusual or powerful ways to do POV:
Sumner Locke Elliott, The Man Who Got Away
Gene Wolfe, Soldier of the Mist – the narration is a scroll
Gene Wolfe, The Shadow of the Torturer – how complex a first-person narrator can be
Daniel Keyes, “Flowers for Algernon”
Alfred Bester, “Fondly Fahrenheit” (and a lot of other Bester too!)
Agatha Christie, Endless Night
Melissa Scott, The Kindly Ones– we never learn the 1st person narrator’s gender, in the 80s that was a novelty, also it’s really great immersive spaceship and planet SF
Roger Zelazny, A Night in the Lonesome October — dog POV
Ted Chiang’s “Story of Your Life” (2nd person direct address)
Choose Your Own Adventure Books (true 2nd person, unlike 2nd person direct address)
For Jo Walton doing interesting POV things see the short story on Tor.com “Escape to Other Worlds With Science Fiction,” and her forthcoming Or What You Will. (Her Small Change books also have an unusual example of alternating first/third)
And for Ada Palmer, see Terra Ignota, which does a LOT of unusual things with POV, and keeps doing more volume by volume.
The discussion also, at the end, included a discussion of examples how strange POVs can get, including the Chintsubu boys’ love manga by Nase Yamato, which adopts the POV of talking penises, and while I don’t particularly recommend as a literary work but, like Asumiko Nakamura’s short story I Am a Piano, is a fascinating example of how outside-the-box a POV can be.
It was my honor last night to, for the second time, present the John W. Campbell Award for Best New Writer, on behalf of last year’s brilliant winner Rebecca Roanhorse who couldn’t attend this year. Many have asked me to post the text of the speech, which was unluckily disrupted by automated voice recognition subtitles on the screen behind me which garbled the words far too hilariously for the audience to keep from laughing, though it was very clear everyone tried hard. It’s funny, I feel that in any other venue on Earth uproarious laughter during a very serious speech would have felt mortifying, but this community is both so welcoming & respectful and so fun & comfortable that, between the sea of faces scrunched by clear attempts to fight the laughter, and audible shouts of “We love you, Ada!” from the front rows during the laughter roars, it really just felt hilarious and warm. I hope I find a video somewhere so I too can enjoy such disasters as “dog mechanism” for “dogmatic” and “Bored of the Rings and Cream of Thrown” for Lord of the Rings & Game of Thrones. More seriously, it was a great honor to speak again at this year’s Worldcon, and I couldn’t be more proud of Jeannette Ng‘s courageous acceptance speech, bringing attention to the crisis and violence happening right now in her home city of Hong Kong, and to the great responsibility we in the science fiction and fantasy community have to make sure that the theme of empire–which has numerous positive depictions in genre literature from space empires to the returns of kings–does not end up celebrating the dangerous, colonial, and autocratic faces of empire, and that as we explore empire in our work (including in my own work) we do so in ways which examine empire’s problems and advance versions of empire which reverse or rehabilitate it, and which affirm the greater values of free-determination, autonomy, and human dignity.
Separately (though very much in that spirit), I’m working on some Charity Fundraising for Refugees, and will put details in this post below the speech, so please look if you’re interested in helping!
JOHN W. CAMPBELL AWARD PRESENTATION SPEECH 2019 (Dublin Worldcon)
First awarded in 1973, the John W. Campbell Award welcomes outstanding new writers to the genre fiction community. Last year’s winner Rebecca Roanhorse could not attend to present the award this year, but as she and I corresponded about what we wanted, we focused first on our excitement that once again we are welcoming a brilliantly diverse group of new voices with perspectives long under-represented in literature, and second on how many of the finalists for this award administered by the World Science Fiction Convention are authors of fantasy.
In 1620 Francis Bacon published the first proposal for the formation of a community of scientists dedicated to exchanging ideas and collaborating to seek knowledge and technology that would improve the human condition generation by generation. When defining this new category of scholar—the scientist—Bacon described not mere encyclopedist compiling unsorted facts as ants pile sand in their anthills, nor dogmatists who weave elaborate spider webs of theory out of the stuff of their own minds untethered to reality. Rather the scientists like honeybees would gather among the fruits of nature and, processing those fruits through the organs of their own being, produce something good and useful for the world. In this first and richest definition of science, authors of genre fiction are all scientists, all gathering among the fruits of culture, literature, history, scientific discovery and social discovery, and processing those treasures through the diverse and diversifying organs of our being to create worlds, unrealities, which, whether fantasy realms, or eerie near futures, worlds of cosmic horror or alternate history, are good and useful for this world which we make wiser by constantly comparing it to others, worse, better, different, alien, or eerily familiar. In historical fiction, alternate history, and fantastic worlds we genre authors and readers test thousands of governments every year, far more than this little blue planet could ever give us room to try, and by so testing them we yield questions, and warnings, and broadened possibilities which enrich scientific understanding, political science, real government policy, and the palette of possible futures that we believe could come about. There are youth activists striking around the globe today inspired in no small way by the heady optimism of dystopian revolutions or the fantasy overthrow of enchanter tyrants, narratives which celebrate our capacity to resist tyrannies by depicting the victory of action over regimes of terror far darker than any lived reality. Sales of Orwell’s 1984 which have skyrocketed worldwide over the past three years as we face new censorship and surveillance technologies, threats to the global free press, and authoritarian resurgence, but so have sales of Tolkien and of Game of Thrones in fact the whole world of genre fiction whose readers debate good government and when a war is just, and by so doing yield, as Bacon promised, those honey treasures that empower our teamwork to keep building a better and more examined human future. So I say, in the best way, we here at the World Science Fiction Convention are all scientists, part of a community doing what Bacon envisioned, studying our world, testing our hypotheses in innumerable simulations, and thereby giving every generation new tools of empowerment, analysis, action, and global progress social as well as technological. In that spirit please join me in honoring six outstanding scientists of worlds beyond our own: Katherine Arden, S.A. Chakraborty, R.F. Kuang, Jeannette Ng, Vina Jie-Min Prasad, and Rivers Solomon.
So, on that note, I’m working with a number of fellow fantasy and science fiction authors, editors, readers & fans to organize two different things su
One project is “WORLDS OF WELCOME” an effort to raise money for refugees, in this case for RAICES to help those in the camps on the US side of the US-Mexico Border, by organize an online auction of items donated by the F&SF community: signed books, craft items, fanworks & merchandise, custom fiction, editors willing to give query critiques, or members of underrepresented groups willing to give sensitivity reads to in-progress fiction. We hope the auction will run in October. We are currently soliciting donations of items for the auction, and a few more people to help organize it. If interested, please email us at WorldsOfWelcome@gmail.com — more information to come in the next months.
Separately, if you’d like to give immediate and ongoing aid to refugees, I’ve just helped launch a Patreon for Cittadini Del Mondo (Citizens of the World), a brilliant microcharity in Rome which cares for the needs of the thousands of refugees in Italy who slip through the international aid system. This heroic and incredible 12-person team provides thousands of refugees with medical services, legal aid, language courses, prenatal care, and runs an Intercultural Library containing literature and kids’ books in refugees’ native languages, to help people who have lost everything reconnect with their languages and cultures. With such a tiny group even a few dollars a month can make a huge difference in their ability to stock the clinic for winter, buy new socks and shoes for new arrivals, or provide a warm and welcoming place where kids can explore the world of imagination as we all treasure. Please check it out (and pass it on!)