FreshRSS

🔒
❌ About FreshRSS
There are new available articles, click to refresh the page.
Before yesterdayYour RSS feeds

A Case for AI Wellbeing (guest post)

“There are good reasons to think that some AIs today have wellbeing.”

In this guest post, Simon Goldstein (Dianoia Institute, Australian Catholic University) and Cameron Domenico Kirk-Giannini (Rutgers University – Newark, Center for AI Safety) argue that some existing artificial intelligences have a kind of moral significance because they’re beings for whom things can go well or badly.

This is the sixth in a series of weekly guest posts by different authors at Daily Nous this summer.

[Posts in the summer guest series will remain pinned to the top of the page for the week in which they’re published.]

 


A Case for AI Wellbeing
by Simon Goldstein and Cameron Domenico Kirk-Giannini 

We recognize one another as beings for whom things can go well or badly, beings whose lives may be better or worse according to the balance they strike between goods and ills, pleasures and pains, desires satisfied and frustrated. In our more broad-minded moments, we are willing to extend the concept of wellbeing also to nonhuman animals, treating them as independent bearers of value whose interests we must consider in moral deliberation. But most people, and perhaps even most philosophers, would reject the idea that fully artificial systems, designed by human engineers and realized on computer hardware, may similarly demand our moral consideration. Even many who accept the possibility that humanoid androids in the distant future will have wellbeing would resist the idea that the same could be true of today’s AI.

Perhaps because the creation of artificial systems with wellbeing is assumed to be so far off, little philosophical attention has been devoted to the question of what such systems would have to be like. In this post, we suggest a surprising answer to this question: when one integrates leading theories of mental states like belief, desire, and pleasure with leading theories of wellbeing, one is confronted with the possibility that the technology already exists to create AI systems with wellbeing. We argue that a new type of AI—the artificial language agent—has wellbeing. Artificial language agents augment large language models with the capacity to observe, remember, and form plans. We also argue that the possession of wellbeing by language agents does not depend on them being phenomenally conscious. Far from a topic for speculative fiction or future generations of philosophers, then, AI wellbeing is a pressing issue. This post is a condensed version of our argument. To read the full version, click here.

1. Artificial Language Agents

Artificial language agents (or simply language agents) are our focus because they support the strongest case for wellbeing among existing AIs. Language agents are built by wrapping a large language model (LLM) in an architecture that supports long-term planning. An LLM is an artificial neural network designed to generate coherent text responses to text inputs (ChatGPT is the most famous example). The LLM at the center of a language agent is its cerebral cortex: it performs most of the agent’s cognitive processing tasks. In addition to the LLM, however, a language agent has files that record its beliefs, desires, plans, and observations as sentences of natural language. The language agent uses the LLM to form a plan of action based on its beliefs and desires. In this way, the cognitive architecture of language agents is familiar from folk psychology.

For concreteness, consider the language agents built this year by a team of researchers at Stanford and Google. Like video game characters, these agents live in a simulated world called ‘Smallville’, which they can observe and interact with via natural-language descriptions of what they see and how they act. Each agent is given a text backstory that defines their occupation, relationships, and goals. As they navigate the world of Smallville, their experiences are added to a “memory stream” in the form of natural language statements. Because each agent’s memory stream is long, agents use their LLM to assign importance scores to their memories and to determine which memories are relevant to their situation. Then the agents reflect: they query the LLM to make important generalizations about their values, relationships, and other higher-level representations. Finally, they plan: They feed important memories from each day into the LLM, which generates a plan for the next day. Plans determine how an agent acts, but can be revised on the fly on the basis of events that occur during the day. In this way, language agents engage in practical reasoning, deciding how to promote their goals given their beliefs.

2. Belief and Desire

The conclusion that language agents have beliefs and desires follows from many of the most popular theories of belief and desire, including versions of dispositionalism, interpretationism, and representationalism.

According to the dispositionalist, to believe or desire that something is the case is to possess a suitable suite of dispositions. According to ‘narrow’ dispositionalism, the relevant dispositions are behavioral and cognitive; ‘wide’ dispositionalism also includes dispositions to have phenomenal experiences. While wide dispositionalism is coherent, we set it aside here because it has been defended less frequently than narrow dispositionalism.

Consider belief. In the case of language agents, the best candidate for the state of believing a proposition is the state of having a sentence expressing that proposition written in the memory stream. This state is accompanied by the right kinds of verbal and nonverbal behavioral dispositions to count as a belief, and, given the functional architecture of the system, also the right kinds of cognitive dispositions. Similar remarks apply to desire.

According to the interpretationist, what it is to have beliefs and desires is for one’s behavior (verbal and nonverbal) to be interpretable as rational given those beliefs and desires. There is no in-principle problem with applying the methods of radical interpretation to the linguistic and nonlinguistic behavior of a language agent to determine what it believes and desires.

According to the representationalist, to believe or desire something is to have a mental representation with the appropriate causal powers and content. Representationalism deserves special emphasis because “probably the majority of contemporary philosophers of mind adhere to some form of representationalism about belief” (Schwitzgebel).

It is hard to resist the conclusion that language agents have beliefs and desires in the representationalist sense. The Stanford language agents, for example, have memories which consist of text files containing natural language sentences specifying what they have observed and what they want. Natural language sentences clearly have content, and the fact that a given sentence is in a given agent’s memory plays a direct causal role in shaping its behavior.

Many representationalists have argued that human cognition should be explained by positing a “language of thought.” Language agents also have a language of thought: their language of thought is English!

An example may help to show the force of our arguments. One of Stanford’s language agents had an initial description that included the goal of planning a Valentine’s Day party. This goal was entered into the agent’s planning module. The result was a complex pattern of behavior. The agent met with every resident of Smallville, inviting them to the party and asking them what kinds of activities they would like to include. The feedback was incorporated into the party planning.

To us, this kind of complex behavior clearly manifests a disposition to act in ways that would tend to bring about a successful Valentine’s Day party given the agent’s observations about the world around it. Moreover, the agent is ripe for interpretationist analysis. Their behavior would be very difficult to explain without referencing the goal of organizing a Valentine’s Day party. And, of course, the agent’s initial description contained a sentence with the content that its goal was to plan a Valentine’s Day party. So, whether one is attracted to narrow dispositionalism, interpretationism, or representationalism, we believe the kind of complex behavior exhibited by language agents is best explained by crediting them with beliefs and desires.

3. Wellbeing

What makes someone’s life go better or worse for them? There are three main theories of wellbeing: hedonism, desire satisfactionism, and objective list theories. According to hedonism, an individual’s wellbeing is determined by the balance of pleasure and pain in their life. According to desire satisfactionism, an individual’s wellbeing is determined by the extent to which their desires are satisfied. According to objective list theories, an individual’s wellbeing is determined by their possession of objectively valuable things, including knowledge, reasoning, and achievements.

On hedonism, to determine whether language agents have wellbeing, we must determine whether they feel pleasure and pain. This in turn depends on the nature of pleasure and pain.

There are two main theories of pleasure and pain. According to phenomenal theories, pleasures are phenomenal states. For example, one phenomenal theory of pleasure is the distinctive feeling theory. The distinctive feeling theory says that there is a particular phenomenal experience of pleasure that is common to all pleasant activities. We see little reason why language agents would have representations with this kind of structure. So if this theory of pleasure were correct, then hedonism would predict that language agents do not have wellbeing.

The main alternative to phenomenal theories of pleasure is attitudinal theories. In fact, most philosophers of wellbeing favor attitudinal over phenomenal theories of pleasure (Bramble). One attitudinal theory is the desire-based theory: experiences are pleasant when they are desired. This kind of theory is motivated by the heterogeneity of pleasure: a wide range of disparate experiences are pleasant, including the warm relaxation of soaking in a hot tub, the taste of chocolate cake, and the challenge of completing a crossword. While differing in intrinsic character, all of these experiences are pleasant when desired.

If pleasures are desired experiences and AIs can have desires, it follows that AIs can have pleasure if they can have experiences. In this context, we are attracted to a proposal defended by Schroeder: an agent has a pleasurable experience when they perceive the world being a certain way, and they desire the world to be that way. Even if language agents don’t presently have such representations, it would be possible to modify their architecture to incorporate them. So some versions of hedonism are compatible with the idea that language agents could have wellbeing.

We turn now from hedonism to desire satisfaction theories. According to desire satisfaction theories, your life goes well to the extent that your desires are satisfied. We’ve already argued that language agents have desires. If that argument is right, then desire satisfaction theories seem to imply that language agents can have wellbeing.

According to objective list theories of wellbeing, a person’s life is good for them to the extent that it instantiates objective goods. Common components of objective list theories include friendship, art, reasoning, knowledge, and achievements. For reasons of space, we won’t address these theories in detail here. But the general moral is that once you admit that language agents possess beliefs and desires, it is hard not to grant them access to a wide range of activities that make for an objectively good life. Achievements, knowledge, artistic practices, and friendship are all caught up in the process of making plans on the basis of beliefs and desires.

Generalizing, if language agents have beliefs and desires, then most leading theories of wellbeing suggest that their desires matter morally.

4. Is Consciousness Necessary for Wellbeing?

We’ve argued that language agents have wellbeing. But there is a simple challenge to this proposal. First, language agents may not be phenomenally conscious — there may be nothing it feels like to be a language agent. Second, some philosophers accept:

The Consciousness Requirement. Phenomenal consciousness is necessary for having wellbeing.

The Consciousness Requirement might be motivated in either of two ways: First, it might be held that every welfare good itself requires phenomenal consciousness (this view is known as experientialism). Second, it might be held that though some welfare goods can be possessed by beings that lack phenomenal consciousness, such beings are nevertheless precluded from having wellbeing because phenomenal consciousness is necessary to have wellbeing.

We are not convinced. First, we consider it a live question whether language agents are or are not phenomenally conscious (see Chalmers for recent discussion). Much depends on what phenomenal consciousness is. Some theories of consciousness appeal to higher-order representations: you are conscious if you have appropriately structured mental states that represent other mental states. Sufficiently sophisticated language agents, and potentially many other artificial systems, will satisfy this condition. Other theories of consciousness appeal to a ‘global workspace’: an agent’s mental state is conscious when it is broadcast to a range of that agent’s cognitive systems. According to this theory, language agents will be conscious once their architecture includes representations that are broadcast widely. The memory stream of Stanford’s language agents may already satisfy this condition. If language agents are conscious, then the Consciousness Requirement does not pose a problem for our claim that they have wellbeing.

Second, we are not convinced of the Consciousness Requirement itself. We deny that consciousness is required for possessing every welfare good, and we deny that consciousness is required in order to have wellbeing.

With respect to the first issue, we build on a recent argument by Bradford, who notes that experientialism about welfare is rejected by the majority of philosophers of welfare. Cases of deception and hallucination suggest that your life can be very bad even when your experiences are very good. This has motivated desire satisfaction and objective list theories of wellbeing, which often allow that some welfare goods can be possessed independently of one’s experience. For example, desires can be satisfied, beliefs can be knowledge, and achievements can be achieved, all independently of experience.

Rejecting experientialism puts pressure on the Consciousness Requirement. If wellbeing can increase or decrease without conscious experience, why would consciousness be required for having wellbeing? After all, it seems natural to hold that the theory of wellbeing and the theory of welfare goods should fit together in a straightforward way:

Simple Connection. An individual can have wellbeing just in case it is capable of possessing one or more welfare goods.

Rejecting experientialism but maintaining Simple Connection yields a view incompatible with the Consciousness Requirement: the falsity of experientialism entails that some welfare goods can be possessed by non-conscious beings, and Simple Connection guarantees that such non-conscious beings will have wellbeing.

Advocates of the Consciousness Requirement who are not experientialists must reject Simple Connection and hold that consciousness is required to have wellbeing even if it is not required to possess particular welfare goods. We offer two arguments against this view.

First, leading theories of the nature of consciousness are implausible candidates for necessary conditions on wellbeing. For example, it is implausible that higher-order representations are required for wellbeing. Imagine an agent who has first order beliefs and desires, but does not have higher order representations. Why should this kind of agent not have wellbeing? Suppose that desire satisfaction contributes to wellbeing. Granted, since they don’t represent their beliefs and desires, they won’t themselves have opinions about whether their desires are satisfied. But the desires still are satisfied. Or consider global workspace theories of consciousness. Why should an agent’s degree of cognitive integration be relevant to whether their life can go better or worse?

Second, we think we can construct chains of cases where adding the relevant bit of consciousness would make no difference to wellbeing. Imagine an agent with the body and dispositional profile of an ordinary human being, but who is a ‘phenomenal zombie’ without any phenomenal experiences. Whether or not its desires are satisfied or its life instantiates various objective goods, defenders of the Consciousness Requirement must deny that this agent has wellbeing. But now imagine that this agent has a single persistent phenomenal experience of a homogenous white visual field. Adding consciousness to the phenomenal zombie has no intuitive effect on wellbeing: if its satisfied desires, achievements, and so forth did not contribute to its wellbeing before, the homogenous white field should make no difference. Nor is it enough for the consciousness to itself be something valuable: imagine that the phenomenal zombie always has a persistent phenomenal experience of mild pleasure. To our judgment, this should equally have no effect on whether the agent’s satisfied desires or possession of objective goods contribute to its wellbeing. Sprinkling pleasure on top of the functional profile of a human does not make the crucial difference. These observations suggest that whatever consciousness adds to wellbeing must be connected to individual welfare goods, rather than some extra condition required for wellbeing: rejecting Simple Connection is not well motivated. Thus the friend of the Consciousness Requirement cannot easily avoid the problems with experientialism by falling back on the idea that consciousness is a necessary condition for having wellbeing.

We’ve argued that there are good reasons to think that some AIs today have wellbeing. But our arguments are not conclusive. Still, we think that in the face of these arguments, it is reasonable to assign significant probability to the thesis that some AIs have wellbeing.

In the face of this moral uncertainty, how should we act? We propose extreme caution. Wellbeing is one of the core concepts of ethical theory. If AIs can have wellbeing, then they can be harmed, and this harm matters morally. Even if the probability that AIs have wellbeing is relatively low, we must think carefully before lowering the wellbeing of an AI without producing an offsetting benefit.


[Image made with DALL-E]

Some related posts:
Philosophers on GPT-3
Philosophers on Next-Generation Large Language Models
GPT-4 and the Question of Intelligence
We’re Not Ready for the AI on the Horizon, But People Are Trying
Researchers Call for More Work on Consciousness
Dennett on AI: We Must Protect Ourselves Against ‘Counterfeit People’
Philosophy, AI, and Society Listserv
Talking Philosophy with Chat-GPT

The post A Case for AI Wellbeing (guest post) first appeared on Daily Nous.

The Fourth Branch (guest post)

“We shouldn’t attempt to fit ‘outreach’ or ‘engagement’ into one of the existing three categories [of research, teaching, or service]. It doesn’t fit neatly into those categories. And, more importantly, all of us should be doing it as part of our jobs, not just a few of us. We are in an all-hands-on-deck situation.”

In the following guest post, Alex Guerrero, professor of philosophy at Rutgers University, argues that we should count engagement or outreach as a distinct component of the job of professor.

This is the fifth in a series of weekly guest posts by different authors at Daily Nous this summer.

[Posts in the summer guest series will remain pinned to the top of the page for the week in which they’re published.]

 


[“The Beautiful Walk” by René Magritte]

The Fourth Branch
by Alex Guerrero

The job of a philosopher in America has been defined in relatively specific terms. These are the terms: research, teaching, and service. How much of one’s life one is expected to contribute to each, the precise percentage and weight, varies from job to job. In some, research is all that matters. In others, teaching is the thing. More rarely, a philosopher wanders deep into administration, lost in a forest of service, and emerges as a thoughtful if pesky manager and builder of things with varying degrees of value. For most of us, we are required to do all of these, more or less well, with more or less joy.

Socrates and Kongzi both might have, for once, been at a loss for words if forced to say whether they were doing research, teaching, or service. But we have contracts, faculty handbooks, promotion guidelines, and legalistic specifications of what we are required to do. None of us are employed as philosophers. We are employed as professors (lecturers, instructors). For the most part, our job is like that of other humanities professors, and unlike that of research scientists and others in STEM fields supported by grants, who do things like “buy out” of teaching and oversee research labs. For us, the job is the core three: research, teaching, and service.

Teaching is the very official role where we are in front of enrolled students, in a class that counts for credit, presenting material, devising and evaluating assignments, and figuring out ways for students to learn what we take to be important about the subject. Our teaching responsibilities are almost always defined by the number of courses taught per academic year, spread out over semesters or quarters, often specified in terms of level and number.

Research is publication. We might wish it were instead about ideas, figuring new things out, moving knowledge forward—regardless of whether that results in articles in peer-reviewed journals or books in academic presses. But we know better. You need 8 publications for tenure in B+ or better peer-reviewed journals. Or 1.5 per year on the tenure clock. Or whatever. What “counts” for research, and how much it counts, is usually clear, even if promotion requirements are rarely specified in detail. Post-tenure, although research productivity factors into further promotions and merit raises, the sense in which it is “required” becomes considerably murkier.

Service is what is required to keep up the pretense that universities and colleges are run by the faculty, rather than by a distinct managerial class. We “serve” on hiring committees, as undergraduate and graduate program administrators, on curriculum committees, on admissions committees, in organizing colloquia and events, in putting people forward and evaluating them for tenure and promotion, as chairs and vice-chairs, deans and deanlets, and on committees for every domain of human complaint and frustration—these are the core of internal, university service. We serve our departments, schools, colleges, and universities.  Many of us expand beyond this to engage in service to the “profession”—running academic journals, professional associations, planning workshops and conferences, creating and supporting broad mentoring and inclusion efforts, and so on.

Most of us were just dropped into this world of research, teaching, and service. I’ve never seen anyone try to justify why these are the three parts of the job. As many have noted, they don’t fit together all that naturally. What makes one an amazing teacher might have nothing do with what makes for a groundbreaking researcher. And we almost expect that whatever makes us good at those things will make us inept at, or at least impatient with, most kinds of “service.” Our graduate training programs do very little to train us to teach, or to administer anything or manage anyone.

The explanation for the three branches seems to be a historically contingent one, with the modern college or university coming to exist with a dual-purpose mission of educating students (teaching) and advancing knowledge (research), and service comes along as a third thing essential to preserving various values relating to those first two. Specifically, the values of academic, expert peer review (for admissions, hiring, publication, research evaluation, promotion) and academic, expert curriculum and course design and implementation.

There are, of course, many ways of rethinking this basic three branch setup, and many institutions that have already reconfigured things so that people are hired into jobs where they will do just one or two of those three things. I don’t want to wade more deeply into those waters. Instead, I want to suggest that, given the pressures confronting colleges and universities, sustaining those institutions in their core dual-purpose mission of educating students and advancing knowledge requires introducing a fourth branch. I’m not sure what name for that fourth branch is best. Here are my two favorite candidates: outreach and engagement.

The basic argument for the fourth branch is simple.

Colleges and universities are supported (1) by the general public, through government funding; (2) by students and their families, through tuition and fees; and (3) by rich people, through donations. What education and what knowledge will be pursued in colleges and universities is not set in stone; it is, rather, a function of what those three groups want and demand. If we want philosophy to be part of the education and part of the knowledge that is pursued in the years to come, we need people in those three groups to want and demand philosophy. And for people in those three groups to want and demand philosophy, we need to reach out to them, engage them, make them aware of what philosophy is and why it is wonderful and valuable. Given what philosophy is, and given our contemporary situation, that task is monumental, and must be undertaken at many different levels, in many ways. No small number of us can do it on our own. Therefore, it should be a part of all of our jobs—quite literally—to do this work.

(We can substitute in almost any humanistic field for ‘philosophy’ in the above argument, with similar implications for a needed fourth branch for the rest of the humanistic fields.)

The basics of this ‘demand side’ story are familiar. Many students and parents think of college and university in terms of relatively short term, career ambitions—how will this major, this degree, this school help me get a job. Many states and nations have begun taking a similar attitude toward all higher education, thinking in terms of contribution to economic productivity. Little actual empirical investigation is involved in deciding that humanistic fields, and fields such as philosophy, in particular, don’t do well by this score. But that is a common public perception. And it is understandable. It seems plausible that a degree in business, health professions, computer science, engineering, or biomedical sciences (the five largest growth majors over the past ten years) would be a more direct route to a job than a degree in English, history, or philosophy (three of the majors that lost the most majors over the past ten years).

There are other factors that affect demand. STEM fields are intimately connected to industries and occupations outside of the academy, so that one might have encountered a person with background in that field. We no longer have the same kind of elite, quasi-aristocratic veneration of those humanistic things that every “educated” person must know. STEM spends more time in the news, as breakthroughs in tech, science, and computing get regular reporting and discussion, and result in products in our pockets and dreams on our screens. In the United States, quality exposure to literature, philosophy, and even history is rare prior to college or university, with many students not having encountered any philosophy before college, and having encountered history only as rote memorization and literature only as forced reading.

Behind the idea that colleges and universities have a dual-purpose mission of educating students and advancing knowledge is a mostly implicit idea about what education and knowledge is valuable and why. For those of us working as professors in a subject domain, we almost certainly see the domain as valuable, and can offer many different compelling reasons why it is valuable, pointing both to intrinsic or final value of the subject, and to instrumental benefits of education and knowledge in the domain. We can go on and on about personal transformation, what it means to be human, learning to think critically about what is important, becoming a democratic and cosmopolitan citizen of the world, and so forth. But for those not already in philosophy, where are they supposed to hear the good news?

It is easy to vilify the administrator class, focused on the bottom line—bottoms in seats, donations to “development” offices, legislative support for public institutions. But it is not their fault that higher education is funded as it is. It is not their fault that we, the professoriate, don’t have unbridled power to force people to study those subjects that we see as most valuable. Professors like the idea of being in control: we get to design the curriculum, plan the syllabus, pick the readings, develop the research projects, evaluate the work, decide what should be published, determine who should be admitted, hired, promoted, esteemed. This all seems right and good to us, given our knowledge and expertise. But that is not the system we have with respect to the very basic facts of higher education in most contemporary political environments. We might wish for administrators who could hold the line, fight the battle for us, make the case for the importance of philosophy and the humanities. And some of them can and do. But many operate in incredibly difficult economic and political environments. They can’t change the basic facts about whose demand matters. And they can’t even do much to affect the substance of what is in demand. They need help. And—at least given our knowledge—we are well positioned to provide it.

Many of us are already involved in various efforts to broaden exposure to and engagement with philosophy. Most involved in this work aren’t doing it thinking to “increase demand” for philosophy, but it plausibly has that effect. There are obviously central enterprises: exposing children and adolescents to philosophy and serious humanities in K-12 education, for example, something that many are already doing. Writing “public facing” philosophy that appears in newspapers, broad circulation prestige venues, trade books, and so on. Creating online philosophy courses and videos and other broad access materials like podcasts. There are also more local, more intimate efforts: organizing a public philosophy week at a public library, running a philosophy club or ethics bowl team at the local high school, organizing community book groups and “meetups” to discuss philosophy, running “ask a philosopher” booths at the train station, farmers’ market, or mall. These activities bring philosophy to people outside of the academy and bring people into philosophy, giving them entry points and a better sense of what the subject is and why it is of value. They also are a lot of fun. And a ton of work to do well. And, for the most part, they are treated as outside of one’s job, falling outside of the big three: research, teaching, and service.

For those involved in this work, a common argument is that it should be included under research, teaching, or service—depending on the details. In a few places, “engagement” work is already included as part of one’s official job requirements. In more places, efforts are being made to think about how to include this work under research, or teaching, or service. I’ve spent the past year on a committee at Rutgers on a “Task Force for Community Engaged/ Publicly Engaged Scholarship,” focused on questions of definition (what is it, what counts) and evaluation (what metrics are available and appropriate, what standards should be used) of this kind of research. In serving on this committee, I’ve learned in detail about dozens of similar efforts at other institutions. In almost every case, the discussion is focused on how to credit the work being done by a small percentage of professors who are doing some “community engagement” or “public facing” work. Can it count instead of other, more traditional academic research? Can it count for teaching or service?

I want to suggest that, for those of us in the humanities, we should understand the importance of this engagement work to our core dual mission of educating students and advancing knowledge in our fields, and we should stop trying to shoehorn it into the traditional three categories. In the same way that service is required to sustain certain valuable features of how education and research is conducted, engagement is required to sustain certain valuable features of what education and research is conducted. Professorial “service” is required due to internal, contingent features of running a college or university, as something instrumentally important to enabling high quality education of students and to advancing serious research and knowledge. In the modern political and economic context most of us are in now, professorial “outreach” or “public engagement” is required due to external, contingent features of running a college or university, as something instrumentally important to enabling high quality education of students and to advancing serious research and knowledge in all academic domains that are of genuine value, including philosophy. We shouldn’t attempt to fit “outreach” or “engagement” into one of the existing three categories. It doesn’t fit neatly into those categories. And, more importantly, all of us should be doing it as part of our jobs, not just a few of us. We are in an all-hands-on-deck situation. It’s not clear that even this huge shift would be enough to save the humanities, but it affords more hope than doing nothing and just praying for favorable shifts in the demand curves.

So, the proposal is this. Add “engagement” or “outreach” as a fourth component of the job of professor, along with research, teaching, and service. The exact percentages might vary, as they already do, across institutions. One model might be 35% teaching, 35% research, 20% service, and 10% outreach. Just as with service, there would be a variety of ways of satisfying the requirement. And it might be assessed over several years, rather just in one particular year.

In addition to the familiar forms above—creating and running K-12 programs and clubs, public facing writing, online courses and spaces for public philosophy discussion, podcasts, local events and courses and reading groups at libraries and other community venues—there might be other work that would count. Creating and publicizing materials about philosophy and its intrinsic and instrumental value, developing materials to connect philosophy to prominent issues of the moment, connecting philosophy to locally popular elements of colleges and universities (such as college sports), strategically lobbying local and state political officials to help explain the value of philosophy and to think about ways in philosophy might be relevant or useful given their political agenda, developing white papers and strategic plans to encourage businesses and industries to seek out philosophy graduates and build philosophy alumni networks, even working on philosophy-related fundraising (either through or, where permitted, outside of, the development office). We need to do work at many different levels, in many different venues and spaces. New Yorker articles and popular books are important, but insufficient to reach or affect the very broad, heterogeneous communities whose interest and demand is relevant.

Just as with service (and teaching and research, for that matter), we shouldn’t expect that everyone will be equally good at or equally drawn to every aspect of outreach. And, just as with service, much of our evaluation of the work will be somewhat less precise than our evaluation of teaching or research. Norms and guidelines will need to be developed and adjusted over time. There is also the question, as with teaching and service, of how to train people to do this work well. Much of the work being done by committees like the one I’ve been on can be of help in thinking through these issues. None strike me as insurmountable.

The basic hope is that, by requiring everyone to do some outreach and engagement work, we can get many more people involved, and have a correspondingly greater effect on the broad understanding of philosophy and its value, with a hoped-for uptick in interest, support, and demand.

There are other potential side-benefits to creating the fourth branch. One is that it might mean that the public face of philosophy will be much more complex and multifaceted, so that it isn’t entirely dominated by a few prominent people who are skilled at prestige publishing or personal branding or whatever we want to call Žižek’s skillset. Another is that, as most people who have done this work will attest, engagement and outreach work provides a kind of beneficial feedback, potentially improving the quality of one’s teaching, research, and service, as one steps back from those activities and considers and attempts to communicate about their value, and to share philosophy with people who have not encountered it before. A third is that it might enable and prepare philosophers and other humanists to push back more effectively against the tendency for humanistic and normative concerns to be overrun by the march of technological, scientific, and commercial “progress” and “innovation.” And it might help the broader public see, raise, and respond to these concerns for themselves, even without any university training in the field.

There might be a worry that focusing on demand and broad interest sets up a zero-sum competition. We might do better, but that will mean some other field, perhaps with a harder case to make, will do worse. Perhaps, although I think a broad push from humanists in this regard might make a considerable difference to public understanding and public perception, even about the basic role of colleges and universities and the value of attending those institutions. Much greater, more structurally supported and organized efforts from professors in this regard might help alter the discussion and push back against the view of higher education as just some kind of pre-vocational training for those who can afford it. Most optimistically, it might alter the public funding dynamics of college and university, reasserting the ideal of affordable liberal arts, humanistic higher education for all as part of an important public component of a genuinely democratic, flourishing society.

Earlier, I mentioned that administrators are often limited in the ways in which they can help us. Here is one: help the humanists help themselves, by building a fourth branch, focused on engagement and outreach, into our jobs. In some cases, we could begin doing this somewhat informally within departments, setting up service positions focused on outreach, and then treating that work as part of service. But, in my view, it will be better to build it in more structurally, with a broader requirement for everyone, and allowing a broader array of ways in which to contribute.

If we want philosophy to survive, we need people to understand what it is and why it should survive. We can’t just rest on our historical laurels or on “get off my lawn” arguments that simply insist that no serious university can exist without a philosophy department. We might be right. But we also might just end up surrounded by unserious universities.


Other posts on public philosophy, engagement, and outreach.

The post The Fourth Branch (guest post) first appeared on Daily Nous.

Some Remarks on Form in Philosophy (guest post)

“When my younger self complained angrily in the margins with scrawls of ‘where is this going?’, he missed the sights and insights that the journey itself provided.”

A kind of philosophical innocence is the subject of the following guest post by Bradford Skow, Laurance S. Rockefeller Professor of Philosophy at the Massachusetts Institute of Technology. Professor Skow works mainly in metaphysics, philosophy of science, and aesthetics, and has a blog/newsletter, Mostly Aesthetics, on Substack, which I highly recommend.

This is the fourth in a series of weekly guest posts by different authors at Daily Nous this summer.

[Posts in the summer guest series will remain pinned to the top of the page for the week in which they’re published.]

 


 

Art gallery displaying a painting of Adam and Eve, a painting of Freytag's Triangle, and a vandalized painting of Van Gogh's field of poppies, with the words "I will argue that P" spraypainted onto it with a stencil

Some Remarks on Form in Philosophy
by Bradford Skow

Herbert Morris read an essay titled “Lost Innocence” at the 1975 Oberlin Philosophy Colloquium. In it, he attempted an analysis of Adam and Eve’s loss of innocence, when Eve was tempted and they ate from the tree of knowledge of good and evil. The knowledge they acquired was not, obviously, perceptual knowledge, nor did they draw some new conclusion from their evidence. It is closer to say, Morris thought, that they acquired “a different way of feeling about what had been before them all along”: “[w]hereas one had earlier felt at ease, felt a kind of natural joy, one now is cursed with an absence of joy, perhaps more, with feeling anxious and bad.” But this itself is just a first approximation, and Morris steers us toward, and then away from, several more accounts of lost innocence, until, after a pause at his preferred one, he closes with a meditation on the nature of evil, and on the morally wise people who have “not been crushed by what they have confronted, but have emerged, in ways mysterious to behold, victorious, capable, despite and because of knowledge, of affirming rather than denying life.” Life-affirmation is not an ending that the essay’s opening paragraphs point you toward. Morris announces only that the essay was stimulated by reflection on the story of Adam and Eve, and he acknowledges early on the “serpentine nature” of his reasoning.

Every age has its favored philosophical templates. Readers of this venue are doubtless familiar, either as writers, or as readers or referees, with one in common use today:

I will argue that P.
Lo, notice that I am now arguing that P.
Reminder that Modus Ponens is valid!
In conclusion, I have argued that P.

On top of all this meta-signposting, not to be neglected is the obligatory here is the plan for this paper, placed at the end of the Introduction, and skipped over by every reader. The basic unit of philosophy is taken to be the argument with numbered premises and labeled conclusion, as the proof is the basic unit of mathematics. Of course books and papers in mathematics also have definitions and explanatory insertions, nor do works of philosophy consist only of arguments—not even Spinoza’s Ethics—but everything is built around them.

At sea between college and graduate school, I pursued a self-directed course of philosophical study, and focus on the argument was the straw keeping me afloat. When chance scudded me into Peter Unger’s paper “The Problem of the Many,” therefore, it seemed scandalous. “Although,” he wrote, “I think arguments are important in philosophy, my arguments here will be only the more assertive way for me to introduce the new problem, not the only way.” What an idea, that a philosophical goal could be achieved by other means.

I am just old enough to have had passed down to me, like myths of a lost age, other templates that were once more common. After his death, various collections of Wittgenstein’s notes and lectures were published, under titles like Remarks on Suchity-Such. A tradition began in which the basic unit of philosophy was—at least in presentation—the remark, sequences of which were presented in numbered paragraphs. On the surface this practice was more collegial and less aggressive, gentlemen (remember this was decades ago) lounging in tweed coats and exchanging thoughts between puffs of their pipes, but of course in reality a seemingly-anodyne “I would like to make a few remarks” was usually a preface to something devastating, or anyway intended to be so.

Because philosophy is so old, and because the ideas of the earliest philosophers are still alive today, it can seem that finding new philosophical positions, or new arguments for existing ones, requires a gold-medal performance, while those practicing other, newer disciplines languorously pick low-hanging fruit on their lunch breaks. I sometimes joke that, because of this, a career-making philosophical achievement can now consist in something as small as drawing a distinction. Or even in erasing one: I was once present when a famous philosopher complained, to the students in his graduate seminar, that one of his undergraduates could not grasp the distinction between numerical and qualitative identity, and I complained back that that student might just be a philosophical genius, seeing clearly that the alleged distinction Mr. Famous was trying to draw was not real. When Amos Tversky was a student, he chose psychology over philosophy in part because there were far fewer giants over whose shoulders one needed to see. However their eventual publications compare to Tversky’s Nobel-worthy life’s-work, I tell philosophy graduate students, they should take heart in the fact that he was a coward who could not muster the courage they have displayed.

Premise, premise, conclusion is the Freytag’s triangle of philosophy, and its de-throning in literary criticism may inspire. Jane Alison, in her book Meander, Spiral, Explode, explores a bunch of alternatives to the exposition, climax, resolution form, including the titular ones, which also makes good philosophical models. If Morris’s essay is a controlled and steady navigation by an experienced hand through a carefully surveyed territory, Arthur Danto is a great meanderer. In just the first few pages of The Transfiguration of the Commonplace he riffs on Euripidean and anti-Euripidean art, the imitation theory, and Sartre, and when my younger self complained angrily in the margins with scrawls of “where is this going?”, he missed the sights and insights that the journey itself provided.

The point is not, or not just, to have more fun, or to unleash unruly intellectual impulses that academic disciplines try to discipline. Form can be expressive in ways that matter aesthetically and even philosophically. Wittgenstein’s Tractatus sometimes achieves an austere oracular beauty that the philosophy-as-lab-report genre cannot equal and to which it cannot aspire. It is also, to some, offputtingly arrogant. To their temperament the Philosophical Investigations may be more congenial. “Bear in mind,” Kripke observed in Wittgenstein on Rules and Private Language,

that Philosophical Investigations is not a systematic philosophical work where conclusions, once definitively established, need not be reargued. Rather, the Investigations is written as a perpetual dialectic, where persisting worries, expressed by the voice of an imaginary interlocutor, are never definitively silenced… the same ground is covered repeatedly, from the point of view of various special cases and from different angles, with the hope that the entire process will help the reader see the problems rightly.

If for Herbert Morris the loss of moral innocence was “like the loss of peace of mind accounted for by acquiring anxiety,” the form of the Investigations—as it struck Kripke anyway, as it presented itself to him—is expressive of a loss of philosophical innocence, and the—possibly appropriate—gnawing philosophical anxiety and uncertainty that some of us cannot shake, but may hope not to be crushed by.


The post Some Remarks on Form in Philosophy (guest post) first appeared on Daily Nous.

The Rigor of Philosophy & the Complexity of the World (guest post)

“Analytic philosophy gradually substitutes an ersatz conception of formalized ‘rigor’ in the stead of the close examination of applicational complexity.”

In the following guest post, Mark Wilson, Distinguished Professor of Philosophy and the History and Philosophy of Science at the University of Pittsburgh, argues that a kind of rigor that helped philosophy serve a valuable role in scientific inquiry has, in a sense, gone wild, tempting philosophers to the fruitless task of trying to understand the world from the armchair.

This is the third in a series of weekly guest posts by different authors at Daily Nous this summer.


[Roy Lichtenstein, “Bull Profile Series”]

The Rigor of Philosophy & the Complexity of the World
by Mark Wilson

In the course of attempting to correlate some recent advances in effective modeling with venerable issues in the philosophy of science in a new book (Imitation of Rigor), I realized that under the banner of “formal metaphysics,” recent analytic philosophy has forgotten many of the motivational considerations that had originally propelled the movement forward. I have also found that like-minded colleagues have been similarly puzzled by this paradoxical developmental arc. The editor of Daily Nous has kindly invited me to sketch my own diagnosis of the factors responsible for this thematic amnesia, in the hopes that these musings might inspire alternative forms of reflective appraisal.

The Promise of Rigor

Let us return to beginnings. Although the late nineteenth century is often characterized as a staid period intellectually, it actually served as a cauldron of radical reconceptualization within science and mathematics, in which familiar subjects became strongly invigorated through the application of unexpected conceptual adjustments. These transformative innovations were often resisted by the dogmatic metaphysicians of the time on the grounds that the innovations allegedly violated sundry a priori strictures with respect to causation, “substance” and mathematical certainty. In defensive response, physicists and mathematicians eventually determined that they could placate the “howls of the Boeotians” (Gauss) if their novel proposals were accommodated within axiomatic frameworks able to fix precisely how their novel notions should be utilized. The unproblematic “implicit definability” provided within these axiomatic containers should then alleviate any a priori doubts with respect to the coherence of the novel conceptualizations. At the same time, these same scientists realized that explicit formulation within an axiomatic framework can also serve as an effective tool for ferreting out the subtle doctrinal transitions that were tacitly responsible for the substantive crises in rigor that had bedeviled the period.

Pursuant to both objectives, in 1894 the physicist Heinrich Hertz attempted to frame a sophisticated axiomatics to mend the disconnected applicational threads that he correctly identified as compromising the effectiveness of classical mechanics in his time. Unlike his logical positivist successors, Hertz did not dismiss terminologies like “force” and “cause” out of hand as corruptly “metaphysical,” but merely suggested that they represent otherwise useful vocabularies that “have accumulated around themselves more relations than can be completely reconciled with one another” (through these penetrating diagnostic insights, Hertz emerges as the central figure within my book). As long as “force” and “cause” remain encrusted with divergent proclivities of this unacknowledged character, methodological strictures naively founded upon the armchair “intuitions” that we immediately associate with these words are likely to discourage the application of more helpful forms of conceptual innovation through their comparative unfamiliarity.

There is no doubt that parallel developments within symbolic logic sharpened these initial axiomatic inclinations in vital ways that have significantly clarified a wide range of murky conceptual issues within both mathematics and physics. However, as frequently happens with an admired tool, the value of a proposed axiomatization depends entirely upon the skills and insights of the workers who employ it. A superficially formalized housing in itself guarantees nothing. Indeed, the annals of pseudo-science are profusely populated with self-proclaimed geniuses who fancy that they can easily “out-Newton Newton” simply by costuming their ill-considered proposals within the haberdashery of axiomatic presentation (cf., Martin Gardner’s delightful Fads and Fallacies in the Name of Science).

Inspired by Hertz and Hilbert, the logical empiricists subsequently decided that the inherent confusions of metaphysical thought could be eliminated once and for all by demanding that any acceptable parcel of scientific theorizing must eventually submit to “regimentation” (Quine’s term) within a first order logical framework, possibly supplemented with a few additional varieties of causal or modal appeal. As just noted, Hertz himself did not regard “force” as inherently “metaphysical” in this same manner, but simply that it comprised a potentially misleading source of intuitions to rely upon in attempting to augur the methodological requirements of an advancing science.

Theory T Syndrome

Over analytic philosophy’s subsequent career, these logical empiricist expectations with respect to axiomatic regimentation gradually solidified into an agglomeration of strictures upon acceptable conceptualization that have allowed philosophers to criticize rival points of view as “unscientific” through their failure to conform to favored patterns of explanatory regimentation. I have labelled these logistical predilections as the “Theory T syndrome” in other writings.

A canonical illustration is provided by the methodological gauntlet that Donald Davidson thrusts before his opponents in “Actions, Reasons and Causes”:

One way we can explain an event is by placing it in the context of its cause; cause and effect form the sort of pattern that explains the effect, in a sense of “explain” that we understand as well as any. If reason and action illustrate a different pattern of explanation, that pattern must be identified.

In my estimation, this passage supplies a classic illustration of Theory T-inspired certitude. In fact, a Hertz-like survey of mechanical practice reveals many natural applications of the term “cause” that fail to conform to Davidson’s methodological reprimands.

As a result, “regimented theory” presumptions of a confident “Theory T” character equip such critics with a formalist reentry ticket that allows armchair speculation to creep back into the philosophical arena with sparse attention to the real life complexities of effective concept employment. Once again we witness the same dependencies upon a limited range of potentially misleading examples (“Johnny’s baseball caused the window to break”), rather than vigorous attempts to unravel the entangled puzzlements that naturally attach to a confusing word like “cause,” occasioned by the same developmental processes that make “force” gather a good deal of moss as it rolls forward through its various modes of practical application. Imitation of Rigor attempts to identify some of the attendant vegetation that likewise attaches to “cause” in a bit more detail.

As a result, a methodological tactic (axiomatic encapsulation) that was originally championed in the spirit of encouraging conceptual diversity eventually develops into a schema that favors methodological complacency with respect to the real life issues of productive concept formation. In doing so, analytic philosophy gradually substitutes an ersatz conception of formalized “rigor” in the stead of the close examination of applicational complexity that distinguishes Hertz’ original investigation of “force”’s puzzling behaviors (an enterprise that I regard as a paragon of philosophical “rigor” operating at its diagnostic best). Such is the lesson from developmental history that I attempted to distill within Imitation of Rigor (whose contents have been ably summarized within a recent review by Katherine Brading in Notre Dame Philosophical Reviews).

But Davidson and Quine scarcely qualified as warm friends of metaphysical endeavor. The modern adherents of “formal metaphysics” have continued to embrace most of their “Theory T” structural expectations while simultaneously rejecting positivist doubts with respect to the conceptual unacceptability of the vocabularies that we naturally employ when we wonder about how the actual composition of the external world relates to the claims that we make about it. I agree that such questions represent legitimate forms of intellectual concern, but their investigation demands a close study of the variegated conceptual instruments that we actually employ within productive science.  But “formal metaphysics” typically eschews the spadework required and rests its conclusions upon Theory T -inspired portraits of scientific method.

Indeed, writers such as David Lewis and Ted Sider commonly defend their formal proposals as simply “theories within metaphysics” that organize their favored armchair intuitions in a manner in which temporary infelicities can always be pardoned as useful “idealizations” in the same provisional manner in which classical physics allegedly justifies its temporary appeals to “point masses” (another faulty dictum with respect to actual practice in my opinion).

Philosophy’s Prophetic Telescope

These “Theory T” considerations alone can’t fully explicate the unabashed return to armchair speculation that is characteristic of contemporary effort within “formal metaphysics.” I have subsequently wondered whether an additional factor doesn’t trace to the particular constellation of doctrines that emerged within Hilary Putnam’s writings on “scientific realism” in the 1965-1975 period. Several supplementary themes there coalesce in an unfortunate manner.

(1) If a scientific practice has managed to obtain a non-trivial measure of practical capacity, there must be underlying externalist reasons that support these practices, in the same way that external considerations of environment and canvassing strategy help explicate why honey bees collect pollen in the patterns that they do. (This observation is sometimes called Putnam’s “no miracles argument”).

(2) Richard Boyd subsequently supplemented (1) (and Putnam accepted) with the restrictive dictum that “the terms in a mature scientific theory typically refer,” a developmental claim that strikes me as factually incorrect and supportive of the “natural kinds” doctrines that we should likewise eschew as descriptively inaccurate.

(3) Putnam further aligned his semantic themes with Saul Kripke’s contemporaneous doctrines with respect to modal logic which eventually led to the strong presumption that the “natural kinds” that science will eventually reveal will also carry with them enough “hyperintensional” ingredients to ensure that these future terminologies will find themselves able to reach coherently into whatever “possible worlds” become codified within any ultimate enclosing Theory T (whatever it may prove to be like otherwise). This predictive postulate allows present-day metaphysicians to confidently formulate their structural conclusions with little anxiety that their armchair-inspired proposals run substantive risk of becoming overturned in the scientific future.

Now I regard myself as a “scientific realist” in the vein of (1), but firmly believe that the complexities of real life scientific development should dissuade us from embracing Boyd’s simplistic prophecies with respect to the syntactic arrangements to be anticipated within any future science. Direct inspection shows that worthy forms of descriptive endeavor often derive their utilities from more sophisticated forms of data registration than thesis (2) presumes. I have recently investigated the environmental and strategic considerations that provide classical optics with its astonishing range of predictive and instrumental successes, but the true story of why the word “frequency” functions as such a useful term within these applications demands a far more complicated and nuanced “referential” story than any simple “‘frequency’ refers to X” slogan adequately captures (the same criticism applies to “structural realism” and allied doctrines).

Recent developments within so-called “multiscalar modeling” have likewise demonstrated how the bundle of seemingly “divergent relations” connected with the notion of classical “force” can be more effectively managed by embedding these localized techniques within a more capacious conceptual architecture than Theory T axiomatics anticipates. These modern tactics provide fresh exemplars of novel reconceptualizations in the spirit of the innovations that had originally impressed our philosopher/scientist forebears (Imitation of Rigor examines some of these new techniques in greater detail). I conclude that “maturity” in a science needn’t eventuate in simplistic word-to-world ties but often arrives at more complex varieties of semantic arrangement whose strategic underpinnings can usually be decoded after a considerable expenditure of straightforward scientific examination.

In any case, Putnam’s three supplementary theses, taken in conjunction with the expectations of standard “Theory T thinking” outfits armchair philosophy with a prophetic telescope that allows it to peer into an hypothesized future in which all of the irritating complexities of renormalization, asymptotics and cross-scalar homogenization will have happily vanished from view, having appeared along the way only as evanescent “Galilean idealizations” of little metaphysical import. These futuristic presumptions have convinced contemporary metaphysicians that detailed diagnoses of the sort that Hertz provided can be dismissed with an airy wave of the hand, “The complications to which you point properly belong to epistemology or the philosophy of language, whereas we are only interested in the account of worldly structure that science will eventually reach in the fullness of time.”

Science from the Armchair

Through such tropisms of lofty dismissal, the accumulations of doctrine outlined in this note have facilitated a surprising reversion to armchair demands that closely resemble the constrictive requirements on viable conceptualization against which our historical forebears had originally rebelled. As a result, contemporary discussion within “metaphysics” once again finds itself flooded with a host of extraneous demands upon science with respect to “grounding,” “the best systems account of laws” and much else that doesn’t arise from the direct inspection of practice in Hertz’ admirable manner. As we noted, the scientific community of his time was greatly impressed by the realization that “fresh eyes” can be opened upon a familiar subject (such as Euclidean geometry) through the exploration of alternative sets of conceptual primitives and the manner in which unusual “extension element” supplements can forge unanticipated bridges between topics that had previously seemed disconnected. But I find little acknowledgement of these important tactical considerations within the current literature on “grounding.”

From my own perspective, I have been particularly troubled by the fact that the writers responsible for these revitalized metaphysical endeavors frequently appeal offhandedly to “the models of classical physics” without providing any cogent identification of the axiomatic body that allegedly carves out these “models.” I believe that they have unwisely presumed that “Newtonian physics” must surely exemplify some unspecified but exemplary “Theory T” that can generically illuminate, in spite of its de facto descriptive inadequacies, all of the central metaphysical morals that any future “fundamental physics” will surely instantiate. Through this unfounded confidence in their “classical intuitions,” they ignore Hertz’ warnings with respect to tricky words that “have accumulated around [themselves], more relations than can be completely reconciled amongst themselves.” But if we lose sight of Hertz’s diagnostic cautions, we are likely to return to the venerable realm of armchair expectations that might have likewise appealed to a Robert Boyle or St. Thomas Aquinas.


Discussion welcome.

 

COMMENTS POLICY

The post The Rigor of Philosophy & the Complexity of the World (guest post) first appeared on Daily Nous.

Rejection Rates Should Not Be a Measure of Journal Quality (guest post)

“If philosophy relies too heavily on rejection rates as a measure for journal quality or prestige, we run the risk of further degrading the quality of peer review.”

In the following post, Toby Handfield, Professor of Philosophy at Monash University, and Kevin Zollman, Professor of Philosophy and Social and Decision Sciences at Carnegie Mellon University, explain why they believe the common practice of using journal rejection rates as a proxy for journal quality is bad.

This is the second in a series of weekly guest posts by different authors at Daily Nous this summer.


 

[Mel Bochner, “Counting Alternatives: The Wittgenstein Illustrations” (selections)]

Rejection Rates Should Not Be a Measure of Journal Quality
by Toby Handfield and Kevin Zollman

Ask any philosopher about the state of publishing in academic philosophy and they will complain. Near the top of the list will be the quality of reviews (they’re poor) and rejection rates (they’re high). Indeed, philosophy does have extremely high rejection rates relative to other fields. It’s extremely hard to understand why we have such high rejection rates. Perhaps there is simply more low-quality work in philosophy than other fields. Or, perhaps, rejection rates are themselves something that philosophy journals strive to maintain. Many journals strive to publish only the very best work within their purview, and perhaps they use their rejection rates to show themselves that they are succeeding.

Like many fields, philosophy also has an implicit hierarchy of journals. Of course, people disagree at the margins, but there seems to be widespread agreement among anglophone philosophers (at least) about what counts as a top 5 or top 10 journal. Looking at some (noisy) data about rejection rates, it does appear that the most highly regarded journals have high rejection rates. So, while we complain about rejection rates, we also seem to—directly or indirectly—reward journals that reject often.

It is quite natural to use rejection rates as a kind of proxy for the quality of the journal, especially in a field like philosophy where other qualitative and quantitative measures of quality are somewhat unreliable. We think it is quite common for philosophers to use the rejection rates of journals as a proxy for paper quality when thinking about hiring, promotion, and tenure. It’s impressive when a graduate student has published in The Philosophical Review, in large part because The Philosophical Review rejects so many papers. Rejection rates featured prominently—among many other things—in the recent controversy surrounding the Journal of Political Philosophy.

We, along with co-author Julian García, argue that this might be a dangerous mistake. (This paper is forthcoming in Philosophy of Science—a journal that, we feel obligated to point out, has a high rejection rate.) Our basic argument is that as journals become implicitly or explicitly judged by their rejection rates, the quality of peer review will go down, thus making journals worse. We do so by using a formal model, but the basic idea is not hard to understand.

We start by asking a very basic question: what is it that a journal is striving to achieve? We consider two alternatives: (1) that the journal is trying to maximize the average quality of its published papers or (2) that the journal is trying to maximize its rejection rate. The journal must decide both what threshold counts as good enough for their journal and also how much effort to invest in peer review. They can always make peer review better, but it comes at a cost (something that is all too familiar).

This already shows why judging journals by rejection rates can potentially be quite harmful. If a journal is merely striving to maximize its rejection rate, it doesn’t much care who it rejects. So, it has less incentive to invest in high quality peer review than does a journal that is judged by the average quality of papers in the journal. After all, if a journal only cares about rejection rates, it doesn’t much matter if a rejected paper was good or bad.

This already is probably sufficient to give one pause, but it actually gets much worse. In that quick argument, we implicitly assumed that there was a fixed population of authors who mindlessly submitted to the journal, hoping to get lucky. However, in the real world, authors might be aware of their chance of acceptance and choose not to submit if they regard the effort as not worth the cost.

A journal editor who wants to maintain a high rejection rate now has a problem. If they are too selective, authors of bad papers might opt not to submit, and a paper that isn’t submitted can’t be rejected. If a journal very predictably rejects papers below a given standard, their rejection rates will go down because authors of less good papers will know they don’t stand a chance of being accepted. A journal editor who cares about their journal’s rejection rate will then be motivated to tolerate more error in its peer review process in order to give authors a fighting chance to be accepted. They use their unreliable peer review as a carrot to encourage authors to submit, which in turn allows the journal to keep their rejection rates high.

We consider several variations on our model to demonstrate how this result is robust to different ways that authors might be incentivized to publish in different journals. We would encourage the interested reader to look at the details in the paper.

Of course, our method is to use simplified models, and in doing so we run the risk that a simplification might be driving the results. Most concerning, in our mind, is that our model features a world with only one journal. Philosophy has multiple journals, although in some fields of philosophy a single journal might dominate the area as the premier outlet for work in that area. Future work would need to determine if this is a critical assumption, although our guess is that it is not.

Although we don’t investigate this in our paper, we think that the process we identify might also exist in other selection processes like college and graduate school admission or hiring. In the US, colleges often advertise the selectivity of their admissions process, and we suspect that they face the same perverse incentives we identify.

Whether you share our intuition about this or not, we think the process we identify is concerning. If philosophy relies too heavily on rejection rates as a measure for journal quality or prestige, we run the risk of further degrading the quality of peer review. We think it is potentially problematic that journals sometimes advertise their rejection rates, lest it contribute to rejection rates being a sought after mark of prestige. Furthermore, we think it’s important that philosophy as a discipline walk back its use of rejection rates as a proxy for journal quality. To the extent that we are doing that now, it may actually serve to undermine the very thing we are hoping to achieve.


 

 

The post Rejection Rates Should Not Be a Measure of Journal Quality (guest post) first appeared on Daily Nous.

A Plea for Synthetic Philosophy (guest post)

“There need not be strict disciplinary boundaries between philosophy and other disciplines.”

In the following post, Catarina Dutilh Novaes, Professor of Philosophy at VU Amsterdam and Professorial Fellow at Arché at the University of St. Andrews explains and makes a case for synthetic philosophy.

This is the first in a series of weekly guest posts by different authors at Daily Nous this summer.


[“Abandoned Schoolhouse” by Gary Simmons]

A Plea for Synthetic Philosophy 
by Catarina Dutilh Novaes

A few years ago, the philosopher (and prolific blogger) Eric Schliesser used the term ‘synthetic philosophy’ to describe the work of Daniel Dennett and Peter Godfrey-Smith. Schliesser presented synthetic philosophy as “a style of philosophy that brings together insights, knowledge, and arguments from the special sciences with the aim to offer a coherent account of complex systems and connect these to a wider culture or other philosophical projects (or both). Synthetic philosophy may, in turn, generate new research in the special sciences…” Schliesser did not coin the term itself: the once influential but by now largely forgotten polymath 19th-century thinker Herbert Spencer (of ‘survival of the fittest’-fame) titled his mammoth 10-volume work covering biology, psychology, sociology and ethics System of Synthetic Philosophy.

But Schliesser can be credited for re-introducing the term to denote an approach in philosophy that has become more pervasive and widely accepted over the last two decades, namely one where philosophers engage extensively with work done in relevant (empirical) disciplines to inform their philosophical investigations and theories. Other recent examples of synthetic philosophers include Neil Levy (see his Bad Beliefs) and Kim Sterelny, who describes his book The Evolved Apprentice in the following terms: “The essay is an essay in philosophy in part because it depends primarily on the cognitive toolbox of philosophers: it is work of synthesis and argument, integrating ideas and suggestions from many different research traditions. No one science monopolizes this broad project though many contribute to it. So I exploit and depend on data, but do not provide new data” (Sterelny, 2012, p. xi). If this is a good description of synthetic philosophy, then it is fair to say that I have been (trying to be!) a synthetic philosopher for about 15 years now, so when Schliesser introduced the term, I adopted it wholeheartedly. (It is for sure much catchier than alternatives such as the cumbersome ‘empirically-informed philosophy’.)

The idea that there need not be strict disciplinary boundaries between philosophy and other disciplines enjoyed some popularity in the 20th century, in particular in the tradition of ‘scientific philosophy’ initiated by Bertrand Russell and continued by the Vienna Circle and later with their ‘heirs’ in the United States such as W.V.O. Quine (who used the ambiguous term ‘naturalism’ to describe the idea of continuity between philosophy and other disciplines) and Hilary Putnam. (Much before that, over the centuries, there was for the most arguably no strict separation between philosophy and other disciplines either; Aristotle was first and foremost a biologist.) But there was also much resistance within analytic philosophy to the idea that philosophical inquiry should in any way be informed by scientific findings (see this piece that I co-wrote on the dissonant origins of analytic philosophy). This resistance is to be traced back to G.E. Moore, who insisted that moral philosophy and ethics in particular were strictly non-scientific, purely conceptual domains. It continued with the so-called ‘ordinary language philosophers’, as exemplified by the damning critique of Carnap’s notion of explication in Strawson’s piece for the Carnap Living Philosophers volume.

To motivate a strict separation between science and philosophy, a point sometimes made is that scientists are involved in the merely descriptive inquiry of telling us how things are, while philosophers are involved in conceptual and (or) normative inquiry as well, which includes looking at how things ought to be understood, and how they ought to be. If true, this point has important methodological implications, as different methods are used for different types of investigation. Methods to investigate how things are include data collection, experiments, field work, etc. Methods to investigate how things should be include conceptual analysis, ‘intuitions’, thought experiments, etc. Some decades ago, however, this presumed neat separation was challenged by the so-called experimental philosophy approach, which prompted what might described as a small methodological crisis in analytic philosophy. Could philosophy be empirical/experimental after all? The X-Phi challenge made it clear that more sustained methodological reflection was needed, and philosophers spent much of the first two decades of the 21st century discussing the ins and outs of different methods for philosophical inquiry (see Williamson’s The Philosophy of Philosophy and the Oxford Handbook of Philosophical Methodology).

It is fair to say that scientific/synthetic philosophy ‘won’ the battle in that the Moorean rejection of engagement with empirical findings in philosophical inquiry has become much less widespread in the last decade. I speak from (anecdotal) personal experience: when I was working on the project that culminated in my monograph Formal Languages in Logic (2012), I often got remarks to the effect that it was all very interesting, but what I was doing wasn’t really philosophy. (My standard response was: well, I’m glad no one is saying that it’s all very philosophical, but not interesting.) By contrast, in later years the research that culminated in my monograph The Dialogical Roots of Deduction (2020) did not typically prompt the same kind of reaction, even though it was just as empirically oriented as my earlier work. The fact that The Dialogical Roots of Deduction won the Lakatos Award in 2022 (sorry folks, time for some shameless self-promotion!) attests to the widespread acceptance of synthetic philosophy as an approach to philosophical inquiry (though philosophers of science are, of course, from the start more amenable to the general idea of engaging with other disciplines).

In my opinion, philosophy is especially well-placed to facilitate much-needed interdisciplinary collaboration between different disciplines. It is now widely recognized that so-called ‘wicked problems’ require complementary approaches to be addressed, but each methodology has its limitations and dead angles. Experimental methods may lack ‘ecological validity’ (lab situations do not really reproduce the phenomena in the wild); quantitative methods are often not very fine-grained and may give rise to spurious correlations and ‘noise’; qualitative methods may be instances of ‘cherry-picking’ and have limited reach. Thus, what has become clear, in particular during the Covid-19 pandemic, is that triangulation of methods is essential for investigating complex problems.

It seems to me that philosophers have much to contribute to interdisciplinarity and triangulation efforts for two main reasons: philosophers are trained to engage in careful conceptual inquiry, clarifying and sharpening significant (scientific) concepts and sometimes introducing new ones (Carnapian explication, conceptual engineering); philosophers may be better able to see the forest rather than only the trees, as it were, by drawing on various scientific disciplines (as suggested in the Sterelny quote above) and noticing connections that may remain unnoticed within each specific discipline. Philosophy has much more potential for synthesis than is often recognized (even by us philosophers), and this is perfectly compatible with the centrality of analysis in analytic philosophy. (Traditionally, analysis and synthesis were viewed as two complementary rather than incompatible processes: you break things down to them put them back together again, usually in a different, more fruitful configuration.) (Note: commenting on an earlier draft of this post, Eric Schliesser remarked that my conception of synthetic philosophy differs in some important respects from his. He thinks that my conception resembles that of Kitcher’s, which he discussed in this blog post.)

True enough, there are also a number of difficulties, pitfalls and risks involved in attempting to do synthetic philosophy. For starters, the approach requires that the philosopher be conversant with various different scientific disciplines; she has to be a ‘polymath’ in some sense, which in the current scenario of scientific hyper-specialization is a formidable challenge. Secondly, there is a perennial risk of conceptual confusion/equivocation and ‘talking past each other’ for lack of a common vocabulary (a familiar problem in interdisciplinarity studies). Third, scientific studies themselves are not always reliable guides, with many important results not being replicated (see the famous ‘replication crisis’ in psychology and other disciplines).

However, while these issues are real and must be taken seriously, I submit that they should not be viewed as knock-down arguments against synthetic philosophy. (I have responses to each of them but I’m running out of space!) Sustained methodological reflection on the ins and outs of synthetic philosophy is still needed, but I hope to have established here at least that synthetic philosophy is an interesting and viable approach for philosophical inquiry.


The post A Plea for Synthetic Philosophy (guest post) first appeared on Daily Nous.

❌