Philosophy and Science Policy: A Report from the Field I

I’m actually going to give a series of reports from the field, including a chapter in a book on Field Philosophy that I’m revising now in light of editor/reviewer comments. In the chapter, I discuss our Comparative Assessment of Peer Review project. For a brief account of Field Philosophy, see the preprint of a manuscript I co-authored with Diana Hicks. That’s also being revised now.

Today, however, I will be focusing on more pressing current events having to do with Plan S. So, I will give a talk at the NJIT Department of Humanities Fall Colloquium Series to try to let my colleagues know what I’ve been up to recently. Here are the slides.

As a philosopher, my approach is not simply to offer an objective ‘report’. Instead, I will be offering an argument.

If we want to encourage academic flourishing, then we need new ways of evaluating academic research. We want to encourage academic flourishing. Therefore, we need new ways of evaluating academic research.

Of course, the argument also refers to my own activities. I want my department to understand and value my forays into the field of science policy. But that will mean revaluing the way I am currently evaluated (which is along fairly standard lines).

Camp Engineering Education AfterNext

This looks like fun!

Thoughts from the Public Philosophy Network 2018 Conference

First, I’ve been away from my own blog for far too long. My apologies. Second, no more ‘Press This’?! Ugh. So, here is a LINK to the full program of PPN 2018.

Most of these thoughts were generated during the workshop run by Paul Thompson on day 1 on ‘Evaluating Public Philosophy as Academic Scholarship’. This issue is important for everyone who would like to see public philosophy succeed; but it is vitally important for those of us on the tenure track, since not being able to evaluate public philosophy as academic scholarship often means that it is reduced to a ‘service’ activity. Service, of course, is seen as even less important than teaching, which is often seen as less important than research. This hierarchy may be altered at small liberal arts colleges or others that put special emphasis on teaching. Generally speaking, though, one’s research rules in tenure decisions. I’ve never heard, or even heard of, any advice along the lines of ‘Do more teaching and publish less’ or ‘make sure you get on more committees or peer review more journal manuscripts’. Whereas ‘Just publish more’ is something I hear frequently.

So, it’s vitally important to be able to evaluate public philosophy as academic scholarship.

I want to add that, although many of these ideas were not my own and came from group discussion, I am solely responsible for the way I put them here. I may mess up, but no one else should be blamed for my mistakes. What follows isn’t quite the ‘Survival Guide’ that Michael O’Rourke suggested developing. Instead, it is a list of things I (and perhaps others) would like to see coming from PPN. (This may change what PPN is, of course. Can a network that meets once in while provide these things?)

We need:

  1. A statement on the value of public philosophy as academic scholarship. [EDIT: The expression of this need came up at the workshop, but no one there mentioned that such a statement already exists HERE from the APA.  Thanks to Jonathan Ellis and Kelly Parker for help in finding it! Apologies to APA for my ignorance.]
  2. A list of scholarly journals that are public philosophy friendly (i.e., where one can submit and publish work that includes public philosophy). The list would need to be curated so that new journals can be added and old ones removed when they fit or don’t fit the bill.
  3. A list of tools for making the case for the value of public philosophy. I have in mind things like altmetrics (see HERE or HERE or HERE), but it could also include building capacity among a set of potential peers who could serve as reviewers for public philosophy scholarship.
  4. Of course, developing a cohort of peers will mean developing a set of community standards for what counts as good public philosophy. I wouldn’t want that imposed from above (somewhere?) and think this will arise naturally if we are able to foster the development of the community.
  5. Some sort of infrastructure for networking. It’s supposedly a network, right? Is there anywhere people can post profiles?
  6. A repository of documents related to promotion and tenure in public philosophy. Katie Plaisance described how she developed a memorandum of understanding detailing the fact that her remarkably collaborative work deserved full credit as research, despite the fact that she works in a field that seems to value sole-authorship to the detriment of collaborative research. Katie was awesome and said she would share that document with me. But what if she (or everyone) who did smart and cool things like this to help guarantee their ability to do public philosophy had a central repository where all these documents could be posted for everyone to view and use? What if departments that have good criteria for promotion and tenure — criteria that allow for or even encourage public philosophy as scholarship — could post them on such a repository as resources for others?
  7. Leadership! Developing and maintaining these (and no doubt others I’ve missed) resources will require leadership, and maybe even money.

I’d be interested in thoughts on this list, including things you think should be added to it.

What a difference a day makes: How social media is transforming scientific debate (with tweets) · deevybee · Storify

This is definitely worth a look, whether you’re into the idea of post-publication peer review or not.

What a difference a day makes: How social media is transforming scientific debate (with tweets) · deevybee · Storify.

Lethal Autonomous Robots (“Killer Robots”) | Center for Ethics & Technology | Georgia Institute of Technology | Atlanta, GA

Lethal Autonomous Robots (\”Killer Robots\”)

Monday, 18 November 2013 05:00 pm to 07:00 pm EST

Location: 

Global Learning Center (in Tech Square), room 129

WATCH the simultaneously streamed WEBCAST at: 

http://proed.pe.gatech.edu/gtpe/pelive/tech_debate_111813/

Debate and Q&A for both

Lethal Autonomous Robots (LARs) are machines that can decide to kill. Such a technology has the potential to revolutionize modern warfare and more. The need for understanding LARs is essential to decide whether their development and possible deployment should be regulated or banned. Are LARs ethical?

via Lethal Autonomous Robots ("Killer Robots") | Center for Ethics & Technology | Georgia Institute of Technology | Atlanta, GA.

‘Big Data’ Is Bunk, Obama Campaign’s Tech Guru Tells University Leaders – Wired Campus – The Chronicle of Higher Education

“The ‘big’ there is purely marketing,” Mr. Reed said. “This is all fear … This is about you buying big expensive servers and whatnot.”

via 'Big Data' Is Bunk, Obama Campaign's Tech Guru Tells University Leaders – Wired Campus – The Chronicle of Higher Education.

Also funny what he says about his own education ….

Me on the Web, According to ImpactStory.org and Altmetric.com

Impact Story is one of the two altmetrics tools that allow individual researchers to find out something about the social media buzz surrounding their activities; the other is Altmetric.com. Although other developers exist, I can’t seem to figure out how I, as an individual, can use their tools (I’m looking at you, Plum Analytics).

There are a few major differences between Impact Story and Altmetric.com from a user standpoint. First, Impact Story is not for profit, while Altmetric.com is a business. Second, Impact Story steers one to create a collection of products that together tell a story of one’s impact. Altmetric.com, on the other hand, steers one to generate figures for the impact of individual products. Third, Impact Story allows for a range of products, including those tagged with URLs as well as DOIs; Altmetric.com only works with DOIs. This means that Impact Story can gather info on things like blog posts, while Altmetric.com is focused on scholarly articles. Finally, and this is a big difference, Impact Story deemphasizes numbers, while Atlmetric.com assigns a number, the Altmetric score, to each product.

Here is my latest Impact Story.

Here is one of my products according to Altmetric; here’s another; and another.

It’s interesting to see how Impact Story and Altmetric differ, both in their approaches and in terms of what they find on the same products.

I think the implications of these tools are enormous. I’d be interested to hear your thoughts!