This is the story of how a school science fair inspired a passion for science communication, a PhD in microbiology, and a valuable perspective on the current AI debate.
Dr Jessica Miles recently participated in the SSP2023 panel on AI and the Integrity of Scholarly Publishing, the writeup from which has just been published on the Scholarly Kitchen.
I caught up with Jessica to chat about how she came to be on the panel, her background in microbiology, and her thoughts on the future of scholarly publishing in an AI world. I got a Barack Obama impression for free đ

Quick links
- From school science fair to SSP2023
- What does Mary Shelleyâs Frankenstein have to do with AI?
- Generative AI â the new normal
- Barack Obama?
- What is the publishing industryâs moat?
- The value of getting things done
From school science fair to SSP2023
John Hammersley: Tell us how you came to be on the SSP2023 panel on AI
Jessica Miles:
It seems so surprising, right? (JH: Not at all!!) I asked myself that same question the other day â I was at the SpringerNature office and I met somebody who works on submissions who Iâd connected with on LinkedIn. He stopped me and said: âI looked at your LinkedIn, your career, and it seems âŠhe didnât say bizarre, but went with âvery interesting!ââŠand Iâd love to chat about it.â
And it made me think â in respect to the SSP AI debate and my participation â that one of the reasons I was invited is precisely because I donât have the normal profile of someone that you might expect to be participating in an AI debate. But if thatâs the case, whatâs interesting enough about my career that makes my perspective on this noteworthy? And the conclusion I came to is that, beyond my experience in scholarly publishing, itâs quite a bit to do with my experience with (and enthusiasm for!) science communication. Iâve been interested in science communication for a long time â when I went to school, the university I picked specifically had a program in science communication.
From an early age I learned about Michael Faraday (and his role as a science communicator) and I thought that would be something really cool to do, even though I didnât know what that was at first! I thought I maybe wanted to do science journalism, eventually landed into research and so when I think about a lot of the different things that Iâve been interested in or done, they really fit kind of nicely within that space of science communication that Iâve been following for a long time.
So getting back to the SSP debate, if you think about the audience being in the scholarly technology space, not an audience of experts on AI, this idea of communicating science in a different way but with this sort of publishing lens comes to mind. And so I think thatâs why I was really excited to get the opportunity to do this and why I think it makes sense because itâs not like Iâm giving a talk at Google or to a group of AI experts, itâs really about communicating science, but at the intersection of something else, something new.
What does Mary Shelleyâs Frankenstein have to do with AI?
John: Yes, I agree itâs really important that different perspectives are included (in these discussions), and that we donât reduce AI to just being about the technology because itâs totally not just that! Chat GPTâs explosion into the public consciousness is a great example â it wasnât so much the technology but the interaction with it that caught the imagination. So I think itâs good when forums do try to include lots of different viewpoints. But I also know what you mean in terms of not feeling qualified, because your background isnât in the technology side of it.
Jessica:
Exactly â my PhD is in microbiology, not machine learning, and I certainly wouldnât call myself an AI expert. So, yeah, itâs a natural question people asked of me (after the debate). And thatâs the answer that I came up with after some reflection! I think having that perspective informs the way Iâm thinking about these technologies. For example, I think about one of the courses that I took where we read Mary Shelleyâs Frankenstein. The text of Frankenstein is not necessarily considered particularly difficult and some might ask, âWhy is this a college level text?â But one of the main ideas of her work is that sheâs interrogating how society is thinking about new technologies â itâs a commentary on science, and I was thinking about that recently because it feels like weâre in another similar moment where weâre having to grapple with this new technology and some people are quite fearful and some people are quite excited.
And some people worry that it will take away the mystery of the world if we can recapitulate human thought, human intelligence.”
Where does that leave us from a philosophical perspective? Weâre starting (or continuing) to ask if anything is sacred anymore? There are a lot of really interesting questions, and having a historical perspective feels like a nice way to approach this moment so that itâs not so overwhelming. Itâs like âokay there are historical parallels, and yes this might buck those trends but at least you have some kind of grounding to approach all of this.â
We spent a lot of time on Frankenstein â we must have spent several months on it â and itâs not necessarily something you would think would require that level of rigor. But when you consider it from the context of society and all the other things that were happening around that time â industrial revolution, and the rapid pace of change that brought â there is very much an allegory meaning behind it that belies its reputation as a simple text.
Generative AI â the new normal
John: Frankenstein is a nice example of the fear a new technology can generate. Yet, weâre very good at adapting to the new normal as new things are developed â stuff that would have been seen as magical and amazing previously is then seen as expected, trivial commonplace, once itâs been (repeatedly) demonstrated.
Generative AI â for text, images, and more â is quickly becoming the new normal. So Iâm curious, because I didnât go to SSP this year and I havenât had a chance to look over the sessions â what did you find most interesting at SSP, either from the sessions or just from the chats you had whilst you were there?
Jessica:
One topic that particularly resonated was that of the scholarly intellectual output of humans â that the ideas arenât machine generated, and that manifests itself in terms of the written text. So you can imagine, especially in humanities, thereâs a very heavy focus on making sure that the text isnât machine generated, or at least that the ideas and the text are very much those of a human being.
On the science side however, it seemed to be not so much a concern that the text itself isnât from humans, but what was seen as more worrying was that thereâs even more active fraud with respect to data outputs. The opening plenary was from Elisabeth Bik, whom you might know from her work in ensuring scientific integrity and thoroughly investigating image manipulation.
John: Yes, Elisabeth Bik is a legend â I donât know how she finds the energy in the face of ever more papers!
Jessica:
She talked about image manipulation and her efforts as a whole, and then focused on the potential implications for these new technologies â not so much on the text side, but on the image side with respect to making synthetic outputs. From her talk, it sounds like weâre in a challenging moment because although the general consensus was that weâre a little bit too early on with those technologies to really see the impact, everyone feels like thereâs a brewing storm in terms of all the people who have had time to use these tools, and learn these tools â that we will see nefarious actors (e.g. paper mills) start to incorporate them in ways that we havenât seen before. And that weâre ill-equipped as a publishing community to deal with it.
John: Itâs interesting you mention that. Tim Vines shared a tweet todayâŠ
Jessica: Tim was my debate partner!
John: Itâs a small world indeed! I know Tim from when he founded Axios Review, back when Overleaf was also a new kid on the block đ He shared this tweet and it really highlighted how close we are to researchers being able to use AI to generate plausible scientific papers.

Jessica:
This feels almost fully autonomous, not fully, but with a sort of minimal intervention. And obviously I havenât looked at this in great detail, but itâs a huge step forward from AI âsimplyâ helping to fine-tune something a researcher has written, like weâve had before with grammar tools.
Barack Obama?
John: Exactly. It feels like weâre almost at the point where a researcher can ask the AI to write the paper, like e.g. a CEO could say to Chat GPT âplease write five paragraphs on explaining why the company retreat has to be canceled this year and make it apologetic and sound like Barack Obama.â
Jessica: (in a deep voice) Folks, the company retreatâŠ
John: âŠ
Jessica: Iâve just been watching his new documentary on Netflix so I have his voice in my mindâŠ
John: Okay.
Jessica: Itâs actually quite good â he interviews people who have different roles at the same companies and internally tries to get at âwhat is working, what is a good jobâ? Itâs very US focused but I thought it was quite interesting. Anyway, pivot, but we can go back to AI!
John: This is a nice aside! I saw a video excerpt of him speaking recently, where he was asked for the most important career advice for young people and I believe he said âJust learn how to get stuff done.â Because there are lots of people that can always explain problems, who can talk and talk about stuff, but if youâre the one that can put your hand up and say âThatâs all right, Iâll sort that, I can handle thatâ, it can get you a long way.
What is the publishing industryâs moat?
John: But yes, back to AI, some of the new generative image stuff is a bit crazy â being able to use it to expand on an existing image, rather than just generate something standalone, suddenly makes it useful for a whole load of new things. And I see Google has now also released a music generator, which generates music for you based on your prompt. Everything seems to be happening faster, at a bigger scale than before, and I can see why scholarly publishing is trying to figure out how to ride this tsunamiâŠ
Jessica:
How are we going to keep up? YeahâŠto borrow a phrase, âWhat is our moat?â
I think a lot of people are thinking about that, especially given that thereâs not only Chat GPT but also all of the smaller models that proficient people can now train on their own. As a scholarly publisher, youâre serving a population that has an over-representation of peopleâyour core demographicâwho are going to be really fluent in these models, so what can you offer to them? What can you offer this group that they canât kind of already do on their own? Thatâs the million dollar question.
John: Publishers would say they try to help with trust in science, and research integrity â for example, through peer review and editorial best-practices. But they also have an image problem, because there is also a tendency to chase volume, because volume generally equals more revenue, and if youâre chasing volume then youâre going to accept some papers that prove to be controversial. Itâs an interesting dynamic between volume, trust and quality.
Jessica:
The volume question is always really interesting because thereâs that perspective where people assume publishers have these commercial incentives to grow volume, which has some validity in an OA context of course, but thereâs the other viewpoint which is that science is opening up and becoming more inclusive) and that almost by definition means publishing more papers from more authors. But restricting what people can publishâŠshould that be up to the publishers? I do think there is a sense that publishers donât want to be in the business of telling people they can or canât publish. Because itâs one thing to say at a journal level, âwe donât think this paper meets the editorial barâ, itâs quite another thing to say, âwe donât think this paper ought to be published anywhere, everâ, right? In fact, I think many publishers would say the opposite: âany sound paper ought to be published somewhere.â We see this view borne out by the increasing investments publishers are making to support authors even before they submit to a journal and also to find another journal, if a paper isnât accepted.
Another consideration that was also mentioned at the conference is that writing papers is the way that scientists communicate with each other. Dr. Bik was saying that hundreds of thousands (I forget the exact number) of papers were published on COVID research in the last three years. And she said, âWhy do we need that many papers?â Well, in retrospect itâs very easy to say that, but if you are an editor working during COVID, as I once was, for which paper are you going to say âwe donât actually need this oneâ? Everything was happening so fast, you didnât know what papers would be the crucial ones in the long run â how can you make that judgment? So I do think more gets published, in response to perceived need.
To be clear: youâre not going to publish anything you donât think is scientifically sound, but most of us are not going to try to set the bar at âwill this be valuable three or five years laterâ? That is quite a different bar than âis this just methodologically soundâ?
And I donât know if weâve gone too far from the question on AI, but with higher volume comes this need to curateâwhich weâve needed for a long timeâand as the volume increases, the need to curate increases. This curation is another value-add for publishers, but also again something that AI can potentially be very good at given reasonable inputs. I want to be careful not to set up a false dichotomy of âpublisher-curationâ versus âAI-curationâ, because publishers are very much already using AI for things like summarization and recommendation engines, but there is the question of whether publishers, moving forward, continue to drive this curation.
John: One reason you write papers as early career researchers is because youâre learning how to write a paper â youâre publishing some results that arenât necessarily all that ground-breaking but itâs a record of what youâve done, why you say your methodology is sound, and so forth. And in doing so you learn how to write a paper. It raises the question of how much e.g. undergraduate work should be published to give say third and fourth year students the opportunity to go through the process of getting their work published.
Jessica:
It is a fascinating question because not only is the pedagogy piece realâthat early trainees, undergrads, etc need to learn how to writeâbut also because publishing is about putting your ideas out there and becoming known to the community.
You can present at conferences (which is another skill), but the scale of impact from that is typically much smaller, and typically it can be difficult to get to the point where you would present at a conference without having published. If you arenât able to publish, then youâre not able to establish yourself within the community. So publishing is critical for early career researchers to get that first step on the ladder.
The value of getting things done
John: That brings us back almost full circle to where we started â you talked about Michael Faraday as a science communicator, and that you were into science as a kidâŠ
Jessica: I definitely was â I did science fairs at school and thatâs actually how I learned about Michael Faraday. And I hadnât really thought about the communication of science as being distinct from the research itself but for some reason the communication aspect specifically really appealed to me. I have always been quite curious and into learning and storytelling, and being able to communicate ideas through stories.

John: As weâre nearly at time, perhaps thatâs a good point to end on â the value of science fairs, and the value of doing something for yourself and getting that experience is one thing AI is not going to replace; it might be able to create the poster for you, or it might write a paper for you, but if youâre the one thatâs got to stand there and say what youâve done, itâs usually pretty obvious if you know what youâre talking about or not, and thereâs no AI substitute for that.
Jessica: Yet!
Jessica Miles holds a doctorate in Microbiology from Yale University and now leads strategic planning at Holtzbrinck Publishing Group. I (John Hammersley) am one of the co-founders of Overleaf and now lead community engagement at Digital Science.