AI as a Thought Partner For Your Research

The initial results of my experiment.

In partnership with

[image created with Dall-E 3 via ChatGPT Plus]

Welcome to AutomatED: the newsletter on how to teach better with tech.

In each edition, I share what I have learned — and am learning — about AI and tech in the university classroom. What works, what doesn't, and why.

Today, I share results from an experiment where OpenAI's o1-pro AI model provided remarkably thorough and helpful feedback on an economics research paper, revealing that advanced AI can serve as a valuable thought partner for us and offering promising applications for research development and student feedback.

As I note, if you want to participate in this experiment with your research, respond to this email to let me know!

📬 From Our Partners:
AI for Presentations

The future of presentations, powered by AI

Gamma is a modern alternative to slides, powered by AI. Create beautiful and engaging presentations in minutes. Try it free today.

Remember: Advertisements like this one are not visible to ✨Premium subscribers. Sign up today to support my work (and enjoy $500+ worth of features and benefits)!

💡 Idea of the Week:
How Good is AI as a Thought Partner?

Can any current AI serve as a genuinely useful thought partner for academic researchers operating at the frontiers of their fields?

To explore this question and provide evidence to you all about the answer, I am conducting a simple experiment with OpenAI's o1-pro, their most advanced reasoning AI model (which requires a $200/month subscription via ChatGPT Pro; its newer, faster, but slightly dumber sibling, o3-mini, is available to all in a low-compute version).

A few months ago, when o1-pro came out, I invited professors to volunteer drafts of their academic papers, which I would then share with o1-pro, asking it to provide thoughtful feedback as a colleague might. The AI would evaluate the paper and offer suggestions for improvement, after which I'd gather the professors' reactions to assess the value of this AI-assisted thought partnership.

Note: If you’d like to participate in this experiment, respond to this email to get in touch! In other words, the experiment is ongoing…

Today, I'm sharing results from one particularly illuminating case: a paper about economics education titled “Relevance, Belonging, and Growth Mindsets in Economics: Differences across Institution Types.” The paper was co-authored by Caroline Krafft, Kristine West, Allen Bellas (my correspondent), Ming Chien Lo, Adriana Cortes Mendosa, Gabrielle Agbenyiga, J Dombroski, Nayomi Her, and Joy Moua. Allen Bellas and Ming Chien Lo are at Metropolitan State University, St. Paul, MN, while the rest of the co-authors are at St. Catherine University in the same city.

Crucially, I did not turn on Deep Research, which would have connected o1-pro with the internet (notably, Deep Research is now available to all paying ChatGPT users; more on it here, here, and here). Instead, I tested o1-pro’s ability to engage with the paper based on its training data alone.

In short, the authors received extensive, structured feedback from o1-pro that impressed them in multiple ways, as researchers and as teachers of research methods.

What's most striking is how little guidance the AI needed, all due to the power of o1-pro — like other reasoning models — to prompt itself.

Let's look at what happened.

The Paper

The paper I shared with o1-pro examines why economics as a field has struggled with diversity and inclusion compared to STEM disciplines, focusing on psychological aspects (relevance, belonging, and growth mindset — collectively "RBG") across different institutional settings, including minority-serving institutions (MSIs) and women's colleges.

This was a substantive, data-driven paper with a robust methodology section, analysis of survey data from 24 colleges and universities, and thoughtful discussion of implications for economics education. At 15 pages, single-spaced (excluding the 4 pages of references and 15 tables), it represented the kind of complex academic work that typically requires careful review from knowledgeable colleagues.

My Prompt

The prompt I used relied on no field-specific expertise, as I have none (I majored in economics a long time ago but let’s be honest here). Lengthier prompts with more background information, criticisms from knowledgeable colleagues of the paper, etc. would likely get even better results (more on this below). My prompt just conveys the broad context of my request, without any special insight into how to proceed.

Here is my prompt:

Here's the deal. I run a newsletter where I help professors with artificial intelligence, whether in their pedagogy or research. I am running a test of AI like you to see how useful you are at helping professors with their research. In particular, I am soliciting research problems, papers, etc from a range of professors, giving them to AI like you, and asking you to act as a highly useful thought partner. Your job is to evaluate or discuss what I give you in a reflective way, as if you were trying to give feedback as a colleague on the professor's work.

In the current instance, I got a paper from a professor and MBA director of an economics and finance program, but it is also coauthored with some colleagues. The submitting professor's primary work is in environmental economics, principles of economics, benefit-cost analysis, and the economics of public policy and social issues.

I will paste the paper draft/manuscript below. Obviously, to offer feedback or useful thoughts to high-level researchers working on this document, you will likely need to focus on specific parts or aspects of it, one at a time. I want you to first analyze it and develop a plan to evaluate or give feedback on it, with a goal of providing feedback that might improve future drafts (not critiques that are merely critical). Present this plan to me first in your first message. Each of its parts should correspond with one of your subsequent messages; that is, I will ask you to continue from one part to the next, after you present the plan with its parts, and you dedicate one message to each part, with me prompting you to continue.

Here is the paper:

“““

<PASTED IT HERE>

“““

My full prompt for o1-pro

The AI Feedback

Given this limited and generic prompt, o1-pro developed its own comprehensive feedback plan with eight distinct parts:

  1. High-Level Framing and Organization

  2. Literature Review and Theoretical Foundations

  3. Data and Sampling Strategy

  4. Measurement of Key Constructs (RBG) and Empirical Methods

  5. Presentation and Interpretation of Results

  6. Discussion, Contributions, and Limitations

  7. Writing Style and Readability

  8. Future Directions and Broader Implications

For each section, the AI provided detailed, thoughtful analysis that balanced acknowledgment of strengths with suggestions for improvement. The feedback totaled over 10,000 words — longer than many published papers themselves!

Here's a sample from the "High-Level Framing and Organization" section to give you a flavor of the feedback:

The abstract provides a succinct statement of the underrepresentation problem in economics and situates it vis-à-vis STEM fields. It concisely introduces the core constructs (relevance, belonging, and growth mindset—RBG) and underscores their link to student outcomes.

[…]

While the abstract does present the key findings, it could benefit from an even more explicit statement about the implications for economics education. For example, specifying why the observed baseline differences (and eventual convergence or divergence) matter for departments and policy-makers would give the abstract added punch.

Some of o1-pro’s feedback on the “high-level framing and organization” of the paper

The feedback wasn't just superficial; it engaged deeply with the paper's methodology. For instance, on sampling strategy:

The paper does an excellent job of identifying a key limitation in existing research—namely, the focus on primarily PWI (predominantly white institution) and co-ed contexts. By deliberately oversampling MSIs and women's colleges, the authors ensure that these identity-focused institutions are well-represented. This design is central to their goal of comparing RBG across diverse settings.

[...]

The procedure of bootstrapping multiple random draws from IPEDS data to achieve a balanced sample across three institutional covariates (Pell-share, selectivity, and student-faculty ratio) is notable. It demonstrates a thoughtful approach to constructing a sample with variation across key institutional traits, rather than relying on ad hoc or convenience sampling.

[...]

The paper notes that 2-year institutions and non-women’s-college MSIs saw higher attrition rates. It might help to discuss possible reasons for this discrepancy (e.g., students dropping out, transferring, or less stable enrollment patterns) and how it might bias comparisons over time. If feasible, a short analysis of whether baseline responders who dropped out differ systematically from those who remained (e.g., by major, prior GPA, or household income) would lend extra credibility to the endline conclusions.

Some of o1-pro’s feedback on the “data and sampling strategy” of the paper

o1-pro also went beyond critiquing the paper to offering thoughtful extensions and implications for future research:

Since the paper frequently references how STEM fields have made somewhat better strides in diversity, systematic comparisons of RBG across different disciplines could provide insights into whether economics is uniquely off-putting or if certain departmental practices are universal.

[...]

Future studies might replicate these RBG measures in fields like computer science, physics, or political science to see if female or minority students similarly lose a sense of relevance over time, or if the pattern is more pronounced in economics.

More o1-pro feedback, this time on future directions

The AI consistently balanced affirmation with constructive critique, maintaining a collegial tone throughout. It identified genuine strengths while suggesting concrete improvements — much like an ideal peer reviewer.

The Professors' Response

The professors were genuinely impressed by the quality and thoroughness of the AI feedback. As Bellas wrote:

The structure of the feedback was great. It was very well organized into sections and, as is usually the suggestion for good feedback, started with the strengths before moving to the opportunities for improvement (OFIs).

Both the strengths and OFIs were mostly on-point and relevant, and as you mentioned, without specific training [or prompting] related to the topic this is what we would expect.

The feedback was also very thorough and specific, so there were a lot of actionable things we can shoot for in revising the paper.

Allen Bellas on the quality of the AI feedback

While they noted that "the length of the feedback was a bit daunting," they appreciated that "the excellent structure made it easier to read through and digest the feedback and to identify specific things to work on."

Bellas reported that their colleagues immediately recognized the educational potential:

A couple of members of our team mentioned that this would be an excellent teaching tool, and that this would be a great way to provide students with high-quality feedback on drafts of their papers. The fact that it is an AI engine rather than the instructor is likely to be less threatening for students and less likely to introduce bias in the grading process, which I think is honestly an issue when I review a draft and then grade a later version of the paper.

Allen Bellas on the pedagogical upshots of this tool

Takeaways and Implications

What does this experiment tell us about AI as a thought partner for academic work? Here are a few of my thoughts:

1. Minimal prompting of the reasoning models like o1-pro can yield high-quality results

The most surprising aspect of this experiment was how little guidance the AI needed. I didn't provide detailed instructions about what constitutes "good feedback" or specify a particular structure. As noted, I don’t know anything about the subject matter, so this prompt is about as good as it’s going to get! Instead, I simply described the basic task in general terms and let the AI determine how to approach it. This may mean that for certain professional tasks, sometimes the best prompt is the simplest one for reasoning models like o1-pro — allowing the AI's capabilities to shine through without constraining them with overly specific instructions. (This matches with the advice of the AI developers, mind you.) But it also leaves room for the possibility that more information in the prompt, which could be provided by a subject-matter expert, would get even better results. I’ve found that the latter can be true in some cases where I do have expertise.

2. AI feedback offers unique advantages over traditional peer review

AI feedback offers several distinctive benefits over human peer review and collaboration:

  • Exceptional thoroughness by default

  • Consistent structure and organization (the use of bullet points is borderline compulsive from these models, I’ve found, but it is useful for this sort of case)

  • No personal biases or academic politics (though there may be biases in the training data, of course)

  • Immediate availability (no waiting for busy colleagues)

  • A balanced approach that acknowledges strengths while suggesting improvements (reasoning models do this almost to a fault)

3. It's particularly valuable for structure and organization

The AI excelled at analyzing the paper's overall structure, flow, and organization — aspects that even experienced academics sometimes struggle to evaluate objectively in their own work. (I’ve definitely had issues with this…) By offering a fresh perspective on how arguments are sequenced and connected, AI feedback can help researchers strengthen the fundamental architecture of their papers.

4. There are still important limitations

Despite its impressive performance, AI thought partnership has clear boundaries. The AI can't evaluate the factual accuracy of novel research findings (without source material being provided to compare), can't bring disciplinary expertise beyond what's in its training data or what’s shared in the prompt, can't fully understand the nuanced dynamics of academic publishing in specific fields, etc.

5. The educational applications are particularly promising

The authors immediately recognized the potential for using AI feedback in their teaching. For students, receiving detailed, non-threatening feedback from an AI might feel less personal and more constructive than instructor comments. The AI can provide a first round of comprehensive feedback, edited by the instructor if needed, allowing instructors to focus on higher-level guidance or areas the AI might have missed. Furthermore, providing this level of feedback from one professor for all 20, 40, 60, or more students would be extremely time consuming.

✨Pedagogical Suggestions

Here are some practical approaches for integrating this technology into your teaching, including some $0 AI models that are comparable to o1-pro (and how to access them):

Note: This rest of this section is visible to only ✨Premium subscribers. Thanks for your support! (If it’s cut off due to length, view the piece online via the “Read Online” link at the top of the email.)

✨ What Premium Members Get

As I improve ✨Premium subscribers’ benefits, it can be hard to keep track, especially if you’re new to AutomatED. Here’s a summary…

Core Benefits

  • $0 monthly webinars on timely AI topics (last one: Feb. 28, on “Using AI to Create Course Content”; next one: on using AI for non-pedagogical productivity, at the end of March), while everyone else pays $25

  • Access to all our practical Guides, including topics like AI-aware assessment design and student AI policies, and Tutorials on how to use AI tools, from using ChatGPT for analyzing teaching data to using LearnLM for tutoring

  • Discounts on our focused courses (like our new custom GPT course)

  • Annual 1-hour consultation to discuss your specific challenges

  • Premium-only sections in weekly newsletters, like the above

You can see all the benefits, including all the Guides and Tutorials, here.

Nearly 100 professors and learning specialists already use these resources to save time and enhance their teaching. Want to join them?

What'd you think of today's newsletter?

Login or Subscribe to participate in polls.

Graham

Let's transform learning together.

If you would like to consult with me or have me present to your team, discussing options is the first step:

Feel free to connect on LinkedIN, too!