graph LR
A[10 examples<br/>Test concept] --> B[100 examples<br/>Refine approach]
B --> C[300 examples<br/>Validate]
C --> D[18,000 projects<br/>Full scale]
Practical Strategies for Using LLMs in Research
2025-07-15
Teal Emery | July 15, 2025

April 2025: Published “Greener on the other side?”
The challenge: Classify 18,000 Chinese lending projects - 🟢 Green | 🟫 Brown | 🔘 Grey | ⚪ Neutral
Traditional: 1,500 hours, $22,500
Our result: 15 hours, $1.58
Why I’m here: Share practical lessons learned
Despite “Green Belt and Road” rhetoric, minimal renewable energy financing
Key finding: 5.8% of Chinese lending went to green projects
The point: This analysis was only possible because LLMs handled the routine classification
Current roles:
Relevant experience:
Philosophy: Build in the open, amplify expertise
Yes, there are problems ⚠
Despite this… 💡
My value add: Practical experience on what works (and what doesn’t)
Each builds on the last
Why AI works this way + strategic thinking
Hands-on practice with immediate applications
From web interfaces to APIs - the China case study
Action planning & Q&A
Today’s Workshop
Companion Book
📖 Full guide: teal-insights.github.io/soas_llm_training
What we’re building: Strategic thinking about AI capabilities
Information overload is real:
💬 Chat: Share one routine task that consumes too much of your time
The opportunity: What if AI could handle the routine so you focus on insight?
Web Interfaces (Our focus)
APIs (Advanced)
Today: Web interfaces | Case study: How APIs enabled our China project
AI capabilities form an irregular, unpredictable boundary
Key insight from Ethan Mollick’s research:
“Paint a dumpster fire in Monet’s style”

“Make a better jagged frontier diagram”

The lesson: You must explore the frontier for YOUR tasks
Where AI consistently saves researchers time:
Literature Tasks 📚
Coding Tasks 💻
These are INSIDE the frontier - reliable time-savers
Think of AI as an infinitely patient research assistant
You Provide
AI Provides
Key principle: AI amplifies expertise, doesn’t replace it
Three major providers, each with strengths:
Context window = Working memory
For academic work: Gemini’s massive context + citations = best choice
Traditional software: Predictable rules, consistent outputs
LLMs: More like a knowledgeable but quirky colleague
This is why prompting matters - it’s how you communicate effectively
✓ Jagged frontier concept
✓ Where AI excels (literature, coding)
✓ Why collaboration works
✓ Tool options
What we’re building: Practical tools you’ll use immediately
LLMs are powerful but not clairvoyant
They need:
Analogy: Like briefing a new research assistant who’s brilliant but knows nothing about your work
CONTEXT: Background information, your expertise, project details
TASK: Specific request with clear action verbs
FORMAT: Structure, length, style preferences
CONSTRAINTS: What to avoid, limitations, requirements
QUALITY CONTROLS: "Ask clarifying questions", "Tell me when unsure"Each element improves output quality
The problem: Starting from scratch every conversation
The solution: Reusable professional context that captures:
Why this matters: Foundation for all future AI interactions
Analyze my professional background and create a 200-300 word
professional context for AI interactions including:
- Domain expertise & background
- Current roles & affiliations
- Research focus areas
- Communication style & audience
- Technical approaches I use
Make it professional but conversational.Step by step:
Success indicator: AI understands your research domain and communication needs
✅ Professional context that:
This becomes part of EVERY future prompt
Now we combine context with structured requests
Without structure ❌ “Help with literature review”
Vague → Generic output
With structure ✅ Context + Task + Format
Specific → Useful output
The magic: Your context + clear structure = targeted assistance
Two powerful additions to any prompt:
“Ask me clarifying questions”
“Tell me when you’re unsure”
These simple phrases dramatically improve reliability
Build on Activity 1 by adding structure:
Try it with a real research need:
💬 Chat: What surprised you about the enhanced response?
✅ What you’ve discovered:
Next: Apply this to your biggest time sink - literature review
The universal research challenge:
Initial literature review is prime “low-hanging fruit” - squarely inside AI’s capabilities
For every potentially relevant paper:
[Professional context]
Please provide the full citation information for this document at the top, then create a detailed structured summary of this entire document, including any appendices.
I need to understand:
- The main argument or purpose
- Key findings or conclusions
- Important data, evidence, or examples
- Any policy implications or practical applications
- Who the intended audience appears to be
Format this as a structured summary with clear headings.Result: 5% merit deep reading, 95% the summary is sufficient
Language Translation 🌍
Jargon Translation 🔄
Both expand your research scope dramatically
[Your context]
This computer science paper uses unfamiliar technical terminology.
Please:
1. Identify the main contribution in plain language
2. Explain how their methods might apply to social science research
3. Highlight any useful techniques for [your research area]
Focus on practical applications, not technical details.Unlock insights from other fields
What are Gems?
Think of it as: Training a specialized research assistant who remembers your preferences
Part A: Test & Refine (10 min)
Part B: Create Gem (15 min)
This combines everything you’ve learned
# Literature Review Assistant - [Your Name]
[Your professional context from Activity 1]
## Your Role
Help me efficiently analyze papers in [your field], identifying
insights relevant to [your research focus].
## For every document:
- Full citation
- Main argument + theoretical framework
- Methods and evidence
- Key findings with page numbers
- Implications for [your specific interests]
- Connections to other disciplines
## Quality Standards
- Distinguish claims from evidence
- Note limitations and biases
- Flag surprising or controversial points
- Suggest related papers if mentionedPart A: Test the workflow
Part B: Create your Gem
✅ You’ve built a complete system:
Each element reinforces the others
You now have a literature review assistant tailored to YOUR research
Both save massive time on routine tasks
Who this helps:
The opportunity: AI as infinitely patient coding tutor and assistant
AI Excels At ✅
AI Struggles With ❌
Success key: You provide research logic and context, AI handles syntax
Promise 🚀
Peril ⚠️
Critical: You need to understand your code. Always validate results.
The dangerous phenomenon where:
Example: Statistical test that always returns p < 0.05
Protection:
[Your context including preferred languages/packages]
I have survey data with columns: household_id, income, education,
region, year. The data has some missing values.
Using R with tidyverse style and base R pipe |>, please:
1. Create a script to calculate and visualize median income by region and education
2. Create plots to help me understand how to handle missing data appropriately
4. Include comments explaining each step
Note: I prefer functional programming (purrr) over loops.When you’re ready for more:
Cursor
GitHub Copilot
Start with chat interfaces, graduate to these when ready
✓ Mental model (jagged frontier)
✓ Three practical tools
✓ Understanding of low-hanging fruit
What we’re exploring: How these tools enable ambitious research
Research question: Given policy chatter about the “Green Belt & Road,” what do we know about China’s role in funding the energy transition in developing countries?
The data: 18,000 lending projects needing classification
The constraint: Ambitious project, limited budget
What you’ve learned 🖥️
What we used 🔧
Same AI, different access method
1,500 hours × $15/hour = $22,500
15 hours × $0.11/hour = $1.58
Real example: “500MW solar power plant with backup diesel generator”
Keyword approach ❌
LLM approach ✅
This is the jagged frontier in action
graph LR
A[10 examples<br/>Test concept] --> B[100 examples<br/>Refine approach]
B --> C[300 examples<br/>Validate]
C --> D[18,000 projects<br/>Full scale]
Key insight: Start small, validate early, scale confidently

Our approach:
Transparency enables trust and progress
China’s green lending: mostly hydro, minimal solar/wind
Only possible through scale - patterns invisible in small samples
Stay with web interfaces when:
Consider APIs when:
Most research stays in the first category
You now have the tools
✅ Mental Models
✅ Three Tools
✅ Understanding of scale
You’re ready to start integrating AI into your research
Choose 2-3 that fit your current work:
💬 Chat: Which actions match your immediate needs?
Your reference book: teal-insights.github.io/soas_llm_training
Essential reading:
Stay connected:
Remember:
The opportunity: More ambitious research while maintaining rigor
The tools: You now have them
Questions & Discussion

SOAS Faculty Workshop • Detailed Guide: teal-insights.github.io/soas_llm_training