Exploring patterns in solo work, small teams, and collaboration | January 2026
About This Research
What this document explores: Patterns in when solo/remote work succeeds vs. when teams/co-location appear beneficial—looking at historical examples, technology companies, and specifically AI research. Includes profiles of researchers, analysis of 20+ AI papers by team size, and frameworks for thinking about your own situation.
Key observations from the data:
Many conceptual breakthroughs came from small teams (2-4 people). LSTM had 2 authors, Attention had 3, GPT-1 had 4. Often traceable to 1-2 key individuals.
Team sizes have grown over time, particularly for scaling work. GPT-3's 14 authors reflects engineering scale, not 14 people having the original idea.
The type of work seems to matter more than blanket rules. Long feedback loops, objective verification, and deep focus appear to favor solo/remote. Coordination-heavy work appears to favor teams.
Research is increasingly distributed. Many successful AI labs operate remotely or with distributed teams.
AI tools are changing what's possible for individuals. The surface area of solo-viable work appears to be expanding.
Historical Patterns of Solo & Remote Success
Work Characteristics That Appear to Enable Solo Success
Characteristic
Potential Reason
Examples
Long feedback loops
Less need for real-time coordination
Darwin (20 years), Einstein (8 years)
Objective verification
Work can speak for itself
Mathematics, code, experiments
Deep focus as bottleneck
Interruptions may be costly
Writing, research, complex coding
Symbolic/digital output
Easy to share remotely
Papers, software, proofs
Where Solo Work Has Succeeded
Mathematics: Ramanujan, Perelman, Grothendieck
Theoretical Physics: Einstein's miracle year, Newton during plague
Writing: Most books
Open Source: Linus wrote Linux in his bedroom
Where Teams Appear Necessary
Experimental Physics: CERN, Manhattan Project
Large-Scale Engineering: Bridges, rockets, OS at scale
PATTERNS OBSERVED
─────────────────────────────────────────────────────────────
Solo/Remote tends to appear in: Teams tend to appear in:
───────────────────────────────── ────────────────────────────
Long feedback loops Short feedback loops
Objective verification Subjective/political decisions
Deep focus is bottleneck Coordination is bottleneck
Symbolic output Physical output
Exploration phase Execution phase
Novel ideas Scaling ideas
Remote-First Companies — Evidence
Successful Remote-First Companies
Company
Type
Remote Status
Valuation
Shopify
Public
"Digital-by-default" (May 2020)
~$188B
Coinbase
Public
Remote-first, no HQ
~$61B
GitLab
Public
All-remote from founding
~$5.6B
Automattic
Private
Remote-first since founding
~$7.5B
Deel
Private
Remote work infrastructure
~$17.3B
PostHog
Private
100% remote, 20+ countries
$1.4B
Supabase
Private
Born-remote
$5B
Linear
Private
Fully remote from start
$1.25B
Research on Remote Work & Innovation
Nature study (2023): Analyzed 20M papers + 4M patents. Found remote teams were less likely to produce breakthrough discoveries vs on-site teams in their dataset.
One interpretation: Remote teams may do more technical work while in-person teams do more conceptual work.
Counterpoint (Oxford 2022): Long-distance collaboration correlated with breakthroughs post-2010, possibly due to better collaboration tools.
Important Context
These studies are pre-AI coding tools. Human + AI collaboration may change these dynamics significantly.
AI Research — Observed Patterns
The 1 → Few → Many Trajectory
A pattern that appears repeatedly in AI research:
Individual vision/prototype — often solo or small team
Small team validates and refines — 2-6 people
Scale execution — 10-30+, where coordination becomes important
OBSERVED PATTERNS IN AI WORK
───────────────────────────────── ────────────────────────────
Often done by small teams: Often requires larger teams:
───────────────────────────────── ────────────────────────────
Interpretability research Frontier model training
Safety/alignment theory Large-scale data operations
Fine-tuning & applications Infrastructure at scale
Open source contributions Regulatory/policy navigation
Novel architectures (early) Multi-modal integration
Blog posts that build reputation Production deployment
Potentially Solo-Friendly Opportunities in AI
Mechanistic interpretability — Chris Olah's reputation came from solo blog posts
Fine-tuning and applications — Accessible without training frontier models
Safety research — May favor deep thinking over coordination
Open source tools — Karpathy's nanoGPT, Hotz's tinygrad
Evaluations and benchmarks — High leverage, doesn't require massive compute
Novel research directions — Early exploration may not need teams
AI Breakthroughs — Team Size Analysis
Major Breakthroughs by Team Size
Breakthrough
Year
Authors
Size
Key Driver(s)
Backpropagation
1986
3
Small
Hinton
LSTM
1997
2
Duo
Hochreiter & Schmidhuber
AlexNet
2012
3
Small
Krizhevsky
Attention
2014
3
Small
Bahdanau
GANs
2014
8
Medium
Goodfellow
Batch Normalization
2015
2
Duo
Ioffe & Szegedy
ResNet
2015
4
Small
Kaiming He
Transformers
2017
8
Medium
Equal contributors
GPT-1
2018
4
Small
Radford
BERT
2018
4
Small
Devlin
GPT-2
2019
6
Medium
Radford
DDPM (Diffusion)
2020
3
Small
Ho
Scaling Laws
2020
9
Medium
Kaplan/McCandlish
GPT-3
2020
14+
Large
Brown/team
2-4
Typical team size for early conceptual work
6-14+
Typical team size for scaling work
98K+
LSTM citations (2 authors)
173K+
Transformer citations (8 authors)
An Interesting Pattern
Early era (1986-2000): Ideas and people seemed to be the constraint → small teams
Modern era (2018+): Compute and engineering seem to be constraints → larger teams for execution
Observation: Team size growth appears more correlated with implementation and scaling than with ideation. The conceptual breakthrough stage often remains small-team territory.
How Team Sizes Have Evolved
Era
Typical Team Size
Primary Bottleneck
1986-2000
1-3
Ideas, people
2009-2014
3-6
Ideas, data
2014-2018
4-8
Ideas, compute
2018-2020
6-14
Compute, engineering
2020+
10-100+ (industry)
Scale, deployment
Researcher Profiles
Alec Radford Independent
Sam Altman called him "Einstein-level genius" — key figure behind GPT
Work Style
Independently built GPT-1
"Hands-on, experimental approach"
Marc Andreessen: "hiding in the corner of AI, quietly researching"
Rarely gave interviews or public appearances
Key Contributions
Paper
Year
Team
Role
GPT-1
2018
4
First author, built independently
GPT-2
2019
6
Co-first author (with Wu)
CLIP
2021
12
First author
Whisper
2022
6
First author
Current: Left OpenAI Dec 2024. Independent researcher collaborating with multiple AI orgs.
Pattern
Stayed hands-on throughout career. Chose autonomy over institutional resources.
Chris Olah Technical Leader
Pioneer of mechanistic interpretability — TIME 100 Most Influential in AI 2024
Work Style
No undergrad degree (Thiel Fellowship)
"Empirical scientist at heart"
Personally reverse-engineers neural networks
Famous for solo blog posts (colah.github.io)
"We can understand them so well that we can go and hand-write weights."
Reputation Built From Solo Work
"Understanding LSTM Networks" (blog)
"Calculus on Computational Graphs" (blog)
"Visual Information Theory" (blog)
Current: Co-founder at Anthropic. Leads interpretability. Still hands-on.
Pattern
Technical credibility came from solo blog posts. Still codes as a co-founder.
Ilya Sutskever Research CEO
Co-founder & Chief Scientist at OpenAI → CEO of SSI
Work Style
Set research direction AND contributed technically
Name on GPT-1, GPT-2, GPT-3, DALL-E papers
The "scaling believer" — pushed the paradigm before formal proof
"The age of scaling is ending... the next breakthroughs now depend on new learning methods, not more GPUs."
Current: SSI (July 2025)
~20 employees, ALL researchers/engineers
No sales, marketing, product managers
$32B valuation
Pattern
Built a structure where CEO = researcher. The entire company is a research lab.
Dario Amodei Full CEO
VP of Research at OpenAI → CEO of Anthropic
At OpenAI
"Collaborating with the team to build GPT-2 and GPT-3"
Maximum leverage path. Uses technical credibility but doesn't do research anymore.
Frontier/Adjacent Researchers
Michael Edward Johnson FrontierIndependent
Consciousness researcher — Co-founded Qualia Research Institute, now Symmetry Institute
Work Style
Almost entirely funded by retroactive donations
Publishes on opentheory.net and Substack
Works with small, distributed teams
No traditional academic position
Active in LessWrong/EA communities
Key Contributions
Principia Qualia (2016) — full-stack research paradigm for consciousness
Symmetry Theory of Valence — valence determined by symmetry of information geometry
Neural Annealing Theory (2019) — unified theory of therapeutic states
AI Consciousness Paradigm (2024) — framework for machine consciousness
Pattern
Demonstrates independent consciousness research is viable. Built reputation through long-form writing and frameworks, not traditional academia.
Michael Levin FrontierAcademic Lab
Bioelectricity & collective intelligence — Director of Allen Discovery Center at Tufts
Work Style
350+ peer-reviewed papers
Multi-institutional research groups
Combines biology, computer science, and cognitive science
Long-term collaborations that deepen over time
Active science communicator (TED talks, podcasts)
Key Contributions
Molecular basis of left-right asymmetry
Bioelectric language of development — voltage gradients drive patterning
Xenobots (2020) — living robots from frog cells designed by AI
Scale-free cognition — intelligence exists at all biological scales
AI Connection
Uses evolutionary algorithms to design organisms
Proposes all intelligence is computational (substrate-independent)
Xenobots were "designed in a computer before being constructed biologically"
Pattern
Shows how frontier research can span biology, AI, and philosophy. Academic lab but intellectually operates like an independent researcher pursuing unconventional directions.
The Researcher Spectrum
PURE RESEARCHER ←――――――――――――――――――――――→ PURE CEO
Radford
Independent researcher
Olah
Technical leader who still codes
Ilya
CEO of research-only startup
Dario
CEO of large company (700+)
Different researchers have found different points on this spectrum. There's no single "right" path — each represents a different tradeoff between hands-on work and organizational leverage.
The Scaling Laws Origin Story
Who Discovered Scaling Laws?
The foundational paper (January 2020) formalized what OpenAI had believed intuitively.
Lead authors: Jared Kaplan and Sam McCandlish
Interesting note: Both were theoretical physicists, not traditional ML researchers.
Author List
Jared Kaplan (Johns Hopkins, physics PhD)
Sam McCandlish (physics background)
Tom Brown, Benjamin Chess, Rewon Child, Scott Gray
Alec Radford, Jeffrey Wu, Dario Amodei
Why Physicists?
Physicists are trained to look for universal scaling relationships. They may have seen AI capabilities vs compute as a physical phenomenon with predictable laws — similar to how physical systems follow power laws.
Most of this team later left to co-found Anthropic.
GPT Timeline
June 2017
Transformer paper — New architecture for translation. The LLM paradigm wasn't yet obvious.
GPT-2 — OpenAI called it "too dangerous to release." 1.5B parameters.
January 2020
Scaling Laws paper — Formal proof that scale = predictable capability gains.
June 2020
GPT-3 — 175B parameters. The paradigm became widely accepted.
Radford's Core Insight
"The underlying generative model learns to perform many of the tasks we evaluate on in order to improve its language modeling capability."
Translation: The model learns QA, translation, summarization as a side effect of predicting the next word. This was a conceptual breakthrough. Scale came later.
Location Considerations
What Tech Hubs Like SF Provide
Potential Advantage
Available Remotely?
Investor access
Partially — remote fundraising is more common now
Talent density
Yes — hire globally
Serendipitous collisions
Harder to replicate remotely
Compute access
Yes — cloud, partnerships
Credibility signaling
Partially
You may not need a hub if:
You have a clear research direction
You have compute access
You have 1-2 high-bandwidth collaborators
You're building reputation through artifacts (papers, code, blog posts)
A hub might help if:
Raising significant capital
Recruiting from existing talent pool
Early-career, need to build network
Your work benefits from frequent in-person collaboration
AI as Force Multiplier
How AI Tools Are Changing Individual Capacity
Old Constraint
How AI May Change It
Coding bandwidth
Cursor, Copilot, Claude Code
Design bandwidth
Midjourney, Figma AI
Research bandwidth
Literature review, synthesis
Ops bandwidth
Automated monitoring, agents
Potential implication: Work that previously required 5-10 people may now be achievable by 1-2 people. This could dramatically expand the viable solo/remote surface area.
The Indie Research Movement
EleutherAI — open source collective producing research
Nous Research — small team, notable impact
Individual researchers with compute grants
arXiv making distribution free
Framework for Thinking About Solo/Remote Viability
Questions to Consider
Question
Solo/Remote May Work Better
Teams May Work Better
How long is the feedback loop?
Days/weeks/months
Hours/minutes
How is quality verified?
Objectively (tests, math)
Subjectively (opinions)
What's the bottleneck?
Deep thinking
Coordination
What's the output?
Digital/symbolic
Physical
What phase is the work?
Exploration
Execution at scale
Potentially high solo/remote viability:
Theoretical research
Writing (books, papers, code)
Design (once vision is clear)
Early-stage product development
Open source maintenance
Content creation
Potentially lower solo/remote viability:
Large-scale engineering
Experimental physics (equipment)
Team sports and performance
Manufacturing
Emergency response
Key Observations
1. Solo/small-team research has produced major work
Radford built GPT-1 independently. Olah's reputation came from solo blog posts. Johnson operates entirely outside academia.
2. Many conceptual breakthroughs came from small teams (2-4)
LSTM (2), Attention (3), AlexNet (3), GPT-1 (4), Diffusion (3). Often traceable to 1-2 key individuals.
3. Team sizes appear to grow for execution
GPT-3's 14 authors reflects scaling/engineering work. Conceptual breakthrough stage often remains small-team territory.
4. Research is increasingly distributed
Many successful AI labs operate with distributed or remote teams. Location may matter less than it once did.
5. The 1 → few → many pattern appears often
Individual vision, small team validation, large team scaling.
6. Some researchers stay hands-on at senior levels
Radford and Olah demonstrate this is possible, but it requires deliberate choice.
7. Frontier research happens in unexpected places
Michael Levin (bioelectricity), Michael Johnson (consciousness) are doing solo/small-team work that intersects with AI from outside traditional AI.
8. Work characteristics may matter more than industry norms
Long feedback loops, objective verification, and deep focus as bottleneck appear to favor solo work.
9. AI tools are expanding what's solo-viable
The capabilities available to individual researchers are increasing rapidly.
10. Distribution channels are diversifying
Blog posts (Olah), open source (Karpathy), Substack (Johnson) can build reputation without traditional publishing.
Summary
This research suggests that for certain types of work, solo or small-team approaches may be viable or even advantageous. The key factors appear to be the nature of the work itself — particularly feedback loop length, verification method, and whether the bottleneck is thinking or coordination. As AI tools continue to develop, the range of work that can be done by individuals or small teams may expand further.