An AI Intern That Runs Your Entire Post-Training Pipeline -- ml-intern on PH
HuggingFace's open-source agent automates lit scans, dataset discovery, training, eval, and iteration. 365 upvotes on Product Hunt. Free and Apache-2.0.

365 Upvotes for "Automate 80% of ML Research"
Launched April 23 on Product Hunt. Maker: HuggingFace. Free, open-source (Apache-2.0).
ml-intern is an agent that automates the entire LLM post-training pipeline. Tell it "improve scientific reasoning" and it searches papers, finds datasets, writes training scripts, trains the model, evaluates results, and iterates. It scored GPQA 32% in 10 hours -- beating Claude Code's 22.99% -- with zero human intervention.
On GitHub, it's at 6,800 stars and climbing 260 per day.
What It Does
ml-intern's automated post-training workflow
It replaces the repetitive parts of ML research. Built on HuggingFace's smolagents framework with native integration across Transformers, TRL, and Datasets.
The pitch: give it a goal, it handles the rest. Paper search (arXiv, Semantic Scholar) -> dataset discovery (HuggingFace Hub) -> training script generation (TRL) -> model training -> benchmark evaluation -> improvement iteration. Full cycle, no human in the loop.
First Impressions
PH comments from ML researchers are enthusiastic. "Why didn't this exist sooner" and "better than an actual intern" are common reactions. People already in the HuggingFace ecosystem especially appreciate the near-zero adoption cost.
The concern: can you blindly trust the results? Fair point -- the datasets and hyperparameters the agent chooses still warrant human review.
Three Key Features
1. End-to-End Pipeline. Paper search through model evaluation in a single command.
2. HuggingFace-Native Integration. Works seamlessly with the full stack -- Transformers, TRL, Datasets, Hub. No extra configuration.
3. Automated Iteration. If evaluation results fall short, the agent automatically runs improvement cycles without waiting for human input.
Pricing
Free. Open-source (Apache-2.0). GPU costs are on you.
Who Benefits
- ML researchers: Automate repetitive experiment setup and training loops
- AI startups: Amplify research capacity on small teams
- Grad students: Explore multiple experimental directions in parallel
Similar Tools
ml-intern GitHub repository main page
- SWE-agent: Automates code bug fixes. Coding, not training.
- STORM: Automates paper writing. Writing, not experiments.
- Hermes Agent: General-purpose self-improving agent. Not ML-specific.
They named it "intern," but this thing delivers senior-level output.
References
관련 기사

GPQA 32% in 10 Hours -- HuggingFace's AI Intern Outperformed Claude Code
An open-source agent that automates the entire LLM post-training pipeline: lit scan, dataset discovery, training scripts, eval, and iteration. 6,800 stars, growing 260/day.

OpenClaw — Why a Local AI Assistant Hit 250K Stars on GitHub
No cloud, no data leaving your device. Connects 50+ platforms including WhatsApp, Telegram, Slack, and iMessage. A weekend project became one of the fastest-growing open-source repos in GitHub history.

95,600 Stars in 7 Weeks -- Nous Research Built an Agent That Improves Itself
Hermes Agent ships a reflection loop, trace-based RL fine-tuning, and multi-LLM routing out of the box. At 1,500 stars per day, it's the fastest-growing agent framework on GitHub.
AI 트렌드를 앞서가세요
매일 아침, 엄선된 AI 뉴스를 받아보세요. 스팸 없음. 언제든 구독 취소.
