Learning to Paint With Model-based Deep Reinforcement Learning
About
We show how to teach machines to paint like human painters, who can use a small number of strokes to create fantastic paintings. By employing a neural renderer in model-based Deep Reinforcement Learning (DRL), our agents learn to determine the position and color of each stroke and make long-term plans to decompose texture-rich images into strokes. Experiments demonstrate that excellent visual effects can be achieved using hundreds of strokes. The training process does not require the experience of human painters or stroke tracking data. The code is available at https://github.com/hzwer/ICCV2019-LearningToPaint.
Zhewei Huang, Wen Heng, Shuchang Zhou• 2019
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Image Inpainting | FFHQ (test) | LPIPS0.697 | 54 | |
| Neural Painting | Landscapes (test) | Pixel Loss0.055 | 18 | |
| Neural Painting | Wiki Art (test) | Pixel Loss0.047 | 18 | |
| Semantic Alignment | Oil Painting Reference Images | CLIP Score (cap.1)0.1879 | 7 | |
| Stroke-based Rendering | Gallery Dataset | PSNR28.58 | 6 | |
| Stroke-based Rendering | User Study | Structure Score3.51 | 6 | |
| Stroke-based Rendering | DIV2K (val) | PSNR27.19 | 6 | |
| Painting Quality Evaluation | Human Evaluation 51 participants (test) | Style Score3.09 | 6 |
Showing 8 of 8 rows