VToonify: Controllable high-resolution portrait video style transfer

by godmode2019on 1/15/2023, 2:02 AMwith 22 comments

by viraptoron 1/15/2023, 3:06 AM

Also reviewed at Two minute papers with a few examples https://youtube.com/watch?v=C9LDMzMRZv8 (hold on to your papers...)

by nineteen999on 1/15/2023, 5:35 AM

Well it's cool and all, but the results sit right at the very deepest part of the uncanny valley.

by morjomon 1/15/2023, 11:45 AM

I wonder what the implementation into e.g live streaming would require.

by marcAKAmarcon 1/15/2023, 4:36 PM

I'm glad things are progressing, but it bugs me that AI is largely being innovated for the use of... things like this? I know this comment is a bit disparaging and minimizes greater achievements, and I apologize for that, but the closeness of content-consumerism and AI is becoming quite off putting.

by techdragonon 1/15/2023, 3:43 AM

Based on the numbers in the paper this is just a little bit too slow for use as a real time video effect. At ~0.1 seconds per frame we just need about a 3x improvement in performance to get to 30fps “real time” video frame rates.

And on that thought since it appears they used nVidia hardware based on the CUDA dependency, it would be interesting to see how this performs on something like an M1/M2 where there’s dedicated ML hardware to help offload and accelerate things.