• 0 Posts
  • 1.17K Comments
Joined 2 years ago
cake
Cake day: August 27th, 2023

help-circle




  • Due to some disagreements—some recent; some tolerated for close to 2 decades—with how collaboration should work, we’ve decided that the best course of action was to fork the project

    Okay, that was always allowed!

    Programming is the weirdest place for kneejerk opposition to anything labeled AI, because we’ve been trying to automate our jobs for most of a century. Artists will juke from ‘the quality is bad!’ to ‘the quality doesn’t matter!’ the moment their field becomes legitimately vulnerable. Most programmers would love if the robot did the thing we wanted. That’s like 90% of what we’re looking for in the first place. If writing ‘is Linux in dark mode?’ counted as code, we’d gladly use that, instead of doing some arcane low-level bullshit. I say this as someone who has recently read through IBM’s CGA documentation to puzzle out low-level bullshit.

    You have to check if it works. But if it works… what is anyone bitching about?










  • I would be shocked if any diffusion model could do that based on a description. Most can’t overfill a wine glass.

    Rendering over someone demonstrating the movement, as video-to-video, is obviously easier than firing up Blender. But: that’s distant from any dream of treating the program like an actress. Each model’s understanding is shallow and opinionated. You cannot rely on text instructions.

    The practical magic from video models, for the immediate future, is that your video input can be real half-assed. Two stand-ins can play a whole cast, one interaction at a time. Or a blurry pre-vis in Blender can go straight to a finished shot. At no point will current technologies be more than loose control of a cartoon character, because to these models, everything is a cartoon character. It doesn’t know the difference between an actor and a render. It just knows shinier examples with pinchier proportions move faster.







  • I fully endorse photorealistic cartoons. Characters can look like anything, without having to find a specific guy and hope fits the role.

    Getting the cartoon onscreen can still involve an actual actor. Diffusion turns whatever you have into whatever you describe. Turning a guy into another guy is not a big ask. It’s how some of this mess started, with Nicolas Cage deepfaked as Superman, and defictionalizing the Stallone version of The Terminator. The target face does not need to be a real person. Three actors can stage nearly any script.

    Same goes for voice-acting. VAs are understandably concerned about being cloned. Nobody’s talking about the opposite: making up what characters sounds like, so any actor can play anybody. Or everybody. You can even substitute, when a scene needs extra oomph - like a band featuring a guitarist for a solo. Same sound… distinct performance.