Skip to content

sugarkwork/sd_expressions_gen

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

32 Commits
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

sd_expressions_gen

This script automatically generates multiple facial expressions for a character by executing the API of the Stable Diffusion Web UI.

Right now, it's just a matter of calling the API of the Web UI from the script. I would like to implement it as an extension of the Stable Diffusion Web UI in the future.

To configure Stable Diffusion Web UI, edit the "webui-user.bat" file and specify/add "--api" to the COMMANDLINE_ARGS.

@echo off

set PYTHON=
set GIT=
set VENV_DIR=
set COMMANDLINE_ARGS=--api
call webui.bat

Launching StableDiffusion.

Open a command line. Install the required modules as follows:

pip install requests pillow PyYAML

git clone.

git clone https://github.com/sugarkwork/sd_expressions_gen
cd sd_expressions_gen

When you run the script, it automatically starts writing a character using the StableDiffusion API. It saves a file for each facial expression of the character.

python expressions_gen.py

config

Specify the base prompt in prompt in config.yaml. You can use components defined in StableDiffusion, such as LoRA and Embedding.

prompt: "1girl, tareme,  pink hair, twintails, red eyes, hime cut, masterpiece, high quality, 
cute, highres, delicate, beautiful detailed, finely detailed, front light, white background, 
standing, sfw, looking at viewer, (upper body:1.1)"

All information such as the step number is stored in the payload variable.

A list of facial expressions is defined in the "expressions" variable. If there are unnecessary facial expressions, you can delete or add them. The key will be the filename and the value will be the actual prompt.

process

This code functions as follows.

  1. Load the image (cn.png) to be set as the input for ControlNet.
  2. Generate a character based on the specified prompt.

expressionless

  1. Blur the face of the generated character according to the mask.png.

br

  1. Feed the blurred image into ControlNet and perform i2i to change the expression prompt.

smile

Repeats the process for the number of expressions defined in the "expressions" variable.

If you prepare a pose image (cn.png) for ControlNet and a face mask (mask.png), you can easily create a standing image that can be used in the game.

facial expressions sample

https://ai.sugar-knight.com/aitest/ex.html

Using ChatGPT, I asked them to come up with lines that sounded like that. The HTML was also made by ChatGPT.

a list of currently defined facial expressions

  • smile
  • crazy
  • sleepy
  • painful
  • endure pain
  • evil smile
  • heart
  • angry
  • happy
  • sad
  • crying
  • bawl
  • confused
  • scared
  • surprised
  • laughing
  • closed eyes
  • half-closed eyes
  • excitement
  • shy
  • shouting
  • kiss
  • ahegao soft
  • ahegao hard
  • ahegao with heart
  • wink
  • get drunk

why did i make this

If we were to create a game in which AI dynamically prepares scenarios and character images, we thought it would be better to automatically generate the character's facial expressions. We verified the implementation of character facial expressions using only the API without going through the Web UI.

future plans

  • Automatically create a face mask image (mask.png) from an image for ControlNet.
  • Create a costume change sample.
  • Make it an extension of the web UI.
  • Randomly select an image to use for ControlNet.
  • Switch API. Web UI API deprecates /controlnet/txt2img. But in my environment ControlNet with /sdapi/v1/txt2img did not work.