CivArchive
    Roukin8-character-LoHa/LoCon/FullCkpt ろうきん8 - Full Checkpoint
    NSFW
    Preview 216885
    Preview 216890
    Preview 216889
    Preview 216887
    Preview 216886

    This post contains three types of fine-tuned results with the same dataset: LoHa, LoCon, and full checkpoint (all with clip skip 1).

    Associated hugging face repository: https://huggingface.co/alea31415/roukin8-characters-partial (more versions can be found there).

    Note that the models are fine-tuned on top of BP, so you get the style I show below by using BP. Using these models on orange could kill the orange style unless using with a properly adjusted weight.

    What is LoHa/LoCon?

    LoHa: Lora with hadamard decomposition (ref https://arxiv.org/abs/2108.06098)

    LoCon: Lora extended to residual blocks

    See https://github.com/KohakuBlueleaf/LyCORIS/blob/lycoris/Algo.md for more explanation

    How to use?

    For webui install https://github.com/KohakuBlueleaf/a1111-sd-webui-locon and use it as a lora

    Why LoHa?

    The following comparison should be quite convincing.

    (Left 4 LoHa with different number of training steps; Right 4 same thing but for LoCon; same seed same training parameters for the two training)

    In addition to the five characters I also train a bunch of style into the model (what I have always been doing actually).

    However, LoRa and LoCon do not combine styles with characters that well (characters are only trained with anime images) and this capacity gets largely improved in LoHa.

    Note that the two files have almost the same size (around 30mb). For this I set (linear_dim, conv_dim) to (16,8) for LoCon and (8,4) for LoHa. However with Hadamard product the resulting matrix could now be of rank 8x8=64 for linear layers and 4x4=16 for convolutional layers.

    P.S. Note that the 20000 step files are provided here so if you want other version please check hugging face

    How about full checkpoint?

    It is still the most performing as illustrated below. Maybe we can get such results with LoHa with larger dimension, better decomposition of the convolutional layers, and by adding back the few layers that are not yet currently trained in the LoCon/LoHa implementation. Let's see.

    Training details

    • LoCon/LoHa: alpha 1, dimensions as specified above, batch size 8, learning rate 2e-4 throughout with constant scheduler, AdamW

    • Full checkpoint: batch size 8, learning rate 2.5e-6 with cosine scheduler, Adam8bit, conditional dropout 0.08

    Example Images

    Images 1 and 4 come from LoCon, 2 and 3 come from LoHa, and 5 and 6 come from full checkpoint.

    Dataset

    I have uploaded the anime part dataset of 3000 anime screenshots. The current format works with Everydream with the multiply.txt and for kohya trainer you should use this script https://github.com/cyber-meow/anime_screenshot_pipeline/blob/develop/utilities/flatten_folder.py

    python flatten_folder.py --src_dir /path/to/dir

    Description

    FAQ

    Comments (7)

    HosiokaMar 9, 2023
    CivitAI

    May I take a look at your config and dataset for LoHA? Im currently experimenting with LoCon aka(LyCORIS).

    If you don't mind sharing them pls do! I could use some research data

    alea31415
    Author
    Mar 9, 2023

    Config is pretty much explained in the description. Do you need any other information?

    As for dataset, there are a bunch of things. For the characters I use 3000 anime screenshots that I just uploaded.

    I however refrain from uploading the works of specific artists but I just grab them from danbooru with tags using imgbrd-grabber and add artist tag at the beginning while dropping character and copyright tag. There are from hundreds to thousands of images for each trained style.

    The total dataset size is of 22000 because I also have other regularization images.

    I use different repeats in different concepts following my own workflow https://github.com/cyber-meow/anime_screenshot_pipeline so taking into account the repeats I train ~20000 steps per epoch

    HosiokaMar 9, 2023

    @alea31415 Oh wow thats quite a big dataset for LoRAs,LoCon(LyCORIS) models the most i've tried is around 700 the thing is though i never got eyes to be correct when training and i still dont understand why do u happen to know? i use and LR of 1e-4 Unet | 5e-5 TE with bf16 usually

    alea31415
    Author
    Mar 9, 2023

    @Hosioka I get good eyes through inpainting only masked area with size 512x512. Eyes are trained into the model especially with face images. Initial generation have small faces so the model may be not that good at it, and this particular happens for certain stylish base models (probably an artifact of Lora transfer?).

    HosiokaMar 9, 2023

    @alea31415 without inpainting at full res my models out of the box after highres fix looks subpar but yes inpainting masked will always fix eyes but I'm really trying to default to not shit eyes but im currently stuck :( its fine though its quite bugging me but not as much.

    tnginakoMar 10, 2023· 1 reaction
    CivitAI

    Hey I tried your Loha method and it works nicely but I might still need more testing. Just dropping by so I could thank you!

    alea31415
    Author
    Mar 10, 2023

    Hey thank you for passing by. Blueleaf is without doubt really the one that leads this project and I am very grateful for his effort as well :)

    LoCon
    Other

    Details

    Downloads
    174
    Platform
    CivitAI
    Platform Status
    Available
    Created
    3/9/2023
    Updated
    5/11/2026
    Deleted
    -
    Trigger Words:
    yamanomitsuha
    sabine
    colette
    beatrice
    adelaide
    edstyle

    Files

    roukin0308-23134steps.safetensors

    Mirrors

    Available On (1 platform)

    Same model published on other platforms. May have additional downloads or version variants.