Tech 5 min read

SeaArt LoRA Training Guide

Overview

I was going to set up a local LoRA training environment, but someone said, “Why not just use the cloud?”

It turns out you can upload a training dataset and create a LoRA on SeaArt (I happened to be subscribed). There’s no environment setup, and if you download a safetensors file you can generate locally.

Benefits

  • No environment setup (training runs in the cloud)
  • Train without owning a GPU
  • Download as safetensors → usable locally
  • Supports many base models: FLUX / SDXL / Pony / Illustrious

Pricing (as of December 2025)

PlanStamina/dayNotes
Free0LoRA must be public
Beginner SVIP300Private LoRA allowed, 3‑day free trial
Standard SVIP700-
Professional SVIP2,100-
  • Training is available on all plans (consumes stamina)
  • The free plan requires your LoRA to be public
  • Choose a paid plan if you want to keep the LoRA private

Preparation

Training data

  • Images: 200 (100 color / 100 monochrome)
  • Size: 512x512 (auto crop & resize by the tool)
  • Variation: face close‑up, bust‑up, full body
  • Captions: auto‑generated and then manually corrected

File layout

dataset/
├── image001.png  (512x512)
├── image001.txt  (キャプション)
├── image002.png
├── image002.txt
└── ...

SeaArt LoRA Training Steps

1. Create an account & log in

https://www.seaart.ai/ja

You can sign in easily with a social account.

2. Go to the training page

“Create → Training → Create dataset”

3. Choose a base model

Pick a model that matches what you’ll run locally (RTX 4060 8GB in this example). With 8 GB of VRAM, SDXL‑class models run.

ModelBaseCharacteristics4060 support
Illustrious-XLSDXLIllustration‑focusedYes
AuthakuMixPonyAnime‑focusedYes
NoobAI-XLSDXLGeneral purposeYes
FLUXFLUXLatest, high qualityNo (tough)

Note: SDXL LoRAs can only be used with SDXL models.

4. Upload the dataset

“Use dataset upload if you already have captioned images.”

  • Upload image and txt files in pairs
  • If filenames match, they will link automatically
  • Skip the tagging algorithm (captions already exist)

5. Check captions

After upload, confirm the following:

✓ トリガーワード(造語)が先頭にある
✓ 英語タグ、カンマ区切り
✓ キャラ固有の特徴(髪色、目色など)は削除済み
✓ 構図タグが正しく入っている

Examples of composition tags

CropTags
Face close‑upportrait, close up, face
Bust‑upupper body
Full bodyfull body

6. Set training parameters

ItemRecommended value
Repeats5 (lower since image count is large)
Epochs10
Batch size2
Learning rate0.0001 (1e‑4)
Network size (Dim)64–128
Network AlphaHalf of Dim (32–64)
Schedulercosine_with_restarts
OptimizerAdamW or DAdaptation

Step count guide

100枚: リピート10 × エポック10 = 10,000ステップ
200枚: リピート5 × エポック10 = 10,000ステップ

7. Preview prompt settings

For generating sample images after training:

my_chara, 1girl, upper body, smile, simple background

8. Start training

“Enter a dataset name → Start training now”

  • There may be a queue (2–4 hours when crowded)
  • Training continues even if you close the browser
  • Paid plans shorten wait time via priority

9. Training complete → Download

“Training history → Confirm the task is complete”

  • Sample images are shown per epoch
  • Choose the epoch with the best results
  • “Download” → Save safetensors locally
  • “Save” → Save on SeaArt (can set private)

Separate color/monochrome LoRA

If you create both

LoRATrigger wordUse case
Color versionmy_chara_colorCover, color pages
Monochrome versionmy_chara_monoMain manga

Keeping separate trigger words makes it easy to switch between them.

Prompt for monochrome generation

プロンプト:
my_chara_mono, 1girl, upper body, smile, monochrome, greyscale, manga, lineart

ネガティブ:
color, colorful

Generating in monochrome from the start looks the cleanest. Converting color → grayscale degrades quality.

Local usage

Where to place the downloaded safetensors

ComfyUI:
ComfyUI/models/loras/my_chara.safetensors

A1111:
stable-diffusion-webui/models/Lora/my_chara.safetensors

Generation on an RTX 4060 8GB

  • SDXL‑class models → work
  • LoRA + ControlNet together → tight on VRAM but works
  • ComfyUI tends to be more VRAM‑efficient

Specify composition with ControlNet

Workflow

1. Clip Studioで3Dデッサン人形ポーズ作成
2. PNG出力(512x512、白背景)
3. ControlNet(Depth推奨)に入力
4. LoRA適用して生成

ControlNet types

TypeCompatibility with Clip Studio 3DUse
DepthExcellentPreserves 3D sense
OpenPoseGoodExtracts pose only
CannyFairContours can sometimes get in the way

Depth is the most stable when starting from 3D.

Security / Visibility

Differences by plan

PlanLoRA visibilityNotes
FreeRequiredCannot use on SeaArt unless public
PaidOptionalCan keep private and still use

To use safely

  • Use a paid plan and set the model to private → only you can use it on SeaArt
  • Download as safetensors → stored fully locally; others can’t use it
  • Don’t post to a model marketplace

Even on the free plan, if you download the safetensors and use it locally, others won’t be able to use it.

Advanced settings

Learning rate & optimizer

OptimizerCharacteristicsLearning rate
AdamWStandard, stable0.0001
AdamW8bitLightweight0.0001
DAdaptationAuto‑tuning1
DAdaptAdamAuto‑tuning1
ProdigyAuto‑tuning1 (auto‑set)

Network settings

ItemDescriptionRecommended
DimHigher reflects finer features64–128
AlphaLower applies LoRA more stronglyHalf of Dim

Scheduler

TypeCharacteristics
cosine_with_restartsStable; supports multiple styles (recommended)
cosineSmooth decay
constant_with_warmupStabilizes the start

Troubleshooting

Features don’t appear

  • Increase total training steps (more epochs)
  • Increase Dim (64 → 128)
  • Make sure the trigger word is in the prompt

Too strong / collapses

  • Lower repeats
  • Raise batch size
  • Lower epochs

Stuck on a specific pose

  • Not enough variation in the training data
  • Check that composition tags (e.g., upper body) are included correctly

Wrap‑up

  1. Subscribe to a SeaArt paid plan for one month
  2. Train a color LoRA → download safetensors
  3. Train a monochrome LoRA → download safetensors
  4. Generate locally with an RTX 4060 + ComfyUI
  5. Use ControlNet (Depth) + Clip Studio 3D to specify composition
  6. Mass‑produce manga panels

Key points

  • Base models: SDXL‑class (Pony / Illustrious)
  • For 200 images: repeats 5 × epochs 10
  • Use different trigger words for color / monochrome
  • Generate in monochrome from the start (post conversion degrades quality)
  • Keep the LoRA private so others can’t use it

In the next article, I’ll actually upload the dataset, create a LoRA, and walk through tuning while reviewing the outputs.