"Fast cutting" is apparently the correct term: https://en.wikipedia.org/wiki/Fast_cutting
This lora is trained on "quick cuts", an editing technique that tells the story of a whole scene in a couple of seconds. I figured it would be suitable for the constrained context window local video producers have to work with.
Consider it experimental, as the dataset is quite limited at the moment. It may collapse completely on unseen data.
It's trained on shot concepts like "wide-angle shot", "mid-shot", "close-up shot" and (which is often used for quick cuts, "extreme close-up shot".
The format is:
A series of quick cuts:
[shot one]
[shot two]
[shot three]
...
Each cut has (about) one sentence description. You may specify angle, too.
It's trained on between 3 and 5 shots, over a very short time. Going for the full 81 frames might make it lose strength.
Tested both with T2V and I2V (but trained on I2V).
Only high noise required.
Description
Larger dataset, slightly larger resultion. Less epochs.