AnimateDiff Legacy Animation v5.0 [ComfyUI] - バイリンガル字幕
We will learn how to make this animation using comfy UI and anime diff.
All workflows link in description below.
Drag and drop the first workflow to get started.
First we have in place.
Then anime.
did, then prompts, then control net, which have batch or single option, then case sampler settings, then video export settings.
First, copy and paste the output folder path where frames will be rendered.
Choose dimension of the output.
Select batch size.
Let's keep it at 72 for this tutorial.
I am using this to an anime model.
I am choosing concept pyromancer lore to add some cool fire effects and putting its weight to around 0.5.
Here you choose the anime diff model.
So prompts you can use any words.
I am using this prompt.
Control net would be turned off by default.
I want to use the directory for open pose reference images.
So I will unmute the directory group.
I had open pose images extracted from my old renders so I will use those.
You can extract them using the cn passes.
extractor work flow.
Now I will unmute the control net node, and enable open pose.
I will also change the FPS at the exporting video to 12 so it does not move fast.
Time to render the queue and wait for it to It came out now we will move to the upscaling workflow.
Now drag and drop the video to video.
upscalor workflow here we will input video here's output path video and settings model
settings animate diff prompts IP adapter case sample upscale value and video out.
Right-click on the video and select Copy As Path and paste it in the input video node.
Leave the output in the settings to default.
I am putting 72 in the load cap as I rendered only 72 frames or you can also put zero if you want to render full length of the video.
Select the same model you as before.
Set the target resolution to 1200.
Make sure to change this FPS according to your video for faster or slower speed.
Everything is good to go.
Now we will hit render.
Just keep in mind that IP adapter will take long to process.
After rendering it will output in the same folder.
Lastly we will use the video 2 video face fixer workflow.
Like previous workflow, it has also similar settings.
Same as before, copy and paste the video path in the input video node.
Set load cap and video settings.
Set model as used before.
I will enter prompts to make Git more details.
Also, I will add up scale to get more better faces.
Make sure to change this FPS according to your video for faster or slower speed.
Then I will start the render.
After face fix, this is how it looks like.
Like wise I did the other two shots and added some frame interpolation for smoothness with flow frames.
I post workflow, tutorials, other on my Patreons for free, so everyone can learn better and can improve their AI artwork.
Thanks for all my Patreons for watching!
the support.
It means a lot to me.
You all keeps me going and help me to keep my tutorials free for everyone.
さらなる機能をアンロック
Trancy拡張機能をインストールすると、AI字幕、AI単語定義、AI文法分析、AIスピーチなど、さらなる機能をアンロックできます。

主要なビデオプラットフォームに対応
TrancyはYouTube、Netflix、Udemy、Disney+、TED、edX、Kehan、Courseraなどのプラットフォームにバイリンガル字幕を提供するだけでなく、一般のウェブページでのAIワード/フレーズ翻訳、全文翻訳などの機能も提供します。

全プラットフォームのブラウザに対応
TrancyはiOS Safariブラウザ拡張機能を含む、全プラットフォームで使用できます。
複数の視聴モード
シアターモード、リーディングモード、ミックスモードなど、複数の視聴モードをサポートし、バイリンガル体験を提供します。
複数の練習モード
文のリスニング、スピーキングテスト、選択肢補完、書き取りなど、複数の練習方法をサポートします。
AIビデオサマリー
OpenAIを使用してビデオを要約し、キーポイントを把握します。
AI字幕
たった3〜5分でYouTubeのAI字幕を生成し、正確かつ迅速に提供します。
AI単語定義
字幕内の単語をタップするだけで定義を検索し、AIによる定義を利用できます。
AI文法分析
文を文法的に分析し、文の意味を迅速に理解し、難しい文法をマスターします。
その他のウェブ機能
Trancyはビデオのバイリンガル字幕だけでなく、ウェブページの単語翻訳や全文翻訳などの機能も提供します。