e-Learning Blog

Blog

2026.03.23

Material Creation
Case Studies
Know-how

Examples of AI Utilization in e-Learning Material Production (From AI Animation to Avatars)

  • Category

  • Recommended Articles

  •  

    Moodle Implementation Support and Operation Services
    e-Learning Material Production Services

    In February 2026, we conducted in-house ISMS training via e-learning.
    This time, we fully utilized AI such as video generation AI to create it, and in this blog, an AI avatar will introduce the behind-the-scenes of the creation process.

    I hope you can get a sense of what can be done (including the limitations of the features).
    * This content was created between January and February 2026. It covers the state of AI technology at that time.

    ◆ AI Tools Used

    To understand the differences between each AI, we used many AIs as listed below.
    Since new versions continue to be released, the situation is constantly changing, but please refer to the video for which ones were easier to use, etc.

    Category AI/Tool Name Provider Usage
    Planning & Scenario ChatGPT OpenAI Scenario Creation, Script Organization
    Planning and Integration GenSpark GenSpark AI Agent, Slide Generation
    Image Generation NanoBanana Google Character Design, 1st Frame Generation
    Voice Generation Gemini2.5 TTS Google Synthetic Voice (Natural Type)
    Video Generation Kling Chinese Company Video Generation
    Video Generation VEO Google Video Generation
    Video Generation SORA OpenAI Video Generation
    Image Editing PhotoShop Adobe Image Fine-tuning, GIF Animation Creation

    ◆ Main Processes and Comparison with Conventional Production (Approximate)

    Although this is a reference estimate of man-hours, using AI is expected to significantly reduce outsourcing costs and lead times.
    In conventional animation production, the operating costs plus outsourcing fees (such as animation production) for this content creation could have cost several million yen, whereas in this AI generation case, the operating costs plus AI usage fees amount to about 100,000 yen (with some compromise on quality).

    Process AI/Tools Used Work Content Required Days
    (Approximate)
    Traditional AI Usage
    1. Planning and Scenario Creation ChatGPT Summarize key points, create a draft scenario, and manually review 3 0.5
    2. Storyboard Creation Manual work Organize the structure of each scene 5 1
    3. Character Draft NanoBanana Character design, generation of multiple patterns 3 0.2
    4. 1st Frame Generation NanoBanana Generate the first image of each scene 3 0.5
    5. Image Review and Adjustment PhotoShop Taste consistency, costume changes, copyright check 2 0.5
    6. Video Generation Kling/VEO/SORA Create video from 1st frame, generate multiple times 15 2
    7. Audio Generation Gemini2.5 TTS/Yukkuri Generate audio from script, have the same sentence read twice 3 0.5
    8. Video Editing Filmora Cut out the good parts, splice them together, and integrate with audio 5 2
    9. GIF Animation Creation PhotoShop Convert the desired content into a looping GIF 1 0.1
    10. Slide Creation GenSpark Generate slide base, adjust manually 5 0.5
    11. Overall Integration Articulate Rise/Storyline Assemble the entire course, add interactive elements 2 2
    12. Final Check Manual work Check overall flow and proofread for typos and errors 3 3
    Total 50 12.8

    (Reference) Video Script

    This time, I will introduce how we created this year’s ISMS training. I will slowly explain what kind of AI we used and what difficulties we encountered.

    "Oh, Ryuno-kun made it? I’m looking forward to it."

    In this year's ISMS training, we fully utilized generative AI to create two pieces of content. Last year, we made animations using Vyond, but there were limitations in expressive power, and with video generation AI becoming practical from now on, I thought it was important to gain experience.

    "Oh, it's all done with AI. Indeed, the quality of generative AI is steadily improving, so we have to keep challenging ourselves."

    This time, we created three parts. The first is a Yukkuri-style video, the second is an animation somewhat inspired by a certain famous anime, and the third is content on countermeasures against supply chain attacks.

    "Each one has a completely different style. Specifically, what kind of AI did you use?"

    The first one I used was ChatGPT. I used it to organize the scenario and script.

    "Ah, that famous one, right?"

    After organizing the key points here, I had it generate a draft script, which I then manually reviewed. Next is GenSpark. This is an AI agent that can combine and use various AIs.

    "So you can integrate and use multiple AIs. That sounds convenient."

    The slides for the supply chain attack countermeasures, both the text and the slides themselves, were created using GenSpark.

    "You generated the video based on that scenario, right?"

    Actually, we first create a single illustration as an image. This is because it’s not possible to make a consistent video just from scripts or text. The image generation tool we used is NanoBanana. This is Google’s image generation AI. We used it to create the character drafts and the very first frame of the video, the so-called 1st frame.

    "Google, huh? Impressive. By the way, what exactly is the 1st frame?"

    By creating the 1st frame as an image, the character's appearance can be unified, and the final image becomes clearer.

    "I see, so you make a video from still images. Did you also create the voice with AI?"

    We used two types of voices. For the Yukkuri parts, we used Yukkuri Voice Maker. This is a standard software actually used in Yukkuri videos. However, the other voices were created with Gemini 2.5 TTS. We also used Gemini 2.5 TTS for other anime parts and for countermeasures against supply chain attacks. The voice is quite natural, but the drawback is that it always includes misreadings.

    "What kind of misreading?"

    For example, multi-factor authentication was read as "tayousei" authentication.

    "What kind of AI did you use for the main video generation?"

    We used three AIs for video generation. At that time, the most stable one was Kling. It's made in China, but we use it via GenSpark without inputting any confidential information.

    "Security considerations are necessary no matter which service you use."

    It was good that they followed instructions straightforwardly and didn’t do unnecessary things. Veo is Google's video generation AI, and Sora is OpenAI's video generation AI. Both have advanced technology, but they often behaved unpredictably.

    "What do you mean by unexpected behavior?"

    Like a stranger suddenly appearing in the background, or in a scene where someone picks up the receiver, for some reason the receiver is placed standing upright.

    "Wow, that's way too mysterious (laughs)"

    We also used video editing tools. We fine-tuned images with PhotoShop, cut and pasted videos with Filmora while combining audio, and finalized the content using Articulate Rise and Storyline.

    "In the end, human hands are still necessary, huh? How did you create each part?"

    First is the Yukkuri part. Normally, I would use a dedicated software called Yukkuri Movie Maker, but since I'm not familiar with the software, I created the movements with video generation AI and combined the audio using video editing software.

    "Haha, that's honest."

    Next is the anime part.

    "I've been curious about this!"

    Using NanoBanana, I created a character that combines the president and an anime character.

    "What, a mix of me and that character?"

    Finally, the e-learning for supply chain attack countermeasures. This doesn’t have anime-like movements but rather simple motions closer to typical e-learning. Items that needed to loop were converted into GIF animations using PhotoShop, and everything else was used as video. The slides were created with GenSpark, and the videos were placed there.

    "You’re using various technologies selectively. Were there any difficulties during production?"

    Both image and video generation struggled to maintain quality, and before we knew it, the outfits had changed. Also, with AI-generated synthetic voices, there were parts that couldn’t be read correctly, so this time there are sections where mispronunciations were left as is.

    "It's important not to aim for perfection and to set compromises according to the project."