Giant Network AI Lab Releases New Multimodal Generation Achievements

【#Tech24H】Giant Network AI Lab, in collaboration with Tsinghua University’s SATLab and Northwestern Polytechnical University, has introduced three multimodal generation technologies in the audio-visual domain. The related research outcomes will be progressively open-sourced on platforms such as GitHub and Hugging Face. The three achievements include the music-driven video generation model YingVideo-MV, the zero-shot singing voice conversion model YingMusic-SVC, and the singing voice synthesis model YingMusic-Singer, demonstrating the team's latest advancements in audio-visual multimodal generation technology.
Editor:Zhang Liyan









