| 4 | 1/1 | 返回列表 |
| 查看: 1365 | 回復(fù): 3 | |||
| 【懸賞金幣】回答本帖問題,作者h(yuǎn)e1wen2zhi將贈(zèng)送您 5 個(gè)金幣 | |||
he1wen2zhi新蟲 (初入文壇)
|
[求助]
大修20天,3個(gè)審稿人,求教大佬們我該怎么改 已有1人參與
|
||
|
求教大佬我該側(cè)重哪方面改 要加什么實(shí)驗(yàn)?zāi)?br /> 第三個(gè)審稿人說我沒用自己提的數(shù)據(jù)集做實(shí)驗(yàn),實(shí)際上我論文中寫了我用了我的數(shù)據(jù)集做的訓(xùn)練,我該怎么合理回復(fù)呢 求求大佬們給我些建議 三條審稿意見如下: Reviewer #1: This paper presents an audio-visual cross-modality generation method for talking face videos with rhythmic head. The studied topic is meaningful. The authors are suggested to further improve the paper from the following aspects. The quality evaluation of the generated audio-visual talking heads is very important for the method design. The authors have used some criteria for evaluation. The authors may give some discussions on whether it is possible to use some quality assessment methods for evaluation. For example using the audio-visual quality assessment methods proposed in 'Study of subjective and objective quality assessment of audio-visual signals', 'Attention-Guided Neural Networks for Full-Reference and No-Reference Audio-Visual Quality Assessment' for evaluation. The authors are suggested to give some discussions on this aspect and the above works. 'The proposed method demonstrates improved performance in terms of video quality compared to traditional approaches' Some discussions about visual quality assessment are suggested to be given here, considering that there are many visual quality assessment studies in the literatures, for example, 'Blind quality assessment based on pseudo-reference image', 'Blind image quality estimation via distortion aggravation', 'Unified blind quality assessment of compressed natural, graphic, and screen content images', 'Objective quality evaluation of dehazed images', 'Quality evaluation of image dehazing methods using synthetic hazy images'. Following the above comments, the quality assessment of multimedia signals is also highly relevant to this work, thus some surveys for quality assessment are suggested to be given in the introduction section of the paper, for example, 'Perceptual image quality assessment: a survey', 'Screen content quality assessment: overview, benchmark, and beyond'. Audio-visual attention is critical for various audio-visual applications. Many audio-visual attention prediction methods have been proposed, for example, 'A multimodal saliency model for videos with high audio-visual correspondence', 'Fixation prediction through multimodal analysis'. The authors may give some discussions on the possibility of using audio-visual attention prediction methods to improve the proposed method. The authors are suggested to give some discussions on this aspect and the above works. Reviewer #2: This paper addresses the generation of realistic talking facial videos by incorporating audio and head pose information. Existing methods lack natural head pose generation and audio synchronization, impacting video realism. The authors propose Flow2Flow, an autoregressive method that encodes audio and historical head poses using a multimodal transformer block with cross-attention. They introduce AVVS, a large-scale dataset for investigating rhythmic head movement patterns. The proposed method generates identity-independent facial motion representations, enabling photo-realistic videos with natural head poses and accurate lip-syncing, as demonstrated through experiments and comparisons with state-of-the-art approaches on public datasets. However, some concerns should be addressed. The organization of the paper could benefit from improvements, e.g., some video synthesis part is introduced in the feature encoding part. The authors pointed out that the full attention structure in the model excessively focuses on a single source during integration, leading to the neglect of crucial information from other modalities. As a result, accurately generating movements for the facial generation task becomes challenging. It would be helpful to provide supporting evidence or examples to further illustrate this issue. Instead of delving into the intricacies of flow theory, it would be more beneficial to focus on incorporating references in the facial attribute generation process. The model utilizes 15 neutral keypoints as facial attributes. It would be valuable for the authors to explore the impact of varying the number of keypoints and investigate whether incorporating certain 3DMM parameters and other types of audio features would enhance the results. The authors have primarily focused on discussing the applications of common loss functions. However, IQA models also have the wide-ranging applications in evaluating generative image, video, audio, and multimedia models, e.g., "Blind image quality assessment via cross-view consistency" and "Comparative perceptual assessment of visual signals using free energy features." The authors are suggested to give some discussions on this aspect and the above works. Additionally, considering the significance of attention mechanism, the authors are encouraged to provide discussions on related works like "Toward visual behavior and attention understanding for augmented 360-degree videos," "Viewing behavior supported visual saliency predictor for 360-degree videos," and "Learning a deep agent to predict head movement in 360-degree images." Reviewer #3: introductions: This paper proposes a normalizing flow based network to generate realistic talking face videos, by using audio and past head poses as inputs. Besides, they also contributes a solo-singing-themed audio-visual dataset called AVVS for research. Strength: 1. Experimental results do show that their methods can generate photo realistic videos with natural head poses and lip-syncing. And the performance looks good. 2. Utilizing normalizing flow model is novel and convincing. Weakness: 1. It is kind of stange that I do not see any experiments on AVVS dataset. Since you are proposing a dataset, I think some experiments should be conducted on it. |
新蟲 (職業(yè)作家)
超級(jí)版主 (文學(xué)泰斗)
No gains, no pains.
![]() |
專家經(jīng)驗(yàn): +21105 |
新蟲 (正式寫手)
| 4 | 1/1 | 返回列表 |
| 最具人氣熱帖推薦 [查看全部] | 作者 | 回/看 | 最后發(fā)表 | |
|---|---|---|---|---|
|
[考研] 招調(diào)劑 +4 | 帆船哥 2026-03-04 | 6/300 |
|
|---|---|---|---|---|
|
[考研] 一志愿天津大學(xué)085600 319分 材料與化工 金屬方向 求調(diào)劑 +6 | 青科11 2026-03-02 | 6/300 |
|
|
[考研] 【求調(diào)劑】293分環(huán)境工程求調(diào)劑材料/化工,服從調(diào)劑,抗壓能力強(qiáng)! +6 | xiiiia 2026-03-04 | 6/300 |
|
|
[考研] 一志愿西交化工專碩288專業(yè)課93求調(diào)劑 +6 | 好運(yùn)好運(yùn)接接 2026-03-04 | 6/300 |
|
|
[考研] 0703化學(xué) 學(xué)碩 理工科均可 不區(qū)分研究方向 總分279求調(diào)劑 +7 | 1一11 2026-03-03 | 7/350 |
|
|
[考研] 085600材料與化工(高分子)290分求調(diào)劑 +8 | wengyujian 2026-03-04 | 8/400 |
|
|
[考研] 316求調(diào)劑 +3 | 林小星發(fā)大財(cái) 2026-03-04 | 3/150 |
|
|
[考研] 264求調(diào)劑 +8 | 26調(diào)劑 2026-03-03 | 8/400 |
|
|
[考研] 292求調(diào)劑 +9 | yhk_819 2026-02-28 | 9/450 |
|
|
[考研] 主題 +3 | realstar2006 2026-02-27 | 3/150 |
|
|
[基金申請(qǐng)] 請(qǐng)問大家,研究風(fēng)險(xiǎn)與應(yīng)對(duì)措施那里, 大家都怎么寫呢 ? +3 | cauasen 2026-03-02 | 3/150 |
|
|
[碩博家園] 2025屆雙非化工碩士畢業(yè),申博 +4 | 更多的是 2026-02-27 | 5/250 |
|
|
[考研] 清華大學(xué) 材料與化工 353分求調(diào)劑 +5 | awaystay 2026-03-02 | 6/300 |
|
|
[考研] 272求調(diào)劑 +9 | 材紫有化 2026-02-28 | 9/450 |
|
|
[考研] 0856求調(diào)劑285 +11 | 呂仔龍 2026-02-28 | 11/550 |
|
|
[考研] 化學(xué),材料,環(huán)境類求調(diào)劑 +7 | 考研版棒棒 2026-03-02 | 7/350 |
|
|
[考研] 275求調(diào)劑 +7 | 明遠(yuǎn)求學(xué) 2026-03-01 | 7/350 |
|
|
[考研] 一志愿華南理工大學(xué)材料與化工326分,求調(diào)劑 +3 | wujinrui1 2026-02-28 | 3/150 |
|
|
[考研] 材料學(xué)調(diào)劑 +10 | 提神豆沙包 2026-02-28 | 12/600 |
|
|
[考研] 調(diào)劑 +3 | 簡(jiǎn)木ChuFront 2026-02-28 | 3/150 |
|