Input
stringlengths 251
41.6k
| Output
stringlengths 137
9.7k
|
---|---|
Below is given review of a research paper from cnoference journal. Please write a summary the review.
### Review:
in this paper the author investigates how to utilize largescale human video to train dexterous robot manipulation skills to leverage the information from the internet videos the author proposes a handful of techniques to preprocess the video data to extract the action information then the network is trained on the extracted hand data and deployed to the real robot with some human demonstration collected by teleoperation for finetuning experiments show that the proposed pipeline can solve multiple manipulation tasks strength the direction explored in this paper is important utilizing the internet video data for robot learning is well motivated especially considering the similarity between human and multifinger hands this direction looks very promising the authors perform experiments with multiple realworld tasks with pick and place pushing and rotating objects weakness although the objective of this paper is very impressive the experiments can not support the introduction and there are multiple overclaims section 4 is titled videodex learning dexterity from youtube however i can not find any evidence that the author utilizes youtube data for learning dexterous manipulation as mentioned in the section on retargeting wrist pose orb slam and the cameras acceleration data are used to compute the camera pose trajectory this information is not readily available in the youtube data the experiments and methods are misaligned with this claim in the introduction line 42 the author mentioned that our key insight is to combine these visual and action priors from passive data with the physical constraints of how robots should move in the world however the method does not consider the surroundings of the human hand and the detection results itself is not accurate how to incorporate physical information into the training data missing literature discussion on previous learning from video works 1 dexmv imitation learning for dexterous manipulation from human videos 2021 this paper focuses also on how to learn dexterous manipulation from human videos the reviewer understands that this literature paper uses simulated tasks while the authors focus on the real robot settings but it seems that similar pipelines are also used in this paper estimating the human hand retargeting and learning from retargeted hand pose 2 the surprising effectiveness of representation learning for visual imitation 2021 this paper also focuses on how to leverage the video data for better learning it also uses a gopro camera to collect a video of each trajectory which is the same as the ego4d dataset used in this paper it shows that by learning from this video data the final manipulation performance can be improved a lot these literature works use very similar methods to achieve robot learning the novelty claims of this paper can also be found in this literature missing details for retargeting wrist pose the detection module frankmocap is a 2d hand detector it is not clear how the author can get 3d keypoints from the hand model in the camera frame also this section is important in the whole technical approach it is better to provide visualization of the final retargeted robot a hand wrist pose and robot arm should also be visualized in figure 3 if they are used in the training if the wrist pose and arm joint pose is not used how to pretrain the action prior missing details about transforms in the equation it is not clear why the author uses t and m to denote pose simultaneously what are the differences if m is also a se3 transformation how to compute the position part of the mworldc1 besides the reviewer can not find any information about how the trobotworld is determined heuristically in both the main paper and supplementary docsepthe authors demonstrate a system in which they combine a few different components to get interesting supervisedlearned open loop behavior of real robot hands doing several different tasks in particular the most notable part of the approach is using videos of human hands as an action prior which informs their supervised mapping strengths good core idea the overall idea of using action priors from human videos via hand tracking to make robots work better is a good idea there are a lot of closely related works but i think they are well referenced in this paper good execution on several key parts the execution details of handling moving cameras with camera pose tracking together with perframe hand tracking seems to be well done i also like just using r3m features out of the box this is smart and interesting to see external validation results of real robots with hands doing a variety of things weaknesses there are various unscientific elements of this paper in its current form while the work is interesting i cant recommend a strong accept for a paper in this form hopefully the list below will help the authors improve both this work and their future work if the authors can address all of the following weaknesses in their rebuttal which i think is all doable and within scope to do in a rebuttal id be happy to move from weak accept to strong accept 1 it seems like the authors are not very upfront about the fact that this method does not produce closed loop policies only on the last page or two is it mentioned that the whole method is open loop this is fine to study the task of i inputting an image of a scene and ii outputting an open loop trajectory but it of course is very limiting the tasks are carefully chosen such that they dont require any closed loop feedback this aspect of their approach is not what most researchers in the field would expect so a common experience of a researcher would be to look over the first handful of pages of this paper and only at the last page or so realize that this is an open loop method please just make this clear up front 2 several false statements in the introduction to build such robotic agents that can operate anywhere we need access to a lot of successful robot interaction data in many environments not necessarily true this is a reasonable hypothesis but one that isnt tested in this paper and it cant be stated as a fact however deploying inexperienced real world robots to collect experience must require constant supervision which is in feasible also not necessarily true but also a very reasonable hypothesis just need to say may require instead most of the inefficiency in robot learning is due to the exponentially large action space an opinion and cant be stated as fact 3 ndps can produce safe and smooth trajectories yes but this is a meaningless statement they can also produce trajectories that are completely unsafe there is nothing about ndpsdmps that provides safety other than a bit of smoothness that may arguably help but there is nothing that helps here with the presence of obstacles in the environment or humans etc this statement probably only serves to confusemislead inexperienced readers please removefix 4 the paper mentions a physical prior as a key component but this is just that it uses dynamic movement primitives it seems im not sure this is the best way to communicate this line 191 also says physicallyaware ndps they dont know anything about contact physics maybe just say second order system or dynamical system or something maybe physicallyinspired but not physicallyaware and whenever it says for example line 269 baselines without a physical prior it should just be instead clear that this just means they dont use dmps 5 line 213 is videodex able to perform general purpose manipulation since the method is open loop the answer is no thats fine and the results are still impressive but should be clarified this is not something that needs to be empirically evaluated its just a result of the formulation 6 its very confusing that citation 44 is used open loop this isnt an intention of the method also is the rnn version closed loop over time its not clear and if its not im not sure how the rnn would be any different if its not used sequentially over time 7 please state exactly how many demonstrations were used for the different experiments 8 in the conclusion this is because training rl in the real world is difficult due to hardware limitations yes but this isnt reason to make the used behavior cloning method open loop instead of closed loop minor dont worry about these too much but i mention these as opportunities to improve the paper further ego4d is not cited on page 2 mentioned but not cited hr is not defined in an equation also i would recommend not using two letters for a math symbol it looks like a matrix h multiplied by a matrix r why use orbslam3 rather than colmap for the poses already running colmap for the calibration docsepvideodex pretrains a policy network with videos with gyroscope and accelerometer data of humans performing a task then finetunes with demonstrating trajectories collected by teleoperating the robot in order to train with the human data they use the approach from 49 for mapping human pose to robot pose and use orbslam355 to account for the camera motion they feed the image data labeled with the outputted pose into a resnet1815 backbone initialized with r3ms6 features and use a neural dynamic policy ndp 13 network to generate actions the paper demonstrates that using human data allows improved performance on 67 tasks pros the paper presents a theoretically simple method of learning from videos of humans the method is demonstrated on 7 different tasks outperforming the baselines without human data on 6 of them cons the writing of the paper is somewhat scattered the analysis of why the proposed approach using ndp rather than a mlp works better with human data could be stronger the paper needs to be much clearer that it relies on gyroscope and accelerometer data from the human videos which is a barrier to truly using internetscale data
### Summary: | this paper studies how to learn dexterous manipulation from human videos in the initial review the reviewer appreciated the direction and realworld experiment but also raised concerns about the need of special sensor for tracking during rebuttal the authors effectively addressed this concern by providing additional experiment results and reviewers were satisfied with the response ac would like to recommend acceptance for this paper |
Below is given review of a research paper from cnoference journal. Please write a summary the review.
### Review:
summary of contributions this paper proposes a new framework to design new loss for gans the authors show that their framework is quite general and encompass a number of existing approaches eg the original gan formulation hinge loss etc they also propose a categorization in three different classes and derive new loss function they then compare experimentally the different existing loss and the new proposed loss that fall under their framework main comment the framework proposed in the paper is interesting since its quite general and the authors are able to derive a large number of existing as well as new loss from it however i think the framework has several limitations 1 the formulation is based on the likelihood ratio which is only defined if the support of g and f match this is known to not be the case in the context of gans 2 the benefit of the framework is not clear while it provides a way to derive new loss its not clear what are the advantages of the new loss theoretically the author argue that it is a hard question to answer and i agree the authors try to answer this question through experiments but i find the experiments not very convincing in particular the authors argue that subclass a objectives are more stable based on the celeba experiment however its not clear to me that the instability is due to a specific choice of objective function it might just be that the hyper parameters where slightly off for the other objectives i believe it would be interesting to understand better the results on celeba in particular maybe to show that some objectives are indeed more stable they can vary several hyperparameters and compare how often each objective is better than the other that would make the results and conclusion much more convincing minor comment the paper is overall clear but the clarity of some sections could be improved i think theorem 1 would be more clear if stated a bit differently simply saying that domegar maximize phidr psid and that r1 minimize phiomegarr psiomegar section 3 is a bit dense the subclasses also seem a bit arbitrary i believe section 5 could be improved by stating more clearly the different observations right now it looks more like a description of the figures than a clear statement of the question that the experiments try to answer and how they answer it docsepthis paper generalizes the minmax problem of gans to form a richer family of generative adversarial networks interestingly most of the wellknown variants of gans can be found in the spectrum of formulations covered by the family proposed in this work in terms of modeling it is evident that the family proposed in the paper is richer than that of fgan the family in this paper is shown to have a connection to wgan except that the lipschitz condition is omitted however under the light of existing works including fgan and other relevant works the obtained theoretical results are not surprising to me in addition apart from providing a richer family this work does not significantly influence the practical aspects of gans i have some following questions 1 if we solve the minmax problem in 2 subjected the fact that phi and psi satisfy eq 9 is it equivalent to minimizing any divergence between two distributions with pdfs f and g 2 dx is not a typical discriminator whose values between 01 providing the probability to distinguish true and fake data is not it d is more similar to a critique whose output values are realvalued is not itdocsepsummary in this paper the authors set out to find what scalar functions will make for a max part of the minmax gan objective they then find such a class of functions and show that only a ratio between two equal probabilities will be admitted as a solution pros the paper nicely introduces a different way of seeing gans not as a difference between the generated and real data but as a an integer of the ratio between generated and real distribution times the discriminator only if the ratio is 1 everywhere is the discriminator unable to maximize the max part of the gan objective further i liked the idea that the discriminator shouldnt just decide what class data belongs to but also estimate the probability ratio specifically in the formulation here the max part is maximized when dx omegarx so maximized iff omega1dx doesnt just classify but says the probability ratio between the two classes if this idea is expanded upon i think the authors could make a novel contribution cons unfortunately the authors have neglected to carefully explain how their contribution relates to previous work its telling that the paper cites only two papers from 2018 one from 2019 and none from 2020 all other citations are from previous years even though 20182020 has been a time of much gan research a key way in which the authors work hasnt been sufficiently compared to previous work is with their main claim we propose a simple methodology for constructing such minmax problems assuring at the same time consistency of the corresponding solution in liu they show a class of of functions where consistency is also guaranteed and the class shown by the authors here is a subset of the class in liu the details are at the bottom of my review further many of the techniques in this paper seem very similar to song where they also investigate the fgan divergence specifically the claims they make in theorem 1 seem very similar to prop 2 in song also the change of measure trick in the introduction can be found in song a detailed comparison of this work to that work would also be helpful since when reading this paper one simply doesnt know what is previous work which has already been done by others and what is the authors novel contribution once the authors address this and one is confident the contribution is indeed novel then the submission would be worth considering details of why this is a subset of whats already been shown in liu there they examine the difference between the target density d in this paper d is f but liu uses f for something else and the generated density g via supfinmathcal fmathbb exsim dysim gfxy so we find the function f in a class mathcal f which maximally separates the classes from d and g now this work proposes to do the same thing but with fxyphidx psidy where phiz intomega10z omega1tpt dt c1 and psizintomega10z pt dt c2 in liu they then split fxy up into two functions m and r such that fxymx y rxy where mxy has the form mxyvxvy this can be done in your case too resulting in here we drop the constants c1 and c2 for simplicity vx intomega10dx pt dt vy intomega10dy pt dt and rxy intomega10dx omega1t 1 ptdt since dx must be in mathcal jomega this integral has an infimum and theorem 4 from liu can be applied to achieve the same results as in this paper song song jiaming and stefano ermon bridging the gap between f gans and wasserstein gans arxiv preprint arxiv191009779 2019 liu liu shuang olivier bousquet and kamalika chaudhuri approximation and convergence properties of generative adversarial learning advances in neural information processing systems 2017 docsepoverall this paper provides impacts on understanding the core of generative models with adversarial optimization problems this paper shows the diverse possibilities of formulating the generative model optimization problems that the researchers can further investigate for better performances also this paper shows that generative models with unexplored losses achieve the best results in various datasets which demonstrates the possibilities of future improvements of generative models overall this paper is valuable to the machine learning community especially for generative models and adversarial training the below are some concerns for this paper but those concerns are not bigger than the advantages of this paper 1 quantitative experiments although the authors provided two tables table 2 and 3 there were not much analyses about the results i understand that it is not an easy problem to understand when should we use which function however it would be great if the authors can discover some trends in the results to demonstrate which type of functions work well with which type of datasets i think it would be great to use some synthetic data with known characteristics of distributions as the target distribution to analyze for understanding this point 2 other types of dataset generative models are widely utilized in computer vision however there are various other types of datasets that can get benefits of generative models such as tabular data and timeseries data it would be good if the authors can provide some simple experiments to demonstrate its generalizability 3 minor points it is not clear to transform between equation 3 and 4 i think this is a critical part in this paper thus it would be good to explain a little bit more for this part the authors explain the differences between fgan and this paper however it is not super clear to understand it would be good to clarify this point to highlight the novelty of this paper after reading other reviews are rebuttals after reading all the reviews from other reviewers and corresponding rebuttals i think this paper is a good paper and enough to be accepted in iclr 1 i think it has a clear difference from fgan it can provide a new loss function for the generative models which can further extend the success of generative models in the future 2 experiments are not super interesting but at least it has some intuitions corresponding to the authors claims 3 general theoretical results for the generative models such as when should we use which loss is a very difficult problem to solve maybe this paper can provide some intuitions for solving that large problem but it seems too much to ask this thing to the authors of this paper without that i think this paper is still worth to present to the iclr readers and participants therefore i am standing on my original score 7
### Summary: | this paper proposed a new family of losses for gans and showed that this family is quite general and encompasses a number of existing losses as well as some new loss functions the paper compared experimentally the existing losses and the new proposed losses but the benefit of this family is not clear theoretically and this work did not also provide the very helpful insights for the practical application of gans |
Below is given review of a research paper from cnoference journal. Please write a summary the review.
### Review:
this paper addresses the problem of moe routing under the cases of different network topologies by allocating another abstraction layer for the topology and designing an auxiliary objective to optimize experiments show very good improvement in terms of speed compared to strong baselines strength 1 the paper offers an important contribution to the ai community at the system level which is probably not difficult to approach for many people working in this field in fact in my humble opinion not so many ai people have the opportunity to access detailed hardware information as cloud users such as with azure or aws 2 the experiments show very good improvement over strong baselines system analysis is clearly presented weakness 1 the paper addresses the system level however since it claims a significant boost of speed without sacrificing the model accuracy it needs to show the accuracy eg at least the lmrelated one with nlprelated metrics 2 line 240 which claims without loss of generality is probably too strong my suggestion is if the solution is good with the current hardware settings the authors can run current codes for other many applications of which codes are available to further solidify their claims 3 likewise why not show the data dispatch distribution of other ranks but only rank 0 if space is limited appendix space is always there 4 in the era of gpus and large data the motivation is led by demonstrating only 128mb of data is probably inefficient probably at least some gbs or even stronger in a combination with different types of data would make a stronger motivation 5 no code is provided maybe not very relevant since the paper addresses the systemrelated level and thus is hard to judge those impacts docsepthe paper proposes a new algorithm to improve training efficiency of mixture of experts models in a distributed training setting by exploiting the network topology information to achieve this the authors propose a new auxiliary loss term incorporating communication bandwidth to encourage tokens to be routed to closer nodes rather than further nodes by applying this new algorithm authors claim that they could achiever faster throughput 101x 477x without losing accuracy on their several different clusters as a result they show a faster wallclock time convergence the communication overhead is one of the major issues for the moe model training and this paper proposes a new method to deal with this problem naturally given the increased usage of moe model technology this is a timely work having a soft guidance seems like a good idea not to hurt the original training dynamics while encouraging locality of token routing and as authors mentioned there have not been this kind of topology aware loss terms before as far as i know however there are a few missing details about model configurations and algorithms asked in the question section and the overall speed gain is minor this paper is focusing on the computation algorithm itself so it might not have direct societal impact docsepsparsely gated mixtureofexpert moe plays a vital role in largescale model training but suffers from both load imbalance and global communication in addition the existing even dispatch approach may cause network contention and worsen the previous challenges this work proposed a topologyaware largescale moe training method called tamoe that can adapt communication volume to fit the underlying network topology without interfering with the model convergence the key ideas are abstracting the dispatch problem as a communication cost optimization problem and then adding an auxiliary loss with patternrelated coefficients experiments show that tamoe provides up to 161x speedup and 477x speedup over deepspeedmoe and fastmoe without accuracy loss strengths this work tried to tackle a very significant and interesting challenge in moe system network topology may worsen the communication and load balance problems during the dispatch in moe the paper is well organized and easy to follow the proposed tamoe method is simple and effective extensive experiments show that tamoe is able to offer noticeable speedup over the stateoftheart under different hardware and model configurations weaknesses the experiments are mostly doen with gpt models it would be better to have models with different neural architectures in the evaluation benchmark it is unclear how tamoe works on other moe using models other than gpts the authors have adequately addressed the limitations and potential negative societal impact of their work
### Summary: | mixtureofexpert moe models have demonstrated a lot of success recently to further improve upon the existing literature this paper studies moe routing for different network topologies this is essentially to deal with the communication overhead of moe training the strategy is to add another layer on top for the topology along with a corresponding objective to optimize the authors also provide experiments demonstrating improved speed of convergence the reviewers were in general positive and liked the idea of the paper the reviewers did however raise issues about lack of clear demonstration that accuracy is not compromised lack of large data and a few other more technical concerns the reviewers concerns seem to be more or less addressed by the authors my overall assessment of the paper is positive i think the general premise of the paper is interesting and the paper has interesting ideas i do agree however that the experiments need to be more thorough i am recommending acceptance but request that the authors follow the reviewers comments to improve their experimental results |
Below is given review of a research paper from cnoference journal. Please write a summary the review.
### Review:
this paper discusses applications of variants of rnns and gated cnn to acoustic modeling in embedded speech recognition systems and the main focus of the paper is computational memory efficiency when we deploy the system the paper well describes the problem of the current lstm especially focusing on the recurrent connection matrix operations which is a bottle neck in this scenario and introduces variants of rnns eg qrnn also these variants may not yield enough performance compared with lstm but 1d convolution andor deep structure helps to avoid the degradation one of the biggest issues of this paper is that they use ctc as an acoustic model while still many real speech recognition applications and major open source kaldi use hybrid hmmdnntdnn lstm cnn etc systems therefore the papers claim on ctc is not along with the current application trends it may be changed near future but still hybrid systems are dominant for example the wsj wer performance listed in table 3 is easily obtained by a simple feedforward dnn in the hybrid system the latest lattice free mmi with tdnn can achieve better performance 2x wer and this decoding is quite fast compared with lstm the authors should consider this current situation of stateoftheart speech recognition also the techniques described in the paper are all based on existing techniques and the paper lacks the technical novelty other comments in abstract and the first part of introduction as i mentioned above ctc based characterprediction modeling is not a major acoustic model the paper needs some discussions about tdnn which is a major acoustic modeling fast and accurate in kaldi p4 first line and represents elementwise multiplication the elementwise multiplication operation was first appeared in eq 1 and it should be explained there section 32 i actually dont fully understand the claims of this experiment based on timit as it is phoneme recognition and not directly related to the real application which is the main target of this paper i think my suggestion is to place these timit based experiments as a preliminary experiment to investigate the variants of rnn or gated cnn before the wsj experiments i did not say that section 32 is useless this analysis is actually valuable and this suggested change about the position of this timit experiment can avoid some confusion of the main target of this paper docsepthis paper present a study on efficient acoustic modeling using neural networksbased model four approaches are presented and evaluated diag lstm qrnn gated convnet and adding a 1d convolution layer the evaluation is done on asr task using wsj and in phoneme classification task using the timit corpus the study show that the inference speed is improved with comparable of better performance than the standard lstm model the findings presented in this paper are interesting and quite useful when one wants to implement a lstmbased acoustic model on mobile devices the paper is well written and easy to ready the main issue of this paper is the lack of novelty the three evaluated approaches diag lstm qrnn and gated convnet are not novel the only novelty is the addition of a 1d convolution which is not enough for a conference like iclr minor comments on the experiments the network quantization approach has been shown to lead to efficient neural networks could the authors provide a comparison between their approach and the quantization approach on the timit experiment the authors could add a decoder and use the per metric instead of the frame accuracy so they could provide comparison with the literature wsj and timit are quite small corpora compared to the available corpora maybe the authors should consider using large corpora like librispeech it could be interesting to see the performance of the presented approaches overall this paper feels more like a technical report the findings could be useful but its novelty is too limited for iclr hence i argue for rejection and suggest that the authors consider submitting the paper to a speech conference like icasspdocsepthis paper investigates a number of techniques and neural network architectures for embedded acoustic modeling the goal is to reduce the memory access and make efficient computation in the meantime to sustain good asr performance overall the paper is well motivated and well written however i have following concerns 1 it is not clear from the paper whether both the training and inference are conducted on embedded devices or only the inference i assume it is the latter but cant find it explicitly mentioned in the paper 2 the exploration carried out in the paper is more on the system level and the novelty is not overwhelmingly significant 3 my major concern is that the reported wers on wsj and phoneme classification accuracy are quite off 2030 wers for wsj do not seem to be usable in real applications honestly i dont even think this performance is better than welltrained gmmhmm acoustic models using a viterbi decoder furthermore there is no clear winners across the investigated architectures in terms of performance one question is if one wants to deploy such an ondevice system which architecture shall be chosen 4 a more general comment on the work explored in the paper first of all the ondevice memory issue puts a heavy constraint on the capacity of acoustic models which will significantly hurt the modeling capability for the dnnbased acoustic models deep learning acoustic models can outperform gmmhmm because they can use large model capacity with very deep and complex architectures when a large amount of training data is available second for ctc when the training data is limited its performance is far worse than the hybrid dnnhmm model let alone a pure endtoend fashion without using external lm and dictionary if wfstbased decoders composition of wfsts of lm dictionary and deblankrepetition are used then the memory issue will surface again
### Summary: | in this work the authors conduct experiments using variants of rnns and gated cnns on a speech recognition task motivated by the goal of reducing the computational requirements when deploying these models on mobile devices while this is an important concern for practical deployment of asr systems the main concerns expressed by the reviewers is that the work lacks novelty further the authors choice to investigate ctc based systems which predict characters these models are not stateoftheart for asr and as such it is hard to judge the impact of this work on a stateoftheart embedded asr system finally it would be beneficial to replicate results on a much larger corpus such as librispeech or switchboard based on the unanimous decision from the reviewers the ac agrees that the work in the present form should be rejected |
Below is given review of a research paper from cnoference journal. Please write a summary the review.
### Review:
the authors introduce the problem of telegraphic summarization given a sentence we want to reduce its size while retaining its meaning with no penalty for grammatical mistakes the main application presented by the author is that of summarizing fictional stories and plays the setting proposed by the author prescribes that the summarized sentence can be obtained by the input sentence by dropping some words so for example the simplest baseline for this problem would consist of simply dropping stop words the approach proposed is basically an autoencoder consisting of a 2step encoderdecoder network in the first step the sentence is encoded into a vector which is in turn decoded to a smooth indicator vector to mask words in the sentence in the second step the masked sentence is encoded into a vector which is in turn decoded into the output summarized sentence the optimization is a tradeoff between recoverability of the input sentence and norm of the indicator vector how many words are dropped in order for the network not to learn repetitive masking patterns eg drop first half of the sentence or drop every other word an additional loss is introduced that penalizes keeping easily inferable words or dropping hardtoinfer words concerns the problem doesnt seem to be wellmotivated also the length of the obtained summarized sentences is 70 that of the original sentences which makes the summaries seem not very useful the proposed complex architecture seems not to justify the goal especially considering that simply dropping stop words works already quite well in order for the presented architecture to beat the simple stopwords baseline an additional loss l4 linkage loss with retention weights which need to be tuned manually as hyperparameters is required theres not enough discussion about the related work by malireddy et al which is extremely similar to this paper a good part of that work overlaps with this paper comparison with literature about abstractive summarization is completely missing minor comments figure 1 indicator encoder should be indicator decoder are negations part of your stop words from your discussion you should make sure that not dont doesnt do not belong to your stop word set how did you optimize the hyperparameters r desired compression the regularization weights and the retention weights were pretrained word embeddings used as initialization whats the average compression of golden sentences docsepthe authors consider the problem of telegraphic sentence compression they train a system in an unsupervised fashion to predict which words can be dropped from a sentence without drastic loss of information to that end they propose a new autoencoding type architecture which uses the extracted words as latent code and most importantly a linkage loss which relates a words perplexity given the summary of its left context to its likelihood of being retained the model itself is sober and well motivated and the linkage loss is to the best of my knowledge original the authors show that their method outperforms some simple baselines in terms of rouge and compression on a small humanannotated test set the paper is generally well written although the initial presentation of the model could be made a little clearer it is not obvious from the text that the decoder takes the text as input figure 2 helps but comes a couple pages later however the authors fail to appropriately justify the choice of their hyperparameters eg the optimum value of r for our experiments was found to be 065 the best value of b was found to be 5 the weights 1 2 3 and 4 have been set to 3 2 50 and 3 respectively for our experiments how is best measured on the validation set which does not have gold references the choice of the specific sparsity constraint one could as well imagine using a simpe l1 regularization for the binarization loss and of chii why not simply use the likelihood could also be better motivated the model also relies on a handcrafted rules section 33 whose effect needs to be made more evident what weights are used in practice how were they chosen we observed that needs to be further developed the authors claim that the quantitative scores are not affected significantly but that is presumably only the rouge score what about annotators preferences most importantly however the task of telegraphic sentence compression whose usefulness is not a priori obvious is barely motivated the author refer to malireddy et al 2018 for a justification but it is important to note that the latter provides a telegraphic summary of a whole document with a compression factor of 037 the claim is that the concatenation of the telegraphic sentence compression can act as a summary of a whole document but given the fact that compression for individual sentences is closer to 069 this is yet to be demonstrated and even if that were true it is unclear whether the cognitive load of reading a sequence of telegraphic sentences would be that much lower than that of reading the original text this paper presents some interesting ideas and is well written but the content is not quite sufficient for publication in addition to the clarifications and justifications requested above the authors are encouraged to apply there methods to full lengths documents which would make for a more substantial contribution docsepthe paper explores unsupervised deep learning model for extractive telegraphic summaries which extracts text fragments eg fragments of a sentence as summaries the paper is in general well structured and is easy to follow however i think the submission does not have enough content to be accepted to the conference first in term of methodology as described in section 3 the paper has little novelty there has been intensive study using various deep learning models on summarization the models described in the paper contain little novelty compared with previous work using autoencoder and lstm for both extractive and abstractive summarization second the paper claims contributions on using deep learning models on telegraphic summarization but the advantage is not well demonstrated for example the advantage of the resulting summary is not compared with stateoftheart sentence compression models with intrinsic evaluation or probably better with extrinsic evaluation by the way it is interesting that the paper argues the advantage of using telegraphic summaries for fictional stories but actually gives an example which looks also very typical in news articles the earthquake tokyo 12 dead example third there has been much work on speech summarization that summarizes with the telegraphic style this is natural considering speech transcripts are often nongrammatical and telegraphic style summaries focusing on choosing informative fragments actually result in usable summaries the authors may consider discussing such work and compare the proposed methods to it
### Summary: | this paper presents methods for telegraphic summarization a task that generates extremely short summaries there are concerns about the utility of the task in general and also the novelty of the modeling framework there is overall consensus between reviewers regarding the papers assessment the feedback is lukewarm |
Below is given review of a research paper from cnoference journal. Please write a summary the review.
### Review:
this work tackles the task of forecasting dynamics in different domains simultaneously using an encoder which is trained to determine the task the inferred latent vector is then used to adapt a forecasting network to the task at hand experiments on three datasets linked to fluid dynamics are then conducted to assess the proposed model pros this is an interesting problem which is quite timely given the development of the field of forecasting physical dynamics using neural networks the proposed solution seems sound and principled moreover it is well motivated and the writing was quite clear the different additions made to the forecaster network are also quite interesting i especially liked the adapad solution to deal with boundary conditions conducting an ablation study also considerably strengthens the paper cons all experiments are conducted on somewhat similar datasets which are based on fluid dynamics pdes it would be nice to see how the model deals with other families of dynamics especially given the fact that the contributions of this work seem geared towards practical considerations the setting of the experiments should be more precise and additional details should be given how are the different datasets constructed what supervision is there exactly regarding the different tasks how many domains are there in each dataset and what are the differences how is the balance between the different domains ect this is a good work on a timely subject the contribution is not groundbreaking but should be significant enough to warrant acceptance docsepthis paper addresses the problem of learning a deep learning model for dynamics forecasting which generalizes to changes in dynamics these changes can be induced by different parameters boundary conditions or external forces the proposed model takes a metalearning approach and proposes to partition data into different heterogeneous domains it consists of two components an encoder which infers timeinvariant features given observed domain data and a forecaster which predicts the dynamics given these features the paper evaluates the proposed approach on several datasets and provides some theoretical insights this paper addresses a new and interesting generalization problem for dynamics forecasting it proposes a model to address different changes in the dynamics evaluation is done on relevant datasets with several baselines and some ablation studies the applicability of the proposed approach is restricted to problems where relevant weak supervision from task parameters is available this seems like an important limitation in realworld applications how valid is this scenario the question of choosing relevant parameters for weak supervision is important for applying this model to other datasets yet the definition of these parameters is unclear how robust is the model when chosen parameters are not useful the performance of wrongenc table 2 tends to say that this model will then fail it is unclear why the model can adapt to changing boundary conditions with adapad as it generates them from features hatzc extracted from data inside the domain and weakly supervised by quantities unrelated to the boundary condition eg mean vorticity or season the theoretical analysis inspired by existing work in multitask learning domain adaptation has some limitations and does not add much value to the paper i have some concerns with the domain adaptation upperbound to the target error in theorem 34 and proposition 35 this upperbound is not minimized thus the target risk can be high ie the model is not guaranteed to adapt well moreover the validity of the theoretical analysis is unclear as several assumptions may not be verified eg bounded loss in theorem 31 proposition 33 lipschitz continuity in proposition 35 theorem 34 requires that the assumptions in theorem 2 in redko et al 2017 are verified yet these assumptions are not mentioned in the paper some ablation studies are missing 1 the contribution of each term in equation 2 and 2 the dimensionality of hatzc which is fixed arbitrarily other questions it would be good to better explain how the experiments include changing boundary conditions between domains the testing scenarios only mention different initial conditions or external forces why do the baselines resnetc and unetc not adapt well despite having access to relevant weak supervision p8 this is the same information used by the proposed model to adapt how redundant is the time invariance term 3rd term in equation 2 with the invariances enforced in the architecture of the encoder this paper tackles a new generalization problem for dynamics forecasting and proposes a model supported by experimental results however this model can only be applied to problems with relevant weak supervision which may not always be available in practise moreover the definition of relevant parameters is unclear and the robustness of the model to the choice of these parameters is not measured which may restrict its application to other datasets there are also unclarities on the ability of the model to adapt to changing boundary conditions with adapad some ablation studies are missing and i have concerns on the theoretical analysis which brings limited value to the paper for this reason i am giving this paper a weak reject postrebuttal comments i thank the authors for their response after studying it the theoretical results still have some major issues and feel disconnected from the model in particular key assumptions are not enforced in the model eg lipschitz continuity and the generalization error of the model in th33 is uncontrolled as the upperbound is not minimized by the model the wasserstein distance between domains is fixed and is high in all generality its use for the model is thus not very convincing on practical aspects the capability of handling boundary conditions should be better justified and evaluated for this reason i keep my score unchanged and recommend rejecting this paper docsepthe paper suggest a remediation for a common problem for dynamics forecasting which is the lack of generalization to other domainstasks the author suggest to tackle this with via a 2 component architecture one for learning the task and one for forecasting in empiricial experiments the authors show the practical feasibility of their approach as a caveat im not an expert in the area so my review remains on a superficial level consequently for which i apologize i overall liked the paper quite a bit the question discussed is relevant the empirical evaluation is very good the theoretical results seem as relevant as they would get and the related work discussed is crisply presented and relevant one question i would have is that results in table 1 are overwhelmingly good with only unetc coming close do we know for these tasks what the theoretical upper bound eg by the right pde system would be is it computationally even possible to compute this upper bound im wondering how much of a gap there still is too close in a similar vein what is the intuition behind dyad resnet being better than dyad unet mostly are there some complementary strengths between dyad and resnet that this combination can exploit better than dyad unet this is a good paper that id like to see accepted for its combination of theoretical results empirical results and methodological novelty docsepthis paper is interested in learning general forecasting models for physical dynamical processes the paper proposes a decomposition of such a model into an encoder that captures the innate properties of the system and a forecaster that autoregressively makes predictions conditioned on the encoded properties this is framed as a metalearning approach and is shown to substantially outperform singletask approaches and offtheshell metalearning approaches across multiple datasets the paper provides some theoretical analysis and qualitative analysis of what is learned overall the paper shows that learning shared models across domains is an important and fruitful way forward for modeling physical processes with machine learning strengths the problem statement is wellmotivated learning generalizable deep learning models across diverse settings is an important open problem experiments use interesting and realworld problems results are strong and appear reliable adapad is an interesting idea specialized to the case of physical complex systems since it is designed to address boundary condition issues visualizations show the model is behaving essentially as expected although there are many design choices that go in to the model each such design choice is wellmotivated aside from some aspects of the theory section the exposition is generally quite clear and wellorganized assumptions are made clear the fact that the encoder can be trained first and independently of the forecaster should be very useful for further rapid developments great to see ese metric used as a complement to raw error table in appendix showing alternatives to adain is very useful in increasing confidence in adain for this application weaknesses the biggest concern is the theory section the multitask learning and domain adaptation results are general results that are not adequately connected back to the specific model and problem the paper is considering yes it is widely accepted that multitask learning and domain adaptation can work well especially when tasks are related in some measurable way and it can be a useful exercise to restate existing theory in the language of your framework but what if any novel claims is the theory implying are there any predictions the theory makes about the particular approach which can be validated in experiments the theoretical bound on error that decomposes the error of the encoder and forecaster is similarly lacking in its interpretation yes it can be a useful exercise to show that the error can be decomposed along the lines of the model but does this bound somehow suggest that the decomposition results in lower error than a monolithic model or is it showing that you can work independently on improving either part of the model and improve the overall error where is there potential for practical value in this theorem for example one place there could be potential to validate the theory is to check in experiments that task pairs with lower wasserstein distance actually support better domain adaptation however in the introduction of the paper it acknowledges that even the slightest change in these features may lead to vastly different phenomena but doesnt that suggest that wasserstein distance may not be a useful metric here for measuring task similarity couldnt turbulence limit the usefulness of such a metric proposition 33 says the bound is strictly looser than the bound in theorem 31 for clarity it would be very helpful to combine the bounds into an inequality showing this strictlylooser property it is not immediately apparent from the statement of the theorems since the inequalities contain different terms as is the theory doesnt really hurt the paper but for the amount of space dedicated to it it doesnt add much the paper could be substantially improved by either 1 adding interpretationpredictionsvalidation of the theory that connect it back to the approach in the paper or 2 removing some of the less useful parts of the theory from the main paper to free up space for more of the interesting analysis of what the model actually learns also it is interesting but a bit counterintuitive that the theory section relies on results in multitask learning and domain adaptation instead of theoretical results from the metalearning literature as is since the paper relies on multitask learning so much it is missing references to related work in multitask learning ie related work outside of modeling physical dynamical systems similarly it would be helpful to mention why there are no comparisons to multitask learning or domain adaptation methods in the experiments why do they not apply here the three terms in the loss function of the encoder are wellmotivated but it is not clear how important each term is ablations on these terms would be very informative for the reader to understand whats generally required to train an encoder in section 5 it says varsepnet employs separation of variables through different loss terms what are these loss terms and how are they different from the ones in the paper in the ablations with no encoder how do adain and adapad work dont they require some z where does this come from if not from the encoder unet does seem it could be at a qualitative disadvantage compared to dyad in terms on number of parameters especially since unet c is one of the more competitive baselines it would be useful to see results for a larger unet c or at least some evidence that the unet is not underfitting the training data additional question of interest overall this is a very important a potentially deep line of research the most exciting promise of such work is the potential of revealing shared regularities across vastly disparate dynamic systems that is across complex physical processes and it seems the approach in the paper could be particularly wellsuited to such research for example the authors could train a single encoderforecaster model across all the datasets in the paper and analyze relationships in the learned encodings across datasets training models across highly diverse domains have been tried in multitask learning eg pretrained transformers as universal computation engines arxiv 2021 the traveling observer model iclr 2021 modular universal reparameterization neurips 2019 one model to learn them all arxiv 2017 is such a generalization part of the longer term vision for this line of work minor comments in section 24 some references would be useful in the sentence ending with the combined force equation there are several inconsistencies in the use of parentheses in citations throughout the paper correcting these would improve readability in last sentence of first paragraph of section 4 the word task could be changed to something like problem since task has another meaning in the paper should the 726 for unetc on ocean currents future be bolded in the last paragraph of section 51 we tried to vary we tried varying or we varied appendix a21 footnote for phiflow is on the wrong page appendix a21 the last paragraph seems like it should be the first paragraph of a22 in proof of proposition b5 there is an extra or missing set of norm bars in the first inequality overall this is very interesting and useful work the problem is wellmotivated and the approach and experiments are carefully designed and generally convincing if the concerns about the theory are addressed i would be happy to increase my score adding the additional info and experiments requested could increase it further and make this a particularly strong paper
### Summary: | the paper addresses the problem of domain generalization for learning spatiotemporal dynamics it proposes a solution where an encoder captures some characteristics of a given environment and a forecaster autoregressively predicts future dynamics conditioned on the characteristics learned by the encoder said otherwise the forecaster learns the general form of dynamics parameterized by an environment representation extracted by the encoder the conditioning is implemented via an adaptive instance normalization mechanism a form of padding is also introduced in order to take into account boundary conditions the two components encoder and forecaster are trained sequentially this approach is casted in a metalearning framework theoretical results inspired by multitask learning and domain adaptation are also demonstrated the model is evaluated and compared to different baselines on three problems and for two different settings varying initial conditions with a given dynamics and dynamics with varying parameters this is a borderline paper it targets a timely and important problem of domain generalization for dynamic environments the proposed solution is original and compares well experimentally to several baselines it allows for better generalization performance for the two test settings considered in the current version the paper however suffers from different weaknesses first there is the imprecision of the arguments and the description of the experiments some of the arguments and claims are vague and sometimes abusive not backed up by evidence for example a central claim is that the encoder learns time invariant quantities characterizing the environment when the learned representations indeed change with a time shift in the input for any environment the same goes for the argument developed for the padding construction it is claimed to model boundary conditions but this is not supported by any theoretical or empirical evidence as noted by the reviewers the theoretical analysis is disconnected from the algorithmic and experimental developments and does not bring much additional value to the paper what is more embarrassing is that some of the claims in this section are overstated and induce incorrect conclusions from theorem 31 and proposition 33 the authors suggest that multitask learning leads to better generalization than learning independently while this is not formally guaranteed by the results this is acknowledged by the authors in a later comment besides the conditions of validity are not discussed while they seem to only cover situations for which the train and the test distributions are the same the same holds for the second theoretical results theorem 34 it is claimed that this result supports the authors idea of training encoder and forecaster sequentially while it does not besides the bounds in this result cannot be controlled as noted by the reviewers and are not useful in practice overall the paper addresses an important topic and proposes new solutions the results are promising and it is indeed an interesting contribution however inaccuracies and incorrect or exaggerated claims make it difficult to accept the current version of the article the article would make a strong and innovative contribution if it were written as a purely experimental article with a detailed description of the experiments and comparisons |
Below is given review of a research paper from cnoference journal. Please write a summary the review.
### Review:
the paper studies the mixture of experts moe architecture which has become popular in nlp recently as a way to increase the capacity of network without increasing depth the authors aim to develop a theoretical understanding of the moe modelconditional computation the authors begin with a formal model for conditionally activated sparse models which can capture common existing moe models the authors use lsh locally sensitive hashing for the gating in moe and use this to derive a few theoretical results regarding the ability to approximate real valued lipschitz functions in rd the authors perform some small scale experiments to back up and verify their theoretical findings post rebuttal thanks to the authors for running the experiments and for sharing the insights like i said earlier it is important to study the theoretical underlining of moes this paper starts with it although as a researchers actively working in moes i do not think that the paper exactly answers the key questions the results proven are expected and not surprising but on the other hand as pointed by the authors nontrivial to prove so i would say that it is a decent paper at the moment and would suggest the authors to keep going in this direction to develop a more thorough understanding so that they can uncover some more fundamental results strengths 1 relevant problem moes are becoming very popular in nlp thus it is important to study their underlying theoretical and working mechanisms the paper tackles this relevant problem 2 lsh locally sensitive hashing authors propose to use lsh for gating this can actually be quite promising in my opinion because it takes the local vicinity into consideration 3 well written it is a very well written paper and is easy to follow i really like the limitations section it is good to see for a change that there is somebody who knows and writes the limitations of their work weaknesses 1 weak experimental evaluation i think that lsh is infact good authors need to perform more experiments to show its effectiveness i am not suggesting to go to the huge model sizes but atleast medium scale models and datasets should be evaluated very well discussed docsepfrom my understanding the main contribution of the paper is as follows 1 the authors capture the sparsity structure of these popular transformers they model the transformers into dsm models 2 they show that the dsm model can represent the lsh model 3 they provide theory on the lsh model these theories can be used to interpret the success of switch and scaling transformers 4 motivated by the theory they proposed a new lshbased model and run toy experiments to show its efficacy see my comments below i have the following concerns 1 in the current manuscript the connection between contribution 12 34 is still a bit vague see my comments in the summary part when reading the current version it is easy to get confused about the main contribution of the paper is it explaining why general sparse model like scale transformer works well or is it designing new lhs methods to save inference costs for me the contribution of the explaining part outweighs the designing part this is because the author didnt provide any realdata experiments on lsh if no realdata experiment is provided lsh is a pure theoretical tool to prove the theory on dsm and the designing part is minor however the current script overemphasizes the designing part causing great confusion for me 2 the performance of dsm on cifar10 does not quite match the theory to support the theory i would suggest the authors run experiments on transformerbased nlp tasks instead of cv tasks on cifar 3 all the theories are built on the function lipschitz assumption it would be better if the authors verify the lipschitz condition is it a necessary condition or it is due to the limitation of the theory if it is the latter case what is the main technical challenge to relax this assumption 4 in line 255 why is random ddegree polynomial a lipschitz function docsepthis paper provides a theoretical treatment of modern sparsely activated networks with the datadependent sparse model dsm model the authors show that the dsm model can simulate modern sparsely activated networks and the locality sensitive hashing lsh model it is proven in the paper that the lsh model can be expressive as a dense network for approximating realvalued lipschitz functions while requiring much fewer flops furthermore experiments are conducted to validate the theoretical findings on lipschitz target functions as well as the cifar10 dataset strength 1 the paper is the first work to treat sparsely activated networks theoretically thus novel to me 2 the paper is wellorganized weaknesses 1 the theoretical analysis is based on the assumption of the llipschitz target function and i am not sure how significant the work is furthermore the neural network size used in the experiment is also very small 2 i am not very sure about the relation between the theoretical findings and experiments theorem 41 and theorem 43 conclude that lshbased sparsely activated networks can be expressive as their dense counterparts when their size and number of samples match as for size the lsh model is measured using hash table size and the dense model is measured using width as a result in experiments i am expecting to see the lsh model is as good as dense ones when buckets width of the dense network however in the figures the width of dense networks is compared to the number of activated units 3 for comparison of dsm and dense networks the authors mention that sparsity helps in both dsm and lsh models using the same number of activated units however it seems that the comparison may be unfair specifically with 64 activated units dsm chooses the best 64 units out of total of 1024 units while the dense one has only 64 units in total it seems unsurprising to me that dsm is better than its dense counterpart yes the authors have addressed the limitations and potential negative societal impact of the work docsepthis paper proposes the dsm model to sparsely approximate lipschitz functions the authors theoretically demonstrate their method in a wide range of scenarios from onelayer shallow neural networks to switch and scale transformers the original idea but i am not sure as i am not familiar with this domain of interpreting dsm as knn is very interesting however the experiment setting is a bit weak and seems to be finished in a rush i have increased the rating from 4 to 5 after rebuttal clarity strengths this paper offers a detailed introduction to the lsm model and other background knowledge weaknesses if the authors can further unify the usage of notations the overall readability will be better for example the authors use s for the sparsity parameter but in sec 30 it switches to k and later k is used as the intrinsic dimension of input distributions the usage of notation ax is also a bit confusing i also recommend the authors add some figures to illustrate their idea for example the euclidean lsm and sec 30 can be well explained by figures originality i am not familiar with this domain so i may not be able to judge this point but still i find the argument in sec 30 interesting it points out a potential direction in that we may interpret neural networks as knn operators the current content can be enhanced in some directions the authors may try to remove the constraints of unit b rows typical network blocks like cnns attention networks and residual connections do not have such unit structures also extending it to deep neural networks will be more attractive quality strengths the theory analysis is careful and in depth but some settings and assumptions need either explanation to justify the necessity or adjustment to cater the practice demands weaknesses weird experiment setting 1 now that the main point is the efficiency of the proposed method why not report inference time flops is a good metric but not enough 2 needs more detailed ablation study specifically detailed study on how sparsity parameter s influence the model accuracy approximation mse and inference time 3 now that the paper put much attention on discussing input distributions the authors should also use input distributions on low dimensional manifold in rn currently it is unclear how the input is sampled 4 from numerical perspectives polynomial may not be good choices as they tend to be extreamly ill conditioned when degree is high the authors may consider bspline or bzier curve for some realistic industrial scenarios also random neural networks even shallow ones may be good candidates theoretical settings need clarify and adjustments 1 it is a bit weird to assume input distributions be uniform can it be replaced as absultely continuous with respect to the uniform distribution lebesgue measure 2 now that the proof is based on euclidean lsh it should be clearly stated in the theorems significance the theory results is good but needs stronger empirical evidence to support it 1 i strongly encourage the authors to add figures to illustrate their concepts and ideas 2 experiments on sota neural networks will be much appreciated 3 needs a more detailed ablation study to justify the theory results
### Summary: | the paper provides a theoretical analysis of sparsely activated neural networks they introduce lsh local sensitive hashing as a new routing function for theoretical analysis and proved a few results on representation power and inference time one reviewer pointed out that the theoretical results are expected and do not provide much interesting insight which i agree with nevertheless this is one of the early papers that study sparsely activated networks and may serve as a starting point i recommend acceptance |
Below is given review of a research paper from cnoference journal. Please write a summary the review.
### Review:
the paper proposes a new approach to inject knowledge into pretrained language representation models plms instead of tuning the original plm parameters the paper plugs in new adapters for knowledge injection to avoid catastrophic forgetting pros injecting knowledge into plms is an advanced topic the authors focus on the catastrophic forgetting problem during knowledge injection evaluation is solid the authors evaluate their model on three downstream tasks and show that the adapters improve the performance the paper is well written and can be easily understood cons the approach is simple but achieves good performance over a variety of tasks i appreciate that the authors conduct the knowledge probing experiment but its p1 is quite low and worse than bert some more explanations are expected docsepsummary the paper proposes a novel approach of incorporating different types of world knowledge sources contained in texts such as facts or linguistic syntax to do this they introduce additional transformers layers between the layers of a pretrained language model such as roberta and term this model as kadapters where the k stands for k different streams of knowledge pros incorporating different sources of information into a pretrained model such as roberta is an interesting idea the proposed approach is simple and interesting and it scales to many different types of information as the different adapters can be trained in parallel with the weights of the pretrained lm being fixed performance gains on different classification tasks such as entity tying questionanswering and relation classification highlights the utility of the approach cons sec 1 introduction in the introduction there are multiple mentions of the phrase rich knowledge but it is unclear what do the authors mean by that in the context of pretrained language models some recent works such as httpsarxivorgabs200208910 httpsarxivorgabs190901066 suggest that pretrained language models do indeed contain a lot of world factual knowledge hence the statement in the paper that pertained lms lack world knowledge is contradicting these works there is also a frequent mention of catastrophic forgetting of knowledge during the finetuning step i tend to disagree that this is necessarily bad for a pretrained model because it has been shown that finetuning pretrained lms perform well in opendomain question answering where some degree of world knowledge is needed furthermore producing entangled representations may not necessarily be a negative thing if multitask learning approaches are able to show an increase in performance due to knowledge injection in table 1 dependency parser doesnt really fall under the same class of knowledge sources such as wordnet or wikidata a dependency parser may be able to provide some sort of syntactic structure of the underlying text moreover such syntactic information is not always generalizable to different domains and thus has the limitation of not being accurate enough the introduction section is not wellmotivated and does not present convincing arguments as to why external knowledge infusion is really required in some tasks it just states that knowledge infusion using kadapter model outperforms roberta models in different tasks in section 31 not enough space has been allocated to explain the adapter model in detail if the authors had used mathematical notation or equations for explanation then it would have been much more clear in section 4 it is mentioned that they select three downstream tasks for evaluating their models however the paper doesnt provide justifications as to why these tasks were selected how can these tasks highlight the importance of kadapter model etc in the results table 2 and table 4 as the performance improvements are somewhat marginal it is important to know if these improvements are statistically significant or not the paper doesnt report if the results are from single run or the mean of multiple runs i have concerns about data leakage during the pretraining step as the factual adapter makes use of supervised relation classification dataset trex i feel that there might be some overlap between entity typing and relation classification datasets used for evaluating the performance of the model the authors should present an analysis as to what degree of overlap if any is present during the pretraining and evaluation tasks the paper lacks a detailed analysis section that could explain as to which test examples are getting correctly classified when using kadapter model in tasks like relation classification entity typing compared to other baseline approaches such as roberta roberta multitask currently the paper pays just too much emphasis on raw numbers or performance improvements in various tasks the results of the probing experiments suggests that bertlarge model vastly outperforms kadapter model on googlere and trex datasets probing datasets this raises an important question over the validity of the results in different downstream tasks for a fair comparison with baselines the authors should compare the performance of the kadapter model with bertlarge multitask across different tasks in almost all of the experiments the authors use roberta as the underlying pretrained language model for demonstrating generalization to different pretrained lms the paper should also evaluate when kadapter model is trained when bertlarge or t5large are used as underlying models in place of roberta grammar errors page 1 3rd line from bottom remains retains section 33 information that concerned information that is concerned section 34 father index is commonly referred to as head index of a worddocsep summary this submission proposes a general method kadapter for injecting knowledge either factual or linguistic into pretrained language models the key architectural property of the approach is that kadapters are isolated from one another allowing the use of multiple adapters without interference these kadapter modules take hidden layer inputs from the main pretrained model eg bert and are pretrained on their knowledge outputs before a finetuning phase where they feed into a joint downstream taskspecific model along with the pretrained model outputs strong and weak points the set of baselines seem strong and the experimental results consistently show that using either factual or linguistic knowledge kadapters improves while using both yields the best results the lama probing experiment is a nice sanity or validation test that the knowledge injection is achieving the desired effect being able to hardcode knowledge into the model in this way could be useful in a variety of applications it is overselling it a bit to say the model captures richer commonsense knowledge however the basic architectural idea is wellmotivated and simple in a good way the supplemental materials mostly provide additional reproducibility details on architectures hardware used learning rates etc recommendation accept or reject with one or two key reasons for this choice i recommend to accept the proposed approach yields strong quantitative performance against solid and relevant baselines and the lama experiments give some support to the hypothesis that it is doing so by capturing knowledge as intended the general design pattern could spur further innovations in modular network designs or knowledge capture strategies as well questions to clarify additional evidence required bertmk inegrates fact triples from the knowledge graph how i can follow the cite but this sentence provides little information inject different types of knowledge independently is it correct to say then that by design there can be no beneficial interactions or synergies among different types of knowledge alternatively in the finetuning phase could different adapters interact or affect each other via the downstream coupling in the taskspecific layers is this observed in practice how should the reader think about the relative magnitude of the presented improvements at one point i see kadapter fl makes significant improvement of but i believe significance is only meant coloquially here section 31 how was this structure chosen what was the motivation or intuition here what limits if any do you foresee with the use of separate parallel knowledge modules like this could we use 10 100 1000 kadapters additional feedback to improve it would be helpful to cite ling and weld 2012 or similar for the definition of loose micromacro f1 or briefly explain it inline in the evaluation setup likewise for the catastrophic forgetting phenomenon affecting other knowledge injection attempts is there some previous work explicitly demonstrating this problem when using multiple knowledge sources if not it would have been interesting to have an experiment of this sort in this work docsep reasons for score the authors propose a plugin based adapter approach to allow for task specific parameter settings without updating the original pretrained model which prevents the potential for catastrophic forgetting while also removing the need for separate models for separate tasks the work seems to build off houlsby 19 as briefly cited but its in plugin nature seems easier to adopt for multiple tasks there is however not direct comparison with it or cooper et al 19 httpsarxivorgpdf190202671pdf which makes it difficult to assess the way in which the adaptors were pretrained was a little unclear to me the experiments are extensive and well done pros 1 the number of experiments run 3 tasks on 6 datasets total are extensive and shows the kadaptor approach can benefit from the factual adaptor in particular in giving better performance over roberta with or without multitask learning 2 the proposed adaptor seems concise and easily expanded to incorporate other knowledge sources though there are few details which could help clarify things see 2 in next section 3 the probing task using lama to show how much factual knowledge has been memorized by the k adaptor roberta facadapter was well done and its discussion was very interesting cons 1 the proposed adapter solution is somewhat similar in nature to that of houlsby 19 and to a lesser extent cooper 19 httpsarxivorgpdf190202671pdf and it feels like an omission to not discuss houlsby 19 and make experimental comparisons against it discussing proscons more thoroughly especially since in the extensive experiments done in this work it is shown the linguistic adapter usually only adds a tenth of a percentage point when using roberta with a single factual adapter in this single adapter case then its not immediately evident how these models would differ and what the advantage is both houlsby and cooper are evaluated on the glue benchmark and provide code 2 i was a little confused as to how the adapters were specifically pretrained and it might be a question of figure 1b but also sections 33 and 34 could have been expanded to clarify it a bit it is my understanding that when pretraining the facadapter on the relation classification task for instance in section 33 for a given example in trexrc two entities and context are passed into roberta whose weights remain fixed while those of the kia units of the facadapter are updated and the final hidden representations of roberta and the facadapter are concatenated to form an input representation of the entities given their context and this used for the actual task is my understanding correct if so im confused as to how the subsequent pooling and concatenation actions are done clarifying this process for 33 and 34 would be beneficial for clarity purposes and its not discussed in the supplemental materials either 3 your roberta large baseline already beats most of what you are comparing against which is fine as your adapters give gains again particularly the facadapter but it also would have been interesting to see what sort of gains would have been achieved using a different less powerful model as the base roberta small or just plain bert and additionally some sort of ablation testing or explanation on the choices made for the adapter networks themselves ie n2 transformers etc hidden layer size etc though its possible this could be left for future work for clarity in figure 2 where you show n x transformer layer and n2 im assuming the first transformer layer feeds directly into the second transformer layer which then feeds into the up projection layer correct if so it might be better just to show two transformer layers like that instead and additionally naming the projection layers up and down projection layer respectively questions during rebuttal period please address and clarify the cons above small typos in abstract we propose kadapter which remains the original parameters remains keeps or retains in introduction they fail to continual learning fail at continual learning it remains the original representation remains leaves pg2 while remaining the original parameters of roberta frozen remaining keeping section 3 it remains the original representation remains keeps 31 different from houlsby et al 2019 add adapter layers in contrast to houlsby et al 2019 who add adapter layers 33 all relations having lees than lees less
### Summary: | the paper augments pretrained language models by introducing adapter where each adapter is another language model pretrained for a specific knowledge source eg wikidata and an objective eg relation classification the representation from each adapter is concatenated to the representation from the generic lm specifically they introduce two adaptors factual mostly derived from wikipedia and linguistic from dependency parser and the experiment shows modest improvements over various benchmarks this is a borderline paper as both methods and experiments are reasonable yet not very novel or strong the clarity of the paper can be improved as pointed by r1 and r4 without any mathematical notations model details are to be interpolated from figures the novelty is limited and experimental rigor can be improved ie for many settings gains are fairly small and no variance reported |
Below is given review of a research paper from cnoference journal. Please write a summary the review.
### Review:
this paper presents a formal analysis for the impact of graph reordering ie ordering the inmemory storage sequence of graph node embeddings on the cache efficiency of near neighbour searches using near neighbour graphs the connection of the graph ordering ie memory layout of the graph nodes and the cache complexity is formulated based on which the cache complexity of the gorder wei et al 2016 is analysed and two other graph ordering corder and porder are proposed experimental results on large real datasets confirm the effectiveness of the analysis and the importance of graph reordering for near neighbour search efficiency strengths 1 the paper has presented a solid analysis for the impact of graph reordering ie ordering the inmemory storage sequence of graph node embeddings on the cache efficiency of near neighbour searches using near neighbour graphs 2 experimental results on large real datasets are presented with an interesting discussion on the results 3 source code of the paper is available weaknesses 1 the proposed method corder is ineffective as discussed in the experimental results perhaps this method can be dropped to make room for adding more details on the experimental settings to the main content of the paper 2 the other proposed method porder is only somewhat better than the existing method gorder table 5 3 a lot of contents have been included as supplementary material which makes the paper somewhat difficult to follow typo degreebased groupingfaldu et al 2019 studies show that studies show that the paper included an interesting discussion on the results and limitations docsepthe paper studies practical performance of different methods for rearranging the node layout in the memory for graphbased approximate nearest neighbor search algorithms hnsw specifically it also proposes a simple modification of the existing methods based on query profiling the paper claims to have up to 3050 improvement in query latency on 100m datasets and is accompanied by code though not clear if it is going to be released strengths simple change that is very likely to be agnostic to the type of the graph algorithm used sizable gains from the algorithm on large datasets sourcecode hopefully will be released with the publication not clear from the text long discussions of the results weaknesses the paper does not have a clear description of the methods no pseudocode or even text description of stepbystep actions i guess the readers are suggested to follow the cited papers but imo there should be at least a sketch of the best solution the source of 1000 queries for porder not discussed in the paper were they used from the train set or the test set there is no implementation of porder in the code which is confusing the code also has bugs eg nonexistent openmp flag it does not compile without fixing the dependencies could had been fixed if you provide dockerfile and there are errors in its description judging the code the construction is done in a single thread if index construction time provided in the paper is for this regime which is not clear from the paper but seems to be the case it should be redone in the multithreaded regime i do not think there is any negative societal impact docsepthis paper proposes to use graph reordering to improve the cache locality of graphbased nearest neighbor search algorithms an analysis is conducted to show why graph reordering works and the experiments show that graph reordering significantly improves performance strength 1 graphbased nearest neighbor search algorithms are very popular and graph reordering improves the performance of graphbased nearest neighbor search algorithms 2 although the conditions are restricted the analysis explains why graph reordering improves performance 3 the profilingbased ordering scheme makes sense weakness 1 the authors should dig deeper to show why recording improves performance currently the explanations are vague eg due to software prefetcher and auxiliary functions the authors may want to make them more specific by explaining how software prefetcher works or conduct some experiments to show where the reduction comes from 2 i believe that graph reorder works for graphbased algorithms in general but it helps to show by experiments that graph reorder improves the performance for algorithms other than hnsw eg nsg or ngt 3 the legends in the figures are too small to read yes
### Summary: | this paper studies how to order inmemory sequences for graph embedding there was a positive consensus that the studied problem is interesting and results are sufficiently discussed there were some concerns on missing results which were addressed during rebuttals |
Below is given review of a research paper from cnoference journal. Please write a summary the review.
### Review:
this paper introduces the pacbayes information bottleneck pib starting from the generalization bound eq 4 which shows that the generalization gap is upper bounded by a function of iws the authors proposes pib which has an additional regularization term of beta iws since the computation of iws is intractable the authors then make several assumptions to simplify its computation arriving at an estimate of iws by tildeiws eq 15 and use sgld to compute it in practice experiments show that 1 there is a twophase transition in sgd training as indicated by tildeiws 2 tildeiws seems to correlate with the generalization gap under different variations of experiment hyperparameters number of hidden layers noise ratio and number of random label data 3 it improves performance compared to l2 and dropout regularization strength this paper addresses an interesting and important problem the proposed pib is novel the experiments show that the proposed tildeiws correlates with the generalization gap and helps improving the performance weakness in order to make the computation of iws tractable the authors make several important assumptions it would strengthen the paper a lot if the paper discuss and perform experiments to show if the assumptions are valid in the experiments the authors run furthermore in section 55 can the authors show the generalization gap together with the train and test acc with different regularization ideally we should see that with pib as the objective the generalization gap is much smaller than the other methods with this we can then be confident that the improvement is due to reduced generalization gap instead of better training in summary this paper is novel but the experiment should be strengthened as detailed in the main review docsepthis paper proposes a new version of the information bottleneck objective for training neural networks this is in part motivated by previouslyderived pacbayes bounds on the generalization error that are proportional to the square root of the mutual information of the weights and the training dataset iws thus this new information bottleneck objective attempts to minimize both the empirical risk and this mutual informationthe paper derives a computationallytractable algorithm for estimating iws then this algorithm is used to show that this quantity is inversely correlated with generalization loss on a variety of neural network architectures strengths the paper proposes an exciting general principle of deep learning as far as i know the contributions here are novel and will be of high interest to the community the authors build on previous work by showing how their ib objective addresses the shortcomings of previous work in this area it is very well written this is a highlytechnical paper and the details are presented in a careful and thoughtful way the experiments are well done and the results support the conclusions specifically this objective is motivated by a pacbound the tightness of which is not clear and various approximations are used to estimate iws the accuracy of these are not immediately clear the experiments address these issues by showing that the motivation and the approximations are reasonable weaknesses the limitations of this method are not discussed clearly for example the paper provides an algorithm for sampling from the weight posterior pws but how does this compare computationally to standard training of a neural network or a estimating the posterior in a bayesian neural network there are some minor grammatical and spelling typos throughout eg infection point an excellent paper with exciting ideas clear presentation and technical depth docsepthe authors propose new interpretation of the information bottleneck ib dubbed pacbayes information bottleneck pib where the ib is defined wrt the mutual information between feature representations t and either inputs x or targets y pib is defined wrt the empirical risk over the dataset s xi yii1n and the mutual information between model parameters and s imathbfw s or information stored in weights the authors show that pib is a bound on generalization error the authors derive a tractable estimator for imathbfw s the authors present an approximate inference method for pmathbfw mid s that utilizes the proposed pib the authors show that pib reflects the hypothesized twophase fitting and compression modes of neural networks across different activation functions network depth and network width they show that imathbfw s yields a good estimator of the generalization error that is robust to label noise they show that their inference method improves generalization across several benchmark datasets strengths to my knowledge this paper has significant technical and empirical novelty the authors do a good job of summarizing previous work and differentiating their contributions this is not my area of expertise but the derivation of pib the estimator for imathbfw s the proposed optimal posterior all look novel and correct the experiments are well done thorough and support the main claims of the paper weaknesses the main weaknesses of this paper are language and clarity i would recommend a thorough grammarly or perhaps external advice the graphs should report means and standard error intervals over multiple random seeds some specifics some technical terms are used before definition eg phase transition in abstract in the abstract ib cannot explain maybe ib theory can as you use in the intro iiw and imathbfw s are redundant would recommend just using imathbfw s third mutual information becomes trivial in deterministic cases please elaborate cite 2 we derive a solution to the intractable can something intractable have a solution maybe approximation is better optimal posterior of pib does pib have a posterior or is the posterior over the weights figure 2 iiw only shows compression phase can the loss also be included in these plots to my knowledge this paper demonstrates significant technical and empirical novelty i believe the main weaknesses can be addressed prior to publication therefore i recommend acceptance however i am not an expert on this topic so my confidence is only a 2 docsepthe authors propose a formulation of the information bottleneck problem replacing the mutual information between input x and latent representation z via the mutual information between the sample s and the weights w obtained from the sample they derive closedform solutions for this mutual information in the gaussian setting and propose an sgld scheme to optimize the objective using this objective and optimization algorithm the authors investigate several interesting scenarios including different activation functions and noisy labels the paper is generally well written and treats an interesting and timely topic the idea to limit the information about the sample that is contained in the weights is not new the authors cite several works that bound the generalization error via this information but this is the first time that i have seen a corresponding cost function implemented in practice there are however a few issues that are not perfectly clear to me the authors cite the literature stating that the generalization gap is limited by isw if the loss is sigmasubgaussian does this hold for the negative loglikelihood in 6 also in 6 is s a random variable or not 4 requires that isw is computed as an expectation over ps while the loglikelihood in 6 is an expectation over pws ie not over ps but over a concrete s how can this be understood connected to this is it safe to call the resulting cost function an information bottleneck cost function i assume that this is better called an iiwregularization rather than an ib cost the ib cost is a very specific formulation that combines a mutual information cost with a mutual information utility whereas here we have a general cost with an additional mutual information cost as regularization term the authors correctly claim that ixt becomes trivial if the network is deterministic more precisely this mutual information becomes infinite in many of these cases see learning representations for neural networkbased classification using the information bottleneck principle by amjad and geiger i believe that this result carries over to isw being infinite for deterministic learning algorithms this may not hold for all learning algorithms but certainly for some my own gut feeling suggests that isw is infinite for sgd with finitely many epochs eg by the fact that there are only combinatorially many options to shuffle the batches but that it is finite for sgld where noise is added to the weights it is therefore not clear to me in which settings the analysis in section 3 is a valid approximation in other words in which settings is the assumption that pws is gaussian valid does it only hold for sgld connected to the point above in which cases is the assumption that pw is gaussian a valid approximation can this gaussian assumption about pw be used to bound isw from above eg for a gaussian learning algorithm can it be shown that the term isw is maximized if w becomes gaussian as well this would be at least intuitive from a channel coding perspective where a gaussian channel input is known to maximize the mutual information through a gaussian channel and which is then known to produce a gaussian channel output in algorithms 1 pls compare line 9 with your equation 15 in 15 you sum over squared inner products in line 9 and 11 you square over the resulting sum of inner products is this difference intended and if so how can it be explained also do we have t0 ge t1 in algorithm 1 in fig 1 why is the mutual information iws evaluated for different layers what is the exact meaning of splitting the iiw between layers in terms of the generalization bound i was assuming that the generalization bounds all consider the entire set of weights and that the proposed pib should do so as well also in fig 1 the discussion of the inflection point is not fully clear in section 51 it is claimed that the variance of the information explodes can this be made more precise eg by writing down the mathematical symbol for this variance furthermore this is not shown in the figures if i remember correctly in all figures why is the mutual information isw so small these numbers do not seem right i would assume that it is necessary to learn more than 102 bitsnats to successfully solve a classification problem in other words while the general trend of iiw seems to be correct i am not convinced of the correctness of the absolute numbers can you provide some intuition about these small numbers is this connected with the proportionality symbol in 14 but going from 8 to 9 it seems to be that additive constants are dropped not multiplicative constants for the sake of clarity i would prefer that footnote 3 is in the main text also in some instances the notation and terminology is not clear eg is s sampled iid in 4 why is the oracle prior called an oracle how exactly is the bootstrapping resampling weight zetak defined why is the temperature beta called the annealing temperature just before 18 at the end of section 52 you write that the l2norm keeps increasing the norm of what a very interesting paper dealing with an interesting and timely topic unfortunately the paper is not perfectly clear throughout all sections
### Summary: | this paper revisits the information bottleneck principle but in terms of the compression inherent in the weights of a neural network rather than the representation this gives the resulting ib principle a pacbayes flavor the key contribution is a generalization bound based on optimizing the objective dictated by this principle which is then tractably approximated and experimentally verified reviews raise concerns about assumptions made to achieve the tractable version and a public discussion debates whether this is truly a pacbayes bound the authors address these adequately another concern is whether improvements in experiments can be ascribed to the new objective authors add new experiments in support of this additional concerns about the clarity of certain aspects of the paper were or were promised to be addressed by the authors overall the perspective of this paper its technical contributions and experimental evaluations appear to be worthwhile to share with the community as they advance the applicability of the information bottleneck principle |
Below is given review of a research paper from cnoference journal. Please write a summary the review.
### Review:
i liked this paper quite a lot although this paper does not belong to my area of expertise i was able to understand the paper clearly because of its lucid exposition experimentally the authors show a novel gnn design with an attention module that has comparable performance to the mlp and outperforms other gnn designs i believe that this will be a valuable contribution to many practical problems unfortunately this work does not have any theoretical results and evaluating the experimental results is outside my range of expertise therefore i would like to defer this paper to my fellow reviewersdocsepmain idea in this paper the authors study the problem of gcn for disassortative graphs the authors proposed the gnan method to allow attention on distant nodes indeed of limiting to local neighbors the authors generalized the idea of graph wavelet with mlp to generate the attention score and utilized it to generate multiple attention heads the authors carried out experiments on several realworld networks 4 assortative and 3 disassortative with comparison to several stateofart gcn methods strength the authors study a very interesting problem of gcngraph embedding or disassortative graphs the proposed method is well motivated with solid theoretical motivation from graph wavelets the proposed model is very intuitive generalization of graph wavelet methods the empirical evaluation is very thorough on seven networks with comparison to about 10 baselines of different kinds weakness though the authors mentioned the use of sparsification of attention for speedup however it mentioned that t is set to zero it is interesting to see how scalable the proposed method is as it needs to have global attention to possibly all nodes an empirical comparison of running time would be very helpful the authors only carry out experiments on three disassortative which are all very small it would be interesting to see more experiments on disassortative graphs alternatively it would be interesting to have an experiment on synthetic graphs where the beta can be controlled and varied smoothly to see how it affects the performance of different algorithms the authors picked only node classification of evaluation tasks it is interesting to see how the disassortative could impact other tasks like graph reconstruction and link prediction docsepthis work propose a new gnn architecture to help gnn break its limitation on only working over homophilic networks the technical is to use introduce graph global attention i think the paper is written okay the motivation is clear the solution is reasonable however i have following criticisms 1 this work has limited novelty observing that gcn cannot work well over heterophilic networks is not a new idea and observation using attention to capture the features from faraway nodes is natural but not novel i do not think that it is reasonable to argue against other works eg 1 that adopts the above idea by saying they are not expressive enough expressiveness sometimes may lead to model overfitting actually chevnet 2 can also capture faraways nodes and be expressive enough why does it not work well i guess that it is due to some overfitting issue moreover if i understand it correctly the limited difference between this work and 3 is most likely the global attention which has very limited contribution 2 although the work claims everywhere to tend to decrease the complexity when computing the global attention one still needs to do computation for every pair of nodes which is of course not scalable for even mediumsized graphs 3 the heterophilic networks used for evaluation are very small with only several hundred nodes why not try larger ones say actor cham in 4 i guess the computational issue comes from the global attention 1 nonlocal graph neural networks 2 convolutional neural networks on graphs with fast localized spectral filtering 3 graph wavelet neural network 4 geomgcn geometric graph convolutional networks postdiscussion update i would like to thank the authors for preparing the rebuttal and attending our discussion however i still think the complexity is a concern of this work i do not think that eq 3 can be implemented within the complexity that the authors claimed moreover if the authors use another way to compute the attention scores that way should be very clearly stated instead of written in a different form given the high complexity i cannot clearly see the advantage of this work in comparison to 1 as the nonlocal attention has been proposed in 1 already 1 nonlocal graph neural networks
### Summary: | this paper proposes a gnn that uses global attention based on graph wavelet transform for more flexible and datadependent gnn feature aggregation without the assumption of local homophily three reviewers gave conflicting opinions on this paper the reviewer claiming rejection questioned the novelty of the paper and the complexity of the global attention mentioned in the paper even through the authors responses and subsequent private discussions concerns about complexity and novelty were not completely resolved considering the authors claim that the core contribution of this paper is to design fully learnable spectral filters without compromising computational efficiency it is necessary to consider why it is meaningful to perform global attention based on graph wavelet transform in the first place in terms of complexity although the wavelet coefficient can be efficiently calculated using the chebyshev polynomials mentioned by the authors in the attention sparsification part n log n is required for each node in sorting resulting in complexity of n2 or more there may still be an advantage of complexity over using global attention in a messagepassing architecture but it will be necessary to clarify and verify that given that the proposed method uses an approximation that limits global attention within k hops also this paper modifies the graph wavelet transform in graph theory which requires a deeper discussion for example as the authors mentioned the original wavelet coefficient psiuv can be interpreted as the amount of energy that node v has received from node u in its local neighborhood the psiuv defined by the learnable filter as shown in equation 3 has a different meaning from the original wavelet coefficient there is insufficient insight as to whether it is justifiable to use this value as an attention coefficient overall the paper proposes potentially interesting ideas but it seems to require further development for publication |
Below is given review of a research paper from cnoference journal. Please write a summary the review.
### Review:
the paper proposes a novel framework for semisupervised learning that solves two issues of previous methods 1 overreliance on labeled data and 2 error accumulation it shows that jointly solving the main task together with another task that discriminates whether the data label is real or not leads to better performance strengths the proposed framework seems to be novel it works well in experiments on a wide range of tasks classification label propagation and data imputation it seems to be potentially beneficial for many domains since it does not have domainrestrictions while many previous ssl methods rely on certain image domain techniques such as consistency regularization and data augmentation weaknesses since the proposed method is only compared with the original pseudolabel method comparing with other extensions of pseudolabelling methods that are mentioned in section 5 will make the contributions more clear in addition to the papers mentioned in section 5 there are a few papers that try to address the error accumulation in semisupervised learning methods that is observed in pseudolabelling for example in defense of pseudolabeling an uncertaintyaware pseudolabel selection framework for semisupervised learning from iclr 2021 and repetitive reprediction deep decipher for semisupervised learning from aaai2020 questions i am not sure if i understood the experiments correctly as the missing rate goes higher do we have more unlabeled samples as explained in the last paragraph of page 6 or do we have more noisylabelled samples as explained in 1st paragraph of section 41 can we show the 3rd task data imputation in figures 2 to 4 one of the benefits of the method seems to be that it can be incorporated into a wide range of ssl algorithms i think the paper demonstrated that it can be used to enhance pseudolabelling method but what kind of other ssl algorithms can scl incorporate minor questions and comments ssl is a very hot topic and there has recently been many advances since the experiments do not compare with many of the recent works it would be better to emphasize why they were not compared for example section 1 has a discussion on how recent ssl methods utilize consistency regularization which relies on heavy data augmentation techniques that is only available in certain domains what kind of value for parameter alpha is used in the image classification for the other two tasks i think the appendix explains that alpha is 1 if we are given a labeled dataset l and unlabeled dataset u it seems we can automatically construct vector m which is explained in end of page 2 if this is correct then why do we need m as an input in algorithm 1 in page 6 what is p introduced in the beginning of section 22 it seems like it is a set from the p in p notation but since it compares with m in the loss function it also looks like a vector typo perforamnce in page 6 should mi mj in the beginning of page 3 be mi mj is y a label space y in y or is it the full set of labels in the training dataset y yl cup yu ideally it would be better to perform several trials and report meanstandard error in table 1 after rebuttal thank you for answering my questions the additional experiments are helpful to have a better understanding about the proposed method it looks like the advatangeous points of the proposed method is now about the low computational costs according to the new experiments including ups rather than better performance although this still may be beneficial for the research community it seems to be slightly less significant and also may affect the storyline i would like to also recommend to put the new experiments with ups in the main paper instead of the appendix the proposed method seems to have some nice benefits but i feel there are a few weaknesses that should be addressed i also have a few questions and it would be helpful if the authors can take a look at the previous section main review docsepthe paper introduces selfinterested coalitional learning scl which is a novel approach to semisupervised learning scl combines the traditional selftraining approach to semisupervised learning with an auxiliary task that infers label observability the empirical results show that in a variety of scenarios scl outperforms both selftraining and the original model this is an interesting paper on a topic with important practical applications semisupervised learning the contribution appears to be original and it is likely to influence future work in the field the authors are explicitly calling out and addressing the two main weaknesses of traditional selflearning approaches error accumulation and overreliance on the labeled data the paper would greatly benefit from an additional section that would provide an intuitive illustrative example of how and why the proposed approach outperforms selftraining ideally it should compare and contrast the convergence of 1 self training 2 the auxiliary task and 3 scl the paper would also benefit by tightening the narrative around the alpha parameter which in the main paper is only discussed in the theoretical framework appendix a provides no value of alpha for the first dataset and it proposed without any justification a value of 1 for the other two domains appendix b is extremely brief and not very helpful the authors make no recommendation on how to tune alpha and the argument that even the worst alpha in the 01 09 range is better than the original model is fairly weak given the wide variations of the accuracy due to changes the value of alpha other comments for table 1 please add three more rows 0 90 and 99 the former is critical to understanding the upperbound performance while the later two will bring scl into a more realistic semisupervised regime where unlabeled data is one or two orders of magnitude more abundant than the labeled data please add to figure 6 the horizontal lines with the accuracy of the original model for each of the three missing rates it is still unclear why did you choose to use only 10of the data for image classification page 6 is scalability to large datasets a concern please spellcheck the paper eg perforamnce on page 4 page 2 please replace more sufficient page 3 jointly solving above two tasks jointly solving the above two tasks page 3 there are some other works embody there are some other works that embody page 4 are impacted the influence are impacted by the influence page 7 please replace well learn overall this paper uses a novel idea to improve the state of the art for semisupervised training docsepthis paper proposes a new semisupervised learning method motivated by the error accumulation problem of typical selftraining paradigms the authors propose to explicitly model the confidence of pseudo labels as an auxiliary task they come up with a selfinterested coalitional learning scl strategy to solve both tasks jointly under the new framework the main task is transformed into a costsensitive learning problem experiments demonstrate that pseudo labels are substantially more accurate with the new method and better performance of the main tasks at different label missing rates pros overall the paper is wellstructured and easy to follow the new method achieves its original goals and improves ssl effectiveness by jointly solving the main and the auxiliary tasks the authors introduce a new scl strategy to solve the problems which can be applied to a broader class of learning problems cons lack of experiments the proposed method is only compared with the selflearning method with the same base learner while this demonstrates how the model is improved with scl it is also necessary to compare with stateofart ssl methods its also valuable to include the supervised method with fullylabeled dataset as a reference in all experiments for data imputation a more common case is that missing state is correlated with inputoutput instead of simply random missing it also checks the method robustness against labeledunlabeled distribution shift compared with original selflearning method the new method has an extra discriminator model which are based on the same base learners as for the main tasks its meaningful and more fair to compare with supervised models of higher capacity the paper doesnt cover how scl can work together with consistency regularization which is commonly used together with selflearning besides i have a few questions although table 1 doesnt have a row for missing rate 0 full dataset it seems scl methods have better accuracy than model trained with full dataset for the first two tasks is this because the scl has double model capacity due to the extra discriminator why is the test accuracy of pseudolabels 100 for scl method in figure 4 are they calculated differently this is an interesting paper from technique perspective but it definitely needs more empirical studies to demonstrate practical value docsepthis paper proposes a new semisupervised learning framework by introducing an auxiliary task that distinguishes whether the pseudolabels are truly labeled or not then this information is used to add a reweighting loss to the main objective experiments on several simple benchmark datasets show that the proposed method outperforms some naive baselines the idea of introducing the auxiliary task that discriminates whether an instance is labeled is quite interesting in effect such a strategy is first introduced in active learning 1 in the vaal method 1 a similar discriminator is introduced to identify where an example is labeled or not which is then used to indicate the uncertainty of an example for active selection therefore the proposed method has a close connection to a recent work in ssl 2 that also employs the uncertainty measure to select highquality pseudolabels i have the following concerns 1 the derivation in section 32 is confusing for example in eq 3 the second equality is incorrect and the term fracdddx should be added also it would be better to change the notation of d discriminator to another one since derivative dx also uses the notation d besides i actually did not understand why mathcallb depends on f since f and mathcallb are from two different branches without sharing network blocks figure 1 2 the experimental section is not convincing and this is my main concern the datasets and the baselines are too simple stateoftheart ssl methods should be employed to support the claims in particular the uncertaintybased ssl method 2 should be compared as i have discussed above the proposed method can implicitly be equal to existing techniques in ssl is the proposed method complementary to existing methods or it is contradictory to some techniques these questions require an indepth empirical analysis overall this work is below the bar of an iclr paper regarding its poor experiments 1 sinha s ebrahimi s darrell t variational adversarial active learningcproceedings of the ieeecvf international conference on computer vision 2019 59725981 2 rizve m n duarte k rawat y s et al in defense of pseudolabeling an uncertaintyaware pseudolabel selection framework for semisupervised learningj arxiv preprint arxiv210106329 2021 interesting idea poor experiments and confusing derivation
### Summary: | this paper proposes a new method for the important problem of semisupervised learning this method relies on an auxiliary task label observability prediction to weight the examples according to the confidence in their pseudolabels so as to avoid the propagation of errors encountered in selftraining limited experiments show that the proposed method can compete with other methods in terms of performance or training time on the positive side all evaluators agree on the potential value of the proposed approach which is generic in nature on the negative side the experimental evaluation although strengthened during the discussion is not yet strong enough to have really convinced of the real merits of the method in particular comparisons with the state of the art still need to be improved in addition the paper would benefit from some rewriting in particular of the mathematics eg the d notation for task b should be avoided as suggested by one reviewer there is a misplaced partial derivative in equation 6 the authors could also simplify their derivation by using the envelope theorem i therefore recommend rejection with an encouragement to strengthen the experimental part and to improve the derivation of the proposed method |
Below is given review of a research paper from cnoference journal. Please write a summary the review.
### Review:
the main goal of this paper is to introduce a simple methodology for optimizing transformer based models for efficiency and effectiveness the paper introduces two main ideas 1a topdown strategy for pruning components of a transformer model given a specific focus say speed the strategy is to consider pruning large coarsegrained components first followed by smaller finergrained components the pruning decision is made based on a significance analysis a component is considered significant for pruning if it from the model does not result in a substantial increase in the models loss as decided by a pruning threshold 2 pruning and approximating techniques for different components for example feedforward networks are pruned by removing weights in groups determined via a hyperparameter for approximating selfattention a signmatching technique for deciding which top k keys to use for computing query x key dot products the main strengths of this work are as follows 1 the techniques do not require training networks from scratch and can be applied directly during finetuning 2 the techniques are simple and should apply widely to most transformerbased models 3 the empirical results support the claim that the technique can yield significant speedup and memoryreductions while maintaining accuracy and even provide improvements in accuracy if that is the pruning goal they show that technique is orthogonal to other models explicitly designed for speed and memory footprint q8bert distillbert and can provide further improvements in both efficiency and effectiveness 4 this is a practical and useful approach that should be widely applicable along with many useful insights about optimizing transformerbased systems i appreciate that the experimental results are reported with averages across multiple runs i dont see any major weaknesses in the paper here are some areas that can be improved 1 the description of the pruning strategies was hard to follow and needed to be tightened up possibly adding equations and some pseudocode to the description should help 2 i am curious to know what components get pruned cross the different models that were optimized i wonder if there are systematic differences between original and distilled models and between autoregressive gpt and autoencoding style models 3 also some level of ablation analysis on the strategies used will be helpful for example if the elements were not ordered based on the granularity would the results be any different since this is an iterative strategy the order should play an important role in selection and utility of the subsequent pruning steps same goes for the set of pruning strategies a related question would be what gives the biggest gains 4 what is the impact on the finetuning time the baseline only requires one finetuning pass does this method require multiple finetuning passes or can the loss thresholds be computed on a smaller subset of the target data this may be a good future work to look into for tasks where the training data is relatively large where one cannot afford to exhaustively search through all the pruning strategies docsepafter reading the rebuttal some of my concerns are addressed by the additional experiments but i also agree with other reviewers that the result is not very surprising as r4 mentioned the proposed method depends on the a specific downstream task where the small general bert can be further pruned for a fair comparison to previous work baselines that are applied to a specific finetuning task need to be compared this paper presents a new framework for creating small finetuned pretrained language models the framework has 3 components 1 a set of transformer components to be pruned 2 a significant analysis for identifying unimportant elements 3 a set of techniques to prune or approximate the transformer element pros 1 the framework is very adaptive by considering different basic elements of the transformer 2 the framework is very efficient by removing large components eg layers attention blocks ffd layers at first and small components eg weight group later 3 the framework gathers multiple different pruningapproximation techniques and tries to explore the limit of pruning pretrained models which is appreciated consquestions 1 is the loss used in significant analysis computed using the development set if the validation loss is used the experiment results in table 1 are not reliable 2 there are many bert pruning papers providing comparison to these papers is very important to evaluate the proposed method can the model prune more weight at the same performance level or can the model perform better at the same pruning ratio 3 it is also helpful to present how much computing resource is needed to prune the network eg how many prunefinetune cycles are needed 4 lack of results of pruning bertbase on glue which is a very standard and common setting 5 in figure 3 why q8bert speed focus is even largerslower than q8bert with the same speed q8bert speed focus is significantly worse than q8bert minor page 5 less the minimum loss seen less than the minimum lossdocsepthis paper presents a method for improving a fineturned transformer in terms of a specific metric such as size speed or accuracy the candidates of removed elements are considered hierarchically with some heuristics and are evaluated in terms of training and validation loss to determine whether they should actually be removed from the model the authors apply their method to several stateoftheart transformer models and show that they can produce fast and compact models without losing much accuracy although the individual techniques employed to realize the whole pruning process are not particularly novel the paper presents a wellthoughtout approach to combine those and reports very promising experimental results i think this is a nice contribution to the community given that the computation cost is increasingly important in dealing with bertlike models it seems to me that the authors used transformers whose weights are shared between different layers like universal transformers or albert maybe i missed something but i think the authors should clarify if this is really the case in the manuscript the entire process of pruning is a bit vague and hard to replicate would it be possible to describe the whole process in pseudo code is algorithm 1 the whole process i think the authors should also describe the computational cost or maybe wallclock time required to perform the proposed pruning processes it seems to me that the search space is rather large and requires a considerable amount of computation p5 we prune the element only if the trainingvalidation loss i think you should be more specific here how did you actually use both the training and validation loss why do you need to look at the training loss when you are interested in the generalization error p5 weight groups of wn why is this wn i thought this should be w minor comments p5 less the less than the p6 doesnt does not p6 attention attention p7 second order secondorder docsepthanks to the authors for the detailed feedback i still have concerns about the clarity of the presentation and some contributions of the papers are not strong enough so ill keep my score summary this paper presents a framework to systematically perform pruning and layer approximation the framework includes a queue of potential elements for compression at each time step the framework will evaluate the head element of the queue try to prune the whole element or perform approximation quantizing pruning attention heads and approximating with signmatching attention and keep the transformation only if the loss in performance is acceptable the paper performs experiments with various models on glue and shows speedups or compressions compared to the original model reasons for score the techniques used in the paper are not novel and the choices on how to apply multiple compression techniques need more justification the experiment results are okay but not surprising the presentation of the paper needs to be polished see below for more details pros 1 i like the insight that approximate finetune approximate cycles doesnt work for finetuning 2 i like the insights used to determine which elements to be examined first start from the larger blocks and later layers i hope this point can be emphasized more and compared with more bruteforce and lessefficient algorithms for example for each round one can choose a layer that causes the least loss of performance to prune you can compare your greedy algorithm with this algorithm to show the gain of using a less efficient algorithm is not significant 3 the signmatching attention proposed in the paper is new i would like to see more emphasis and ablation studies on the effectiveness of this special module cons 1 it is wellknown that compressing the model is easier during the finetuning phase 1 2 i dont think this should be a contribution to emphasize for the paper 2 the whole compression framework has a single global error bound combining this with the greedy layerbylayer approach taken by the framework will the following case be possible a layer that is early in the queue causes a huge drop of accuracy and thus makes all the future layers impossible to remove because the global error bound has been reached a better way is to only remove the layer with the lowest loss reduction it will be better to justify this point with an ablation study or at least show the final pruned model doesnt have this issue in the paper 3 at the end of page 5 when optimizing for speed however removing weight groups with low significance from arbitrary locations does not help since it introduces unstructured sparsity in the weight matrix that can be difficult to exploit to achieve speedups its true that if you remove random entries in a matrix will not help for the actual speedups but you can remove an arbitrary set of rows of the matrix and then restructure the weight matrix ie concatenate all the remaining rows to form a new matrix to make it efficient for modern parallel hardwares 4 i dont really understand the point of using accuracy as the final goal if the framework is for compression the goal should be about speedup or size if accuracy really matters it should be enforced as the threshold instead of as the final goal also i dont see the difference in the framework between using speedup or size as the goal since all the thresholds are defined by loss 5 the results in the paper are okay but compared to previous works in computer vision 3 it seems that the model size can be further compressed 6 there are multiple places where the presentation can be improved a its more clear to use a pseudocode instead of a diagram in figure 2 b it should be more clear to present table 1 as multiple tables c its better to put the results comparing with previous works in a table in the middle of page 8 minor comments on page 5 3rd paragraph from the bottom less the minimum loss less than minimum loss references 1 jiao xiaoqi et al tinybert distilling bert for natural language understanding arxiv preprint arxiv190910351 2019 2 shen sheng et al qbert hessian based ultra low precision quantization of bert aaai 2020 3 han song huizi mao and william j dally deep compression compressing deep neural networks with pruning trained quantization and huffman coding arxiv preprint arxiv151000149 2015
### Summary: | this paper introduces a set of techniques that can be used to obtain smaller models on downstream tasks when finetuning large pretrained models such as bert some reviewers have noted the limited technical novelty of the paper which can be seen more as a combination of existing methods this should not be a reason for rejection alone but unfortunately the results in the experimental section are also a bit weak eg see 14 there are not very insightful analysis and it is hard to compare to existing work for these reasons i believe that the paper should be rejected 1 dynabert dynamic bert with adaptive width and depth 2 training with quantization noise for extreme model compression 3 mobilebert a compact taskagnostic bert for resourcelimited devices 4 squeezebert what can computer vision teach nlp about efficient neural networks |
Below is given review of a research paper from cnoference journal. Please write a summary the review.
### Review:
the paper is a natural extension of 1 which shows the importance of spectral normalization to encourage diversity of the discriminator weights in a gan a simple and effective parametrization of the weights similar to svd is used w usvt is used along with an orthonormal penalty on u and v and spectral penalty to control the decay of the spectrum unlike other parametrizations of orthogonal matrices which are exact but computationally expensive the proposed one tends to be very accurate in practice and much faster a generalization bound is provided that shows the benefit of controlling the spectral norm experimental results show that the method is accurate in constraining the orthonormality of u and v and in controlling the spectrum the experiments also show a marginal improvement of the proposed method over sngan 1 however the following it is unclear why one would want to control the whole spectrum when theorem 2 only involves the spectral norm in 1 it is argued that this encourages diversity in the weights which seems intuitive however it seems enough to use spectral normalization to achieve such purpose empirically according to that same paper it would be perhaps good to have an example where sn fails to control the spectrum in a way that significantly impacts the performance of the algorithm while the proposed method doesnt overall the paper is clearly written and the proposed algorithm effectively controls the spectrum as shown experimentally however given that the idea is rather simple it is important to show its significance with examples that clearly emphasize the importance of controlling the whole spectrum versus the spectral norm only revision figure 1 is convincing and hints to why sngan acheives slow decay while in principle it only tries to control the spectral norm i think this paper is a good contribution as it provides a simple and efficient algorithm to precisely control the spectrum moreover a recent work 2 theorem 1 provides theoretical evidence for the importance of controling the whole spectrum which makes this contribution even more relevant 1 t miyato t kataoka m koyama and y yoshida spectral normalization for generative adversarial networks feb 2018 2 m arbel d j sutherland m bin kowski and a gretton on gradient regularizers for mmd gans nips 2018 docsepthe paper builds on the experimental observations made in miyato et al 2018 in which the authors highlight the utility of spectral normalization of weight matrices in the discriminator of a gan to improve the stability of the training process the paper proposes to reparameterize the weight matrices by something that looks like the singular value decomposition ie w u e vt four different techniques to control the spectrum of w by imposing various constraints on e have been discussed for maintaining the orthonormality of u and v penalties are added to the cost function the paper also derives a bound on the generalization error and experimentally shows the desirable slow decay of singular values in weight matrices of the discriminator other experiments which compare the proposed approach with the sngan have also been given 1the paper puts a lot of stress on the stability of the training process in the beginning but clear experiments supporting their claim related to improved stability are lacking 2it would be helpful for the readers if more clarity is added to the paper with respect to the desirability of slow decay of singular values and spectral normalization 3the point regarding convolutional layers should be part of the main paper docsepthis paper proposes to parameterize the weight matrices of neural nets using the svd with approximate orthogonality enforced on the singular vectors using orthogonal regularization as opposed to eg the cayley transform or optimizing on the stiefel manifold allowing for direct efficient control over the spectra the method is applied to gan discriminators to stabilize training as a natural extension of spectral normalization this method incurs a slight memory and compute cost and achieves a minor performance improvement over spectral normalization on two benchmark image generation tasks im a bit back and forth on this paper on the one hand i think the ideas this paper proposes are very interesting and could provide a strong basis off which future work can be builtthe extension of spectral normalization to further study and manipulation of the spectra is natural and very promising however the results obtained are not particularly strong and as they stand do not in my opinion justify the increased compute and memory cost of the proposed methods the papers presentation also wavers between being strong there were some sections i read and immediately understood and impenetrable there were other sections which i had to read 510 times just to try and grip what was going on ultimately my vote is for acceptance i think that we should not throw out a work with interesting and potentially useful ideas just because it does not set a new sota especially when the current trend with gans seems to suggest that top performance comes at a compute cost that all but a few groups do not have access to with another editing pass to improve language and presentation this would be a strong relevant paper worthy of the attention of the iclr community my notes the key idea of parameterizing matrices as the svd by construction but using a regularizer to properly constrain u and v instead of the expensive cayley transform or trying to pin the matrices to the stifel manifold is very intriguing and i think there is a lot of potential here this paper suffers from a high degree of mathiness substituting dense notation in places where verbal explanation would be more appropriate there are several spots where explaining the intuition behind a given idea particularly when proposing the various spectrum regularizers would be far more effective than the huge amount of notation in the authors defense the notation is generally used as effectively as it could be my issue is that it often is just insufficient and communication would be better served with more illustrative figures andor language i found the way the paper references figure 1 confusing the decays are substantially different for each layerare these all supposed to be examples of slow decay layer 6 appears to have 90 of its singular values below 05 while layer 0 has more than 50 if this is slow decay what does an undesirable fast decay look like isnt the fast decay as shown in figure 2 almost exactly what we see for layer 6 in figure 1 what is the significance of the sharp drop that occurs after some set number of singular values the figure itself is easy to understand but the way the authors repeatedly refer to it as an example of smooth singular decays is confusing what is doptimal design this is not something commonly known in the ml literature the authors should explain what exactly that doptimal regularizer does and elucidate its backward dynamics in an appendix if space does not permit it in the main body does it encourage all singular values to have similar values does it push them all towards 1 i found the brief explanation encourages a slow singular value decay to be too briefconsider adding a plot of the doptimal spectrum to figure 1 so that the reader can easily see how it would compare to the observed spectra ideally the authors would show an example of the target spectra for each of the proposed regularizers in figure 1 this might also help elucidate what the authors consider a desirable singular value decay and mollify some of the issues i take with the way the paper references figure 1 the explanation of the divergence regularizer is similarly confusing and suffers from mathiness a fact which i believe is further exacerbated by its somewhat odd motivation why if the end result is a reference curve toward which the spectra will be regularized do the authors propose 1 a random variable which is a transformation of a gaussian 2 to take the pdf of that random variable 3 discretize the pdf 4 take the kl between a uniform discrete distribution and the discretized pmf and 5 ignore the normalization term if the authors were actually working with random variables and proposing a divergence this might make sense but the items under consideration are singular values which are nonstochastic parameters of a model so treating them this way seems very odd based on figure 2 it looks like the resulting reference curves are fine but the explanation of how to arrive there is quite convolutedi would honestly have been more satisfied if the authors had simply designed a function a polynomial logarithmic function perhaps with a hyperparameter or two to control the curvature our experimental results show that both combinations achieve an impressive results on cifar10 and stl10 datasets please do not use subjective adjectives like impressive a 65 improvement is okay but not very impressive and when you use subjective language you run the risk of readers and reviewers subjectively disagreeing with you as is the case with this reviewer please also fix the typo in this sentence it should at least be achieve impressive results or achieve an impressive improvement on section 3 what is generalization supposed to mean in this context its unclear to me why this is at all relevantis this supposed to be indicating the bounds for which the discriminator will correctly distinguish real vs generated images or is there some other definition of generalization which is relevant does it actually matter for what we care about training implicit generative models what exactly is the use of this generalization bound what does it tell us what are the actual situations in which it holds is it possible that it will ever be relevant to training gans or to developing new methods for training gans experiments i appreciate that results are taken over 10 different random seeds if the choice of gamma is unimportant then why is it different for one experiment i found footnote 4 confusing and contradictory for figure 3 i do not think that the margin is significantit constitutes a relative 65 improvement which i do not believe really justifies the increased complexity and compute cost of the method i appreciate table 1 and figure 4 for elucidating a how orthogonal the u and v matrices end up and b the observed decay of the spectra appendix please change table 7 to be more readable with captions underneath each figure rather than listed at the top and forcing readers to count the rows and match them to the caption what is the difference between sngan and spectral norm in this table or is that a typo and it should be spectralconstraint i would like to see a discussion of table 7 interpretation of why the spectra look that way and why they evolve that way over training for each regularizer minor typos and grammatical mistakes throughout as per the cifar10100 website httpswwwcstorontoedukrizcifarhtml the torralba citation is not the proper one for the cifar datasets despite several recent papers which have used it intro last paragraph generation bound should be generalization bound page 4 paragraph 2 last sentence problem is misspelled
### Summary: | all the reviewers agree that the paper has an interesting idea on regularizing the spectral norm of the weight matrices in gans and a generalization bound has been shown the empirical result shows that indeed regularization improves the performance of the gans based on these the ac suggested acceptance |
Below is given review of a research paper from cnoference journal. Please write a summary the review.
### Review:
the paper is about a method for synthesizing binaural audio from a mono recording of a single speakers speech first i think the title is too general the paper does not attempt to convert all possible sounds but it tries to convert a single speakers monaural speech signal to binaural audio where the speaker is moving i think this inherent assumption is important since the method will probably not work for multiple overlapping audio sources i suggest changing the title to neural synthesis of binaural speech of a single moving speaker the first part of the network neural time warping is an interesting component that is capable of adjusting the delays conditioned on the location and orientations of the source and microphone such that a location dependent binaural audio is formed by estimating timevarying delays of the original mono recording separately for two channels it is believable that such a module would be helpful for a single moving speaker however such a model would not help or work when there are more than two active audio sources a separation module would be required for that scenario neural time warping is an autoregressive model which can work online the second stage convolutional network which uses conditioned hyper convolutions is also an interesting architecture that takes the warped signals and applies timeconvolutions with kernels obtained from the conditioning input which has the timevarying locations and orientations of the source and the microphone the section about the loss function is also interesting in that the time domain l2 loss is shown to not work well for accurate phase estimation so the authors propose to add a separate phase loss term to compensate for that i think it would be better if figure 2 is replaced with a plot of epsilonyhat versus amplitude error divided by yhat in a and versus the phase error in b it could be clearer than the current 2d figure which is hard to interpret the use of sine activation is not well justified sine activation is useful in the first layer of a signal representation network which is different from a signal prediction network i do not see how and why that could be helpful here in terms of comparisons 25d method uses visual information as conditional information to generate complex masks to produce binaural audio in this paper visual information is replaced with the spatial conditioning information it would help to get more information about the window size and hop size used in 25d since they may be an important factor that relates to the amount of delays they can introduce for wavenet comparison it was not clear how the wavenet model was trained to generate binaural data did it use any conditioning information if so how was it applied in an autoregressive way with randomized sampling the wavenet audio example sounded noisy which is not typical of wavenet generated audio it looks like the dsp method can utilize a listener specific hrtf which may be difficult to incorporate for the proposed neural model is it an important issue how does the model generalize to unseen speakers and rooms the training and testing strategy uses the same room and the same speakers would we have any problem when the monaural audio is recorded in some other room with some other speaker in figure 8 maybe it is ok not to draw the original binaural signal for every method in general i liked the neural warping and conditional convolutional components which are interesting and i liked the analysis about the loss function the approach is an interesting way to obtain binaural version of a monaural mono speaker recording in a room the dataset produced for the paper would also be useful for research update after revision the revision improved the paper thanks for taking care of my comments justification of sine activations generalization to unseen speakers experiment are nice additions the new title is a bit better and i think it may be ok since the goal is to perform a moving source simulation for single speech sources multiple speech sources can be simulated separately and added together as mentioned the authors may consider possibly a better name neural binaural synthesis from mono speech which emphasizes that the synthesized target is binaural speech from a single speech recording just a few more points 1 i think it is essential in wavenet to apply the model in an autoregressive fashion over samples just using the network architecture and the loss function from wavenet is not equivalent to using a wavenet model since an essential part of the model is the autoregressive sampling which makes sure the samples are dependent and coherent without autoregressive sampling the resulting sound is poor as observed by the authors so i suggest to emphasize that autoregressive sampling is not performed in the paper to avoid misleading the readers 2 more explanation of 25d is appropriate one wonders if using a larger stft window size would improve its resultsdocsepstrengths 1 the paper is well written it includes clear math notations and figures readers can easily follow the thought process of the authors for example figure 2 shows the relation of l2 loss and phase loss with respect to target energy indicating the importance of penalizing phase loss in the end to end system the same observation is supported by figure 3 2 strong results the proposed end2end model significantly outperforms previous sota in terms of objective measures and subject tests the video demo is very convincing the model improved spatialization and sound quality 3 high novelty this paper proposes to impose monotonicity and causality to the learned warping function which incorporates the physics of sound propagation i am excited to another example of applying domain knowledge to an endtoend model the model includes two novel components the neural warp network compensates the errors from geometry warp the temporal convolution works as a post processing module to account for reverberation and other effects ablation study shows both components are critical to be improved 1 the caption for figure 4a seems to be incomplete 2 it would be good to include a table to compare the proposed model with baselines in terms of model size and inference speeddocsepthis paper presents a neural networkbased model to generate binaural audio given a singlechannel audio and positions of sourcelistener their angles the authors developed a dataset of binaural audio which will be released on acceptance technical details and model architecture are available in the body of the paper whereas additional details such as baseline dspbased approach proof and dataset are available in the appendix the model was evaluated using the dataset developed in this work a demo video demonstrating the capability of the model is also provided as a supplementary material there are a few parts need to be addressed 1 it is unclear why dtwbased warping is required iirc the warpfield here can represent not only a shift but also other monotonic causal such as repeating if there is only delay between left and right just having a shift is enough isnt it it is great if the authors can explain the motivation to use warpfield more clearly 2 the use of hyperconvolution is an interesting idea the equation 5 uses conditional temporal convolution however audio generative models such as wavenet uses a different architecture gated convolution the gating mechanism can give additional nonlinearity and so im wondering if you can evaluate the performance of hyperconvolution against gated convolution 3 too large confidence intervals in table 4 although there were many evaluations the confidence intervals were pretty large and there were overlaps among them eg small overlap between dsp and ours in cleanliness large overlaps between spatialization and realism between dsp and ours with this result it is difficult to claim that there was a significant improvement over the baseline system please check your results and design the experiment more carefully to figure out whether there is any significant difference between them conducting sidebyside comparision is one possiblity comments this paper claims that it works in real time but no information about speed such as realtime factor hardware specification are provided sampling rate information is not explicitly provided in the experiment section 06 mos difference is large not a bit modern wavenet models often use mixtureoflogistics refer to parallel wavenet paper for details as output rather than mulaw to achieve better quality
### Summary: | interesting method for binaural synthesis from moving monoaudio nice insight into why l2 isnt the best loss for binaural reconstructions interesting architectural choice with nice results nicely motivated and clearly presented idea especially after addressing the reviewers comments i agree with the idea of a title change while i think its implied that the source is probably single source making it explicit would make it clearer for those not working in a closely related topic hence neural synthesis of binaural speech from mono audio as suggested in the review process sounds quite reasonable |
Below is given review of a research paper from cnoference journal. Please write a summary the review.
### Review:
summary the paper considers the adversarial attacks via a surrogate model constructed using data from a different domain the authors propose a defense from such attacks by a special kind of adversarial training inspired by the idea of domain adaptation the idea can be useful but raises a lot of questions especially when looking at the evaluation of the proposed approach reasons for score i vote for a reject as some findings are intriguing while the experimental results are questionable the first major concern is why do authors consider nlp models and attacks in the paper it is much easier to work with image datasets and if the general idea is new i suggest to start from this point to verify that the considered domain adaption works well in this scenario also the proposed attack is not new it is just a surrogate model attack but using a surrogate model training on the data from a different domain as the authors suggest due to the unavailability of the initial domain data also for this new attack the authors dont compare a surrogate model attack trained using the same domain data which would be interesting to compare the authors use only one dataset which is a bit strange for modern papers for this dataset they dont provide a full study limiting the scope of experiments to particular pairs of sourcetarget domains from the paper it is not clear how widely applicable are obtained results the comparison is not full there are a lot of options to be tuned for alternative approaches like adversarial training or other defenses the hyperparameter selection for them has a crucial effect on their success so we cant say that the proposed approach works better than others major concerns only one dataset considered i think that the inclusion of additional datasets at least three would improve the paper and make the conclusion by the authors more solid usage of surrogate models trained on other dataset is not new for general adversarial attacks 1 mentioned in the paper 2 and for adversarial attacks in nlp 3 lstm is not the state of the art model for the processing of nlp data 42 what attack do you use not explicitly specified so the results cant be verified by replication of the described experiments table 2 will benefit from adding afterattack accuracy for the original domain if it is similar to the presented accuracies then why bother with a new method table 3 comparison is not fair as we have no details about training for each approach eg we dont know how many additional examples we add during adversarial training also note that the stateoftheart for adversarial training is different from described in the paper see 4 5 table 4 afterdefense accuracy for what model is presented because it should be different for lstmgrucnn attack model tables 234 i suggest to keep the list of pairs target domain substitute domain similar for all tables to be sure that the presented examples are not cherrypicked also please consider running your approach on all pairs target domain substitute domain and aggregating all these results domain adaptation models from my experience are not easy to train it is interesting to access the quality of the models for different runs of learn2weight is it stable etc 1 nicolas papernot patrick mcdaniel and ian goodfellow transferability in machine learning from phenomena to blackbox attacks using adversarial samples arxiv preprint arxiv160507277 2016a 2 cheng s dong y pang t su h zhu j 2019 improving blackbox adversarial attacks with a transferbased prior in advances in neural information processing systems pp 1093410944 3 fursov i zaytsev a kluchnikov n kravchenko a burnaev e 2020 differentiable language model adversarial attacks on categorical sequence classifiers arxiv preprint arxiv200611078 4 shafahi a najibi m ghiasi m a xu z dickerson j studer c goldstein t 2019 adversarial training for free in advances in neural information processing systems pp 33583369 5 aleksander madry aleksandar makelov ludwig schmidt dimitris tsipras and adrian vladu towards deep learning models resistant to adversarial attacks iclr 2017 proposed minor improvements table 1 demonstrates one example that breaks the semantics of the attacked sentence can you provide good examples of why your approach work definition 1 is not a definition is x one instance or many instances in this definition also not specified that x and x should be similar equation 1 why you avoid standard number of equations beginequation labeleqsampleequation sample text endequationdocsepsummary this paper is about generating adversarial examples for some target model and protecting from such attacks authors consider a setting when an adversary has access to some similar to target domain data and can use this data to generate a surrogate model using this surrogate model an adversary can generate adversarial examples that apparently also fool the target model then authors also propose a defense mechanism from this type of attack learn2weight this is a learnt network that for a given example returns perturbation of weights to the target model which will be applied to the target before inference this model is trained by a defender on synthetic domains generated as perturbations to the target data overall this type of an attack is interesting the paper is well organized and written and easy to follow enough background is given for a reader to follow without the need to research around or going to appendix well done on clarity i do have a problem understanding how effective this attack is compared to other blackbox attacks and how the proposed defense compares to standard domain generalization methods like learning domain invariant features 1 one concern i have is about practicality an availability of such similar domains for testing authors used amazon multidomain sentitment classification where domains are easily available but how would you attack a pretrained imagenet for example what domains are similar and further more how much data for these similar domains you need to have to train a good enough surrogate model also you dont really have a way to calculate that your data is close to the actual target data 2 definition 2 fa wt fsa requires an access to your model f so i would not call this type of attack without access to the target model 3 how does this attack compares to any other black box attack that uses target model it really should be in table 2 if other attacks are able to make target model performance worse than this type of attack it is of less value to defend from a weaker attack 4 algo 3 what are the adversarial perturbations you are talking about 5 i am not sure algorithm 2 is the best way of doing it why not to try any of domain generalization techniques eg train on all domains with an adversarial head tries to distinguish between domains or mmd or whatever may be this way you wont need learn2weight model at all since you are already learning domain invariant features only minor table 2 what are u boldening i would expect the bolden result to be per source model book and the worse performance you get so dvd attack gives the lowest after attack accuracy you are boldening baby which is the weakest domain on which your attack mode is trained for an attack algo 2 compute weights of f trained on tywtwt just assign 0s docsepin this paper the authors propose a learn2weight framework to defend against similardomain adversarial attacks experimental studies on amazon dataset are done to verify the proposed learn2 weight the paper is not easy to follow the presentation and organization should be further improved here are the detailed comments 1 adversarial attacks are widely used in various application domains eg computer vision ref1 and reinforcement learning ref2 it is necessary to discuss with these related works and highlight the difference and importance of adversarial attack methods on nlp tasks ref1 adversarial examples that fool both computer vision and timelimited humans ref2 minimalistic attacks how little it takes to fool deep reinforcement learning policies 2 the authors highlight domain adaptation theory several times please give a clear description on what it is 3 where is table 1 used in the main content 4 regarding definition 2 the following two points are unclear 1 is fsa the true label of a based on the figure 1 a only correctly classified source samples are used while the definition does not show this 2 why fawt fsa f is the target classifier are you generating the domaininvariant samples 5 the rationale of similar domain adversarial attack is confused it is more reasonable to use source data to help generate target adversarial samples x which confuse the classifier to deviate the label fx neq fx where x is the original target sample however the paper generates source adversarial samples which naturally may confuse the target classifier due to the domain divergence it is unclear why and how these source adversarial samples can contribute to the robustness of the target classifier 6 regarding the accuracy drops in table 2 it is highly possible caused by the data shift between different domains how to differentiate the importance of the data shift and adversarial in the accuracy drops 7 the technical part is not easy to follow the sections 51 to 53 are not linked well it is necessary to give more contents on the motivation and flow of these algorithms instead of just putting them in algorithm charts 8 why target data is used in algorithm 2 and also transfer loss optimization in the introduction target domain information is assumed to be unavailable moreover algorithm 2 is to reduce the domain divergence if i understand correctly i am quite curious how the proposed method differentiates from other transfer learning methods update thanks for the authors response after reading the response and the other reviewers comments i think the paper needs to be further improved and thus i will keep my score
### Summary: | the submission considers a new attack model for adversarial perturbation in a framework where the attacker has neither access to the trained model nor the data used for training the model the submission suggests adomain adaptation inspired attack learn a different model on a similar domain and generate the adversarial perturbations using that model the authors then also develop a defense for this type of attack and provide some empirical evaluations of the resulting losses on a few nlp benchmark datasets the paper refers to the literature on domain adaptation theory to motivate their suggested defense but this analysis remains on an intuitive rather than a formally rigorous level furthermore the empirical evaluation does not compare to a variety of attacks and the defense is only evaluated with respected to the selfsuggested attack this is a very minimal bar for a defense to meet the reviewers have criticized the submission for the rather minimal extend of empirical evaluation given that the submission also doesnt provide a sound theoretical analysis for the proposed attack and defense i agree with the reviewers that the submission does not provide sufficient novel insight for publication at iclr in contrast to some of the reviewers i do find it legitimate and maybe recommendable even to focus on one chosen application area such as nlp i dont see a requirement to also present experiments on image data or reinforcement learning applciations however i would recommend that the authors highlight more explicitly what general lessons a reader would learn from their study this could be done through a more extensive and systematic set of experiments or a through analysis in a well defined theoretical framework |
Below is given review of a research paper from cnoference journal. Please write a summary the review.
### Review:
this paper proposes a detailed analysis on pruning heuristics and its applications to early pruning it thoroughly analyzed magnitudebased pruning losspreservation based pruning and gradientnorm based pruning the paper demonstrated the results on cifar10 and cifar100 datasets its very timely research to guide the audience which heuristic is better my major concern is the novelty over existing pruning heuristics since the techniques have all been proposed before the other concern is the evaluation and the scale of the dataset given the results in table 2 different by less than a percent and cifar training is very noisy its hard to tell the difference just like the lottery ticket hypothesis works on cifar but does not work on imagenet different pruning heuristics needs to be verified on the large scale imagenet dataset in order to be convincing docsep summary this paper studies different families of pruning criteria and their impact on training dynamics especially early training stemming from the observations authors provide improvements to the 1st and 2nd order saliency methods pros authors provide simple and useful explanations to various pruning criteria that are based on the taylor approximation of the loss functions even the authors dont mention this in the contributions they propose some improved versions of existing criteria for example the updated taylor score with theta2gtheta or absolute valued grasp this is great and it might worth focusing on these criteria further providing further evidence on their usefulness currently they seem a bit arbitrary for example why not third power theta3gtheta or additive biasing of magnitude gthetactheta i recommend authors to run their versions in unstructured setting too cons authors choose to focus on structured pruning since resulting networks are dense and acceleration is straightforward however they miss an important work on structured pruning 1 this relatively wellknown work shows that pruned structured networks can be trained to full accuracy from scratch in other words their value lies on doing some kind of an architecture search over layer widths the motivation of the work needs to be revisited in the light of these results since we can retrain pruned networks from scratch it probably doesnt matter which neuron we choose and therefore which criteria is better unstructured pruning doesnt have this training from scratch issue and i recommend authors to at least include and maybe shift the focus to unstructured pruning but requires specially designed hardware han et al 2016a or software elsen et al 2020 while results in this paper are applicable in both settings our experimental evaluation focuses on structured pruning due to its higher relevance to practitioners all neural networks require special hardware if you want to accelerate them i think a better motivation here is to point out to the difficulties at accelerating sparse operations and limited availabilitysupport for such operations in existing frameworks and i am not sure how useful structured pruning algorithms are given the results of 1 the larger the magnitude of parameters at a particular instant the smaller the model loss at that instant will be this is likely to be true in simple settings however it is not a sufficient condition especially for networks with batch norm you can arbitrarily scale neurons if there is a batchnorm and you can comeup with arbitrary ordering if needed i recommend rephrasing this observation andor stating the assumptions better i dont remember seeing any assumption on the network itself how the regularization or gradient noise will effect this statement thus the parameter with the most negative value for tgt is likely to also have a large negative value for thtgt this is not clear to me assume 1d case where t 1 gt2 ht1 tgt2 thtgt2 i can see the correlation in the figure but it doesnt seem like an obvious thing maybe because hessian dont have many negative eigenvalues rating i found the results and analysis interesting however motivation needs to be updated the work would also benefit from including unstructured pruning experiments minor points recent works focus on pruning models at initialization frankle carbin 2019 lottery ticket paper prunes after training and show existence of some initializations that achieve good performance equations 67 fracdldt gtheta2 assuming gradient descent shouldnt be a learning rate than magnitudeagnostic techniques which methods are these as far as i see all methods use magnitude information in their formulas directly or indirectly in table1 i recommend authors to bold both scores if they lie within the std of each other so that we can identify which improvements are significant it would be nice to show how the temperature parameter is used for grasp 1 httpsarxivorgabs181005270 docsepsummary the authors study proposed importance metrics for pruning neuronschannels in deep neural networks and analyze what properties of parameters are favored by each approach by studying the relationship between model parameters gradients 2nd order derivatives and loss through this analysis they develop a rich understanding of the consequences of different pruning criteria and use their understanding to propose modifications to existing techniques that produce higher quality models across different settings pros the framework used by the authors is clear and easy to understand but also very general the authors mix of empirical results and theoretical analysis makes a convincing case for the accuracy of their observations the authors go beyond observation and analysis and use their insights to design new approaches to pruning that outperform existing techniques the paper is well written and well organized cons this paper has few limitations the main limitation is that all experiments were conducted on relatively small datasets cifar given that is has been shown that some techniques in model compression produce stateoftheart results on small tasks but fail on larger models and datasets 1 2 id encourage their authors to further validate their insights on a larger dataset ie imagenet comments i found that the authors waited a long time to explain the term gradient flow which was important in sections 13 but not fully detailed until the start of section 4 on page 1 the authors say in parenthesis that gradient flow is gradient descent with infinitesimal learning rate but i found this explanation was not clear the second sentence of section 4 the evolution over time of model parameters gradient and loss was much more clear id encourage the authors to potentially work some of these details earlier into the text references 1 httpsarxivorgabs190209574 2 httpsarxivorgabs200303033 docsepthe paper contributes to explaining why saliency measures used for pruning trained models may or may not also be effective for pruning untrained or minimally trained models by developing the relationship between those saliency measures and different forms of the norm of model parameters based on the evolution of model parameters via gradient flow basically derivatives wrt time this result leads to several interesting interpretations that could shed some light on ongoing efforts to understand recent methods of pruning earlyon eg pruning at initialization or after minimal training and potential extensions to existing saliency measures the idea of employing gradient flow is novel for its purpose and seems to be accurately executed the main concern is that there is a gap between the flow model and the actual optimization method used in this work sgd with momentum or more generally standard optimization methods for deep learning in this regard the claim of evolution dynamics seems a bit exaggerated and remains as theoretical experiments are strictly speaking not entirely valid to support it either minor related work is written as if pruning is only done via saliencybased methods eg pruning frameworks generally define importance measures without taking into account various others such as optimization based methods employing sparsity inducing penalty terms on a different but related note the writing becomes a bit loose when it comes to referencing existing methods it is worth correcting them and clarifying the scopefocus of this work further questions why do you study structured pruning only the provided reasons unstructured pruning requires specially designed hardwares or softwares or higher relevance to practitioners dont seem valid enough if the purpose really lies in analyzing can you provide any results for unstructured pruning can you provide evidence to support the claim grasp without large temperature chooses to prune earlier layers aggressively besides raghu et al 2017 based on tables 1 and 2 the proposed extension to losspreservation method works the best while the differences across different methods seem a bit marginal is my understanding correct
### Summary: | this paper proposes a broad framework for unifying various pruning approaches and performs detailed analyses to make recommendations about the settings in which various approaches may be most useful reviewers were generally excited by the framework and analyses but had some concerns regarding scale and the papers focus on structured pruning the authors included new experiments however which mostly addressed reviewer concerns overall i think is a strong paper which will likely be provide needed grounding for pruning frameworks and recommend acceptance |
Below is given review of a research paper from cnoference journal. Please write a summary the review.
### Review:
summary this paper aims to explain dropout from the lens of game theoretic interactions let x denote the input of a deep neural net dnn intuitively the interaction between two variables xi and xj quantifies how much the presenceabsence of the jth variable affects the contribution of the ith variable to the output of the dnn with the above definition in place the authors show theoretically and empirically that dropout reduces the interactions between input variables of dnns as this type of interactions turn out to be strongly correlated with overfitting the authors suggest that dropout alleviates overfitting by reducing interactions between input variables or activation units of dnns based on this understanding of dropout an alternative regularization technique is proposed which explicitly penalizes pairwise interactions between variables strengths 1 the paper is well written and clearly presented 2 although it is already well known or at least widely accepted in the community that dropout reduces dependencies among activation units in dnns the explanation of dropout from the perspective of game theoretic interactions is interesting and it is supported both theoretically as well as empirically weakness 1 hyperparameter settings eg optimizationrelated ones to reproduce the results are not provided it is not clear what dropout rate was used in the experiments is it 05 in all cases different dropout rates should be investigated before claiming the superiority of the proposed interaction loss regularization over dropout 2 experiments are curried out on only one task classification one type of data images and one family of dnns convolutional neural nets however the paper draws quite general conclusions regarding the understanding of dropout from the perspective of game theoretic interactions therefore considering at least one more task involving a different type of data and another family of dnns would reinforce the findings of this paper 3 computational time analysis of the proposed interaction loss and training time comparisons with dropout are lacking additional comments 1 dropout is used both at convolutional and at fully connected layers however one can argue that applying dropout to convolutional layers does not make sense owing to the sparsity of connections in this type of layers 2 i would recommend revising the title of the paper what is proposed is more of an alternative regularization form to dropout than an improvement for the latter docseppaper summary the authors provide a novel interpretation of dropout regularization using banzhaf interactions a tool from game theory pros the authors are able to mathematically prove that dropout is capable of suppressing neural coadaptations the latter being one of the reasons for overfitting visualizations are also provided in this respect on a dataset for face analysis through their mathematical analysis authors are able to improve upon the classical dropout training by making it more compatible with batch normalization so that these two classical regularization strategies show a better complementarity cons some of the results does not read well like table 3 or figure 4 but this is really minor and fixable preliminary evaluation i believe that the overall analysis provided by authors is complete and interesting so i am akin to call for a full acceptance of the paper which i deem suitable for such a venue like iclr in order to improve their paper i would encourage authors to better investigate over the following aspect since many times authors established a principled connections between dropout and neural activations it would be very interesting to discuss the relationship with the present work and another paper gomez et al targeted dropout neurips workshops 2018 in which a computational variant of dropout is proposed such that the dropout rate depends upon neural activations postrebuttal evaluation i have carefully read the response provided by authors and checked the revised manuscript i confirm my preliminary acceptance ratedocsepsummary this paper analyzes the effect of dropout on interaction between units in a neural network the strength of the interaction is measured using a metric that is used in game theory to quantify interaction between players in a cooperative game the paper shows that dropout reduces highorder interaction as measured by this metric and that reduction in interaction is correlated with better generalization the paper introduces a new regularizer that explicitly minimizes the metric and claims that using this regularizer instead of dropout has some advantages pros the idea that dropout reduces overfitting by breaking up complex coadaptations and regularizing interactions is widely believed to be true however this paper tries to explicitly quantify the amount of interaction and presents theoretical and experimental evidence that interaction reduces as a result of having dropout cons the proposed metric is hard to compute exactly since it requires summing over exponentially many terms each term requiring a forward prop through the network the assumptions made in computing this metric approximately seem unclear to me appendix h i could not understand what probability distributions are being expressed and why in particular how is the term in eq 38 approximated by the one in the first line of eq 41 the paragraph after eq 40 was also unclear it is not discussed how this metric for evaluating interaction strength compares to something conceptually simpler like the hessian nabla2ij l which directly measures the dependence of the networks loss on pairs of input variables and its magnitude is proportional to the interaction strength the paper mentions that an advantage of the proposed loss is that the weight lambda applied to the interaction loss can be explicit controlled whereas the strength of dropout cannot be controlled section 4 advantages unlike the interaction loss people cannot explicitly control the strength of dropout this does not seem correct the dropout probability provides such as control mechanism for dropout for the experimental results in table 3 it is not mentioned what value of the dropout probability was used whether this value was tuned for each architecture and which network layers was dropout applied in these factors can have a significant impact on overall performance on the other hand the lambda parameter for the proposed interaction loss is tuned so the resulting comparison is not fair it is not clear what additional insight this metric provides about dropout beyond confirming what is intuitively apparent that having randomly dropped neurons will make it harder for the network to learn highorder interactions other comments and suggestions the introduction includes a discussion around banzhaf value without describing what it means the concept of banzhaf value might be new to many readers in the ml community i would suggest including a short explanation to give some intuition about what it means before discussing it in more detail the output of the dnn corresponds to the score f would it make sense to say that negative loss corresponds to the score f rather than the output of the network award reward or utility im not familiar with game theory literature so im not sure if award is a commonly used term there the title of the paper is a bit misleading as it seems to suggest that the paper is about using dropout in game theory ie solving problems in game theory using dropout post rebuttal the authors addressed the concerns around the clarity of the paper and added useful additional experiments i will increase my score to 7docsepsummary the paper proves that dropout can suppress the strength of interactions between input variables from the perspective of game theory it further improves the utility of dropout by introducing an explicit interaction loss experimental results verify the theoretic proof and the effectiveness of the proposed loss strengths 1 the paper introduces a new perspective of game theory to understand dropout 2 experiments are conducted on various datasets to support the theoretic proof and the proposed interaction loss concerns 1 although i have no background in game theory i try my best to understand the terminology and the analysis however i do not have the ability to verify the correctness of its proof thus i cannot evaluate the main contribution of this paper for experimental results the conclusion that dropout suppressing the input interactions is not a new story 2 it would be more interesting if the author can further explain the disharmony between dropout and bn from the perspective of game theory
### Summary: | the paper introduces a gametheoretic framework to improve our understanding of dropout all reviewers appreciated the contribution of the paper while they had a number of questionssuggestions almost all of them were adequately addressed three reviewers are satisfied and recommend acceptance while a lone reviewer is on the fence heshe admits heshe is less knowledgeable about game theory overall i think this paper makes a solid contribution to iclr |
Below is given review of a research paper from cnoference journal. Please write a summary the review.
### Review:
this paper first thoroughly analyzes the difference in distributions of weights and activations in addernet and then proposes a new quantization algorithm by redistributing the weights and the activations strengths 1 this paper conducts a thorough study of the dilemma in addernet quantization and proposes an effective method to solve this problem 2 the paper is clearly presented 3 i am glad to see that the accuracy drop is within 2 for imagenet even at 4 bits weaknesses suggestions 1 the accuracy and energy comparisons with quantized cnn seem to be not very adequate better to compare the accuracy drops in quantized cnns as well as the currently presented accuracy drop in addernet so the reader can get the full information whether addernet is more advanced as compared to cnns in terms of quantization 2 addernet is a specific neural network it is not clear whether the proposed methods can be generalized to other neural networks will similar distribution properties 3 only classification results are shown how about other downstrem tasks eg detection segemetation or other network architectures eg vits with adder kernels 4 does the addernet compatiable with ssl pretraining eg mae pretraining and how the quantization scheme different for pretraining stage and finetuning or normal training stages 5 how is the latency or throughput in real devices i am curious of this since there is no available efficient cuda implementation of addernet opensourced as of now also quantization is also not efficient for general cpugpu devices have you deployed the trained models to real devices yet i am open to further boost the score or champion this paper if the rebuttal is also sound and can somewhat solve my quesionsconcerns the comparison with cnn quantization method seems to be not very adequate docsepthe paper proposes a new quantization scheme for addernets specifically the authors propose to cluster model weights in the channel dimension where each cluster is assigned its own scale value for quantization this ensures that the scales can better represent the range of values which may be different for different weight channels the authors further propose to absorb the error caused by clamping the weights inside the layer bias which helps restore accuracy finally the proposed method removes outliers when quantizing the activations therefore tailoring the scale to better represent the valid range of data strengths the paper is wellwritten and the ideas are clearly explained the proposed method significantly improves the accuracy after quantization compared to prior methods weaknesses some of the claims are not backed up by the method specifically the authors mention that a shortcoming of prior work is using the same scale for weights and activations which is decided based on either of the two and therefore may not best fit the other the proposed method also adopts the same scheme where the scales are still determined by either the weights or the activations with the only difference being the increased granularity of the scale choices due to the channel clustering please find more details on this in the next section some questions remain regarding applying the method to new models eg how to determine the number of clusters for new benchmarks the authors have not discussed the limitations or potential negative social impact of their work docsepthis manuscript focuses on the problem of the quantization of addernet the author has investigated the difference between addernet and traditional networks based on the differences the dedicated quantization is achieved by redistributing the weights and activation of addernet in the quantization method three techniques are proposed to overcome the bit waste and over clamp problems including clusteringbased grouping quantization range clamp of weights and outlier clamp of activations experimental results show the effectiveness of the proposed method for quantizing addernet with different bit widths pros the manuscript is easy to follow the analysis of the difference between conventional quantization methods for cnn and that for addernet is interesting the statistics of the activations and weights of a pretrained addernet are good as a new kind of efficient neural network addernet how to effectively quantize it is a challenging problem quantization of such nn would put forward a faster and more energyefficient model the proposed method includes clusteringbased grouping quantization range clamp of weights and outlier clamp of activations is promising for addressing the bit waste or overclamp issues within addernet which is also verified by the extension experiments cons besides the flops and energy it would be great to report the inference time of the proposed method it is highly recommended to add a more detailed recap about addernet which will make the whole manuscript smoother especially for those who do not familiar with it in fig 3 how is the 1262 calculated there is no detailed explanation about it in line 45 l1norm quantization is unclear does it mean an l1normbased quantization method or quantization for l1norm operation minor issues there is a strange rec symbol in line 193 there are several minor grammar issues for example in line 21 wellverified should be well verified yes docsepquantization is an effective method to further reduce the energy consumption of addernets however previous addernets quantization methods cannot properly handle the challenge of large differences in weights and activations and often lead to a large accuracy degradation especially in the case of lowbit 4bit this paper first reveals the key reasons for the poor accuracy of previous addernets quantization methods namely over clamp and bits waste then a novel quantization method for addernets is proposed experiments on several datasets and models demonstrate the effectiveness of the proposed method strengths 1 the paper is extremely well structured and easy to follow with motivation wellexplained 2 to my knowledge this paper is by far the most comprehensive and systematic study of the quantization of addernets through thorough analysis this paper concludes two main reasons for the poor accuracy of previous addernets quantization methods namely over clamp and bits waste which are insightful the proposed scheme of the clusteringbased weights grouping and the lossless range clamp for weights are interesting and novel 3 extensive experiments on different models and datasets superior performance compared to other addernets quantization methods the thorough ablation studies verifies the effectiveness of each components the distributions of weights and activations fig1 in appendix demonstrate that the proposed method can effectively solve the problem of over clamp and bits waste leading to a higher quantized performance weaknesses 1 the values in fig 4 are too small to read the authors are required to refine them 2 the histogram for int4 weights adjacent to over clamp is significantly higher fig1 in appendix however this phenomenon is not expressed in the top of fig1 c the authors are advised to revise this detail for better presentation the authors have discussed the limitations and potential negative societal impact of their work in appendix
### Summary: | the reviewers were mostly positive about this paper 8664 while the negative reviewer did not update the review or respond after the authors response i do not see any major issues remaining the suggested method seems interesting novel and achieves good empirical results |
End of preview. Expand
in Dataset Viewer.
README.md exists but content is empty.
Use the Edit dataset card button to edit it.
- Downloads last month
- 32