Notes On Naacl 2021 Oral Presentation On Semantic Parsing
Here are the brief notes that I took from listening to the oral presentations of great authors (mainly focused on text-to-SQL)
Latent Compositional Representations improve systematic generalization in grounded answering
They make trees on all latent
Tree helps whether the model is really correct or not.
Beam search can help
Learning from executions for semantic parsing
Executable program <- model that only uses labeled data
Lower bound that uses the labeled examples
Grammar-based <- forcing at training time, more correct that you do that in test time.
Assume that topk is not always executable
Pretrained parser already has some knowledge about what is right and what is wrong -> have to somehow renormalize the topk items
Trying to utilize the unlabeled data. How are we going to do learning from that?
not all the programs are not executable. Assign non-executable programs have very small loss.
Executability is a free signal. It is weak because we don’t know the answer.
We maximize marginal likelihood for all possible executable programs -> but all is infeasible. We utilize the beam searched query. But it seems kind of unfair. Encourage model to explore more executable models during program(motivation)
provide sparsity on the executable programs <- They can be executable but not correct(second motiviation)
y is our program, 0 otherwise. We want p (distribution of our parser) to be as similar as Q. Executable programs are always in this family. There is only 6 executable programs. It is discrete but the domain for executable program is continuous(?)
Distance metric is KL divergence. As long as p is close to Qx, we are happy. It could be anywhere to q. x
Grammar-based decoding <- Limit search space to executable programs.
We already do that. Grammatical != Executable difference? <- column type constraints?
Compositional generalization for neural semantic parsing via span-level supervised attention
-> span-level supervised attention
encourage models with span-level alignments? Semantic parser is seq-to-seq example.
contains syntac-level MT limitations
uses the outputs from this syntactic level alignments -> find bounding box
Incorporating external knowledge to enhance tabular reasoning
how do you represent tabular data to text? Can we incoporate external knowledge to model?
- poor representation of tabular information
- Implicit lexical knowledge
- presensce of distracting
- missing domain knowledge about keys
We use entity types for better table representation. Prune table according to it.
Preprocessing helps on alpha1 and alpha3.
Learning to synthesize data for semantic parsing
similar to back translation
Hand-crafting rules are not very good objective -> you don’t know what the dataset will look like.
Utterance doesn’t make sense, but maybe the model can learn from the correspondance.
Leave a comment