Presentation on theme: "Authors Sebastian Riedel and James Clarke Paper review by Anusha Buchireddygari Incremental Integer Linear Programming for Non-projective Dependency Parsing."— Presentation transcript:
Authors Sebastian Riedel and James Clarke Paper review by Anusha Buchireddygari Incremental Integer Linear Programming for Non-projective Dependency Parsing
Cutting plane Algorithm Structured Prediction Figure above shows a convex region of feasible solutions defined by several constraints. The grid indicates where inside the polygon the feasible integer solutions lie The dot represents the optimal solution (for the linear programming problem) gained from maximizing x 1 + x 2 Note that although it is not an integer solution it is the upper bound for an optimum one
Why this paper? Integer Linear Programming (ILP) was applied for inference of sequential conditional random fields(Roth and Yin,2004). Exponential number of constraints required to prevent cycles occurring in the dependency graph. Model these constraints using ILP is too large to solve efficiently.
What does the paper solve? Method which extends applicability of ILP to more complex set of problems. Instead of adding all the constraints authors tried to solve with a fraction of constraints. Solution is examined and additional constraints are added if required. Procedure is repeated until all the constraints are satisfied.
What’s the problem the authors picked? Authors applied dependency parsing approach to Dutch due to language’s non-projective nature. Took the parser of McDonald et al as starting point for our model. “I will come at twelve and then you will get what you deserve”.
What’s Dependency Parsing? Task of attaching words to arguments. Dependency graph “kom” attached to “ik”. “kom” is head for “ik” and “ik” is child. Dependency tree every token must be the child of exactly one other node, either another token or dummy root token. It cannot have cycles like “en”-> “kom” -> “ik” - >”en”.
How the model looks like? X is a sentence, y is a set of labelled dependencies F(I,j,l) multi dimensional feature vector of the edge from token I to token j with label l
Constraints Decoding/Inference in this model is to maximize T1 for every non-root token in x there exists exactly one head; the root token has no head T2 There are no cycles This corresponds to maximum spanning tree problem.
Linguistic constraints A1 Heads are not allowed to have more than one outgoing edge labelled l for all l in a set of labels U. A1 tells that there can be only one “subject” in the sentence C1 In a symmetric coordination there is exactly one argument to the right of the conjunction and at least one argument on the left. C1 applies to “and”,”or”,”but”
Some more linguistic constraints ● C2 In an asymmetric coordination there are no arguments to the left of the conjunction and at least two arguments to the right. ● Example is “both” having arguments to its left. ● There are other such constraints total of 8 constraints in the paper.
Process ● The function to be maximized is called objective function Ox. ● Variables Vx, e(I,j,l) is 1 if edge exists else 0 ● Base constraints ● T1
Constraints representation A1 C1
Incremental Constraints T2
Algorithm For a sentence x, Bx is the set of constraints that we add in advance. Ix are the constraints we add in incrementally Ox is objective function and Vx is a set of variables including variable declaration.
What happens in the algorithm? Solve(C,O,V) maximizes the objective function O with respect to the set of constraints C and variables V. Violated(y,I) inspects the proposed solution (y) and returns all constraints in I which are violated
Interesting result The number of iterations is most polynomial with respect to number of variables. In practice this technique converges quickly i.e. less than 20 iterations in 99% approximately 12,000 sentences. Yielding average solve times of less than 0.5 seconds.
Experiment How much do our additional constraints help improve accuracy? How fast is our generic inference method in comparison with Chu-Liu- Edmonds algorithm? Can approximations be used to increase the speed of our method while remaining accurate? Data Alpino Treebank 13,300 sentences with a average length of 14.6 tokens Environment Intel Xeon with 3.8 Ghz and 4 GB RAM. Mixed Integer Programming lib IP_Solve. Code ran on Java and called JNI-wrapper aroung IP_solve lib Feature Sets Along with POS tags there are additional attributes like gender,number and case. Combined attr of head to child
Results Accuracy Nl is the number of tokens with correct head and label and Nt is the total number of tokens. Unlabelled accuracy Nu is the number of tokens with correct head
Accuracy Results table bl means baseline without any linguistic constraints which is compared it to a system with additional constraints (cnstr) There are problems the system suffers from poor next best solution due to inaccurate local score distributions.
Runtime Efficiency Average solve time (ST) for sentences with respect to no of tokens in each sentence. Approximation Total runtime of the system is competitive to CLE.
Discussion Higher order features by using extended set of variables and a modified function which is likely to increase runtime. Fast for real world applications. Large time for first iteration after that algorithm uses last state to efficiently search for solutions in the presence od new constraints. Exponential blow when the cycle constraints are added
Conclusion Novel approach to inference of ILP Efficiently use ILP for dependency parsing Slower than baseline approach but parses large sentences with more than 50 tokens. Parsing time can be significantly reduced using a simple approximation like from q=10 to q=5 which will only marginally degrades the performance from 85% to 84% in the table we have seen in approximation.