ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2107.03277
13
6

Linear-time calculation of the expected sum of edge lengths in random projective linearizations of trees

7 July 2021
Lluís Alemany-Puig
R. Ferrer-i-Cancho
ArXivPDFHTML
Abstract

The syntactic structure of a sentence is often represented using syntactic dependency trees. The sum of the distances between syntactically related words has been in the limelight for the past decades. Research on dependency distances led to the formulation of the principle of dependency distance minimization whereby words in sentences are ordered so as to minimize that sum. Numerous random baselines have been defined to carry out related quantitative studies on languages. The simplest random baseline is the expected value of the sum in unconstrained random permutations of the words in the sentence, namely when all the shufflings of the words of a sentence are allowed and equally likely. Here we focus on a popular baseline: random projective permutations of the words of the sentence, that is, permutations where the syntactic dependency structure is projective, a formal constraint that sentences satisfy often in languages. Thus far, the expectation of the sum of dependency distances in random projective shufflings of a sentence has been estimated approximately with a Monte Carlo procedure whose cost is of the order of RnRnRn, where nnn is the number of words of the sentence and RRR is the number of samples; it is well known that the larger RRR, the lower the error of the estimation but the larger the time cost. Here we present formulae to compute that expectation without error in time of the order of nnn. Furthermore, we show that star trees maximize it, and give an algorithm to retrieve the trees that minimize it.

View on arXiv
Comments on this paper