Given a matrix , a linear feasibility problem (of which linear classification is a special case) aims to find a solution to a primal problem or a certificate for the dual problem which is a probability distribution . Inspired by the continued importance of large margin classifiers in machine learning, this paper aims to deepen our understanding of a condition measure of called \textit{margin} that determines the difficulty of both the above problems. To aid geometrical intuition, we establish new characterizations of the margin in terms of relevant balls, cones and hulls. Our main contribution is analytical, where we present generalizations of Gordan's theorem, and beautiful variants of Hoffman's theorems, both using margins. We end with shedding some new light on two classical iterative schemes, the Perceptron and Von-Neumann or Gilbert algorithms, whose convergence rates famously depend on the margin. Our results are relevant for a deeper understanding of margin-based learning and proving convergence rates of iterative schemes, apart from providing a unifying perspective on this vast topic.
View on arXiv