Edmonton, May-June 2003 | Fei Sha and Fernando Pereira
The paper by Fei Sha and Fernando Pereira from the University of Pennsylvania explores the use of conditional random fields (CRFs) for shallow parsing, a task in natural language processing that identifies non-recursive cores of various phrase types. CRFs combine the strengths of generative models and discriminative classifiers, offering advantages over both approaches. The authors demonstrate that CRFs can achieve performance comparable to or better than state-of-the-art noun-phrase chunking methods on the CoNLL task, a standard evaluation dataset for shallow parsing. They achieve this by using modern optimization algorithms, such as preconditioned conjugate gradient (CG) and limited-memory quasi-Newton (L-BFGS), which provide better convergence properties compared to iterative scaling methods. The paper also includes extensive comparisons between different training methods and discusses the impact of feature sets and parameter tuning on performance. The results highlight the effectiveness of CRFs in shallow parsing and suggest that they can be a competitive solution for learning shallow parsers, potentially replacing lexicalized PCFG models in future applications.The paper by Fei Sha and Fernando Pereira from the University of Pennsylvania explores the use of conditional random fields (CRFs) for shallow parsing, a task in natural language processing that identifies non-recursive cores of various phrase types. CRFs combine the strengths of generative models and discriminative classifiers, offering advantages over both approaches. The authors demonstrate that CRFs can achieve performance comparable to or better than state-of-the-art noun-phrase chunking methods on the CoNLL task, a standard evaluation dataset for shallow parsing. They achieve this by using modern optimization algorithms, such as preconditioned conjugate gradient (CG) and limited-memory quasi-Newton (L-BFGS), which provide better convergence properties compared to iterative scaling methods. The paper also includes extensive comparisons between different training methods and discusses the impact of feature sets and parameter tuning on performance. The results highlight the effectiveness of CRFs in shallow parsing and suggest that they can be a competitive solution for learning shallow parsers, potentially replacing lexicalized PCFG models in future applications.