Accelerated training of conditional random fields with stochastic gradient methods

Date

Authors

Vishwanathan, S
Schraudolph, Nicol
Schmidt, Mark W.
Murphy, Keven P.

Journal Title

Journal ISSN

Volume Title

Publisher

Association for Computing Machinery Inc (ACM)

Abstract

We apply Stochastic Meta-Descent (SMD), a stochastic gradient optimization method with gain vector adaptation, to the training of Conditional Random Fields (CRFs). On several large data sets, the resulting optimizer converges to the same quality of solution over an order of magnitude faster than limited-memory BFGS, the leading method reported to date. We report results for both exact and inexact inference techniques.

Description

Citation

Source

Proceedings of 23rd International Conference of Machine Learning

Book Title

Entity type

Access Statement

License Rights

DOI

Restricted until

2037-12-31