Hidden 1-Counter Markov Models and How to Learn Them

Hidden 1-Counter Markov Models and How to Learn Them

Mehmet Kurucan, Mete Özbaltan, Sven Schewe, Dominik Wojtczak

Proceedings of the Thirty-First International Joint Conference on Artificial Intelligence
Main Track. Pages 4857-4863. https://doi.org/10.24963/ijcai.2022/673

We introduce hidden 1-counter Markov models (H1MMs) as an attractive sweet spot between standard hidden Markov models (HMMs) and probabilistic context-free grammars (PCFGs). Both HMMs and PCFGs have a variety of applications, e.g., speech recognition, anomaly detection, and bioinformatics. PCFGs are more expressive than HMMs, e.g., they are more suited for studying protein folding or natural language processing. However, they suffer from slow parameter fitting, which is cubic in the observation sequence length. The same process for HMMs is just linear using the well-known forward-backward algorithm. We argue that by adding to each state of an HMM an integer counter, e.g., representing the number of clients waiting in a queue, brings its expressivity closer to PCFGs. At the same time, we show that parameter fitting for such a model is computationally inexpensive: it is bi-linear in the length of the observation sequence and the maximal counter value, which grows slower than the observation length. The resulting model of H1MMs allows us to combine the best of both worlds: more expressivity with faster parameter fitting.
Keywords:
Uncertainty in AI: Bayesian Networks
Agent-based and Multi-agent Systems: Formal Verification, Validation and Synthesis
Machine Learning: Bayesian Learning
Machine Learning: Time-series; Data Streams
Uncertainty in AI: Graphical Models