Abstract | ||
---|---|---|
Hidden Markov Models (HMM) have been successfully used in applications such as speech recognition, activity recognition, bioinformatics etc. There have been previous attempts such as Hierarchical HMMs and Abstract HMMs to elegantly extend HMMs at multiple levels of temporal abstraction (for example to represent the user's activities). Similarly, there has been previous work such as Logical HMMs on extending HMMs to domains with relational structure. In this work we develop a representation that naturally combines the power of both relational and hierarchical models in the form of Logical Hierarchical Hidden Markov Models (LoHiHMMs). LoHiHMMs inherit the compactness of representation from Logical HMMs and the tractability of inference from Hierarchical HMMs. We outline two inference algorithms: one based on grounding the LoHiHMM to a propositional HMM and the other based on particle filtering adapted for this setting. We present the results of our experiments with the model in two simulated domains. |
Year | DOI | Venue |
---|---|---|
2008 | 10.1007/978-3-540-85928-4_17 | ILP |
Keywords | Field | DocType |
hidden markov models,logical hmms,inference algorithm,logical hierarchical hidden markov,propositional hmm,previous work,previous attempt,activity recognition,abstract hmms,modeling user activities,hierarchical hmms,hidden markov model,particle filter,hierarchical model,speech recognition | Abstraction,Activity recognition,Markov model,Computer science,Inference,Particle filter,Relational structure,Theoretical computer science,Artificial intelligence,Hidden Markov model,Machine learning | Conference |
Volume | ISSN | Citations |
5194 | 0302-9743 | 16 |
PageRank | References | Authors |
0.86 | 11 | 5 |
Name | Order | Citations | PageRank |
---|---|---|---|
Sriraam Natarajan | 1 | 482 | 49.32 |
Hung Hai Bui | 2 | 1188 | 112.37 |
Prasad Tadepalli | 3 | 1182 | 152.65 |
Kristian Kersting | 4 | 1932 | 154.03 |
Weng-Keen Wong | 5 | 817 | 59.67 |