欢迎来到麦多课文档分享! | 帮助中心 海量文档,免费浏览,给你所需,享你所想!
麦多课文档分享
全部分类
  • 标准规范>
  • 教学课件>
  • 考试资料>
  • 办公文档>
  • 学术论文>
  • 行业资料>
  • 易语言源码>
  • ImageVerifierCode 换一换
    首页 麦多课文档分享 > 资源分类 > PPT文档下载
    分享到微信 分享到微博 分享到QQ空间

    Applying the Hidden Markov Model Methodology for .ppt

    • 资源ID:378518       资源大小:410.50KB        全文页数:44页
    • 资源格式: PPT        下载积分:2000积分
    快捷下载 游客一键下载
    账号登录下载
    微信登录下载
    二维码
    微信扫一扫登录
    下载资源需要2000积分(如需开发票,请勿充值!)
    邮箱/手机:
    温馨提示:
    如需开发票,请勿充值!快捷下载时,用户名和密码都是您填写的邮箱或者手机号,方便查询和重复下载(系统自动生成)。
    如需开发票,请勿充值!如填写123,账号就是123,密码也是123。
    支付方式: 支付宝扫码支付    微信扫码支付   
    验证码:   换一换

    加入VIP,交流精品资源
     
    账号:
    密码:
    验证码:   换一换
      忘记密码?
        
    友情提示
    2、PDF文件下载后,可能会被浏览器默认打开,此种情况可以点击浏览器菜单,保存网页到桌面,就可以正常下载了。
    3、本站不支持迅雷下载,请使用电脑自带的IE浏览器,或者360浏览器、谷歌浏览器下载即可。
    4、本站资源下载后的文档和图纸-无水印,预览文档经过压缩,下载后原文更清晰。
    5、试题试卷类文档,如果标题没有明确说明有答案则都视为没有答案,请知晓。

    Applying the Hidden Markov Model Methodology for .ppt

    1、AIDA01-1,Applying the Hidden Markov Model Methodology for Unsupervised Learning of Temporal Data,Cen Li1 and Gautam Biswas21Department of Computer Science Middle Tennessee State University2Department of EECS Vanderbilt University Nashville, TN 37235. USA. biswasvuse.vanderbilt.edu biswashttp:/www.vu

    2、se.vanderbilt.edu/biswas June 21, 2001,AIDA01-2,Problem Description,Most real world systems are dynamic, e.g., Physical plants and Engineering systems Human Physiology Economic systems Systems are complex: hard to understand, model, and analyze But our ability to collect data on these systems has in

    3、creased tremendouslyTask: Use data to automatically build models, extend incomplete models, and verify and validate existing models using the data available.Why models?Formal, abstract representation of phenomena or processEnables systematic analysis and prediction Our goal:Build models, i.e., creat

    4、e structure from data using exploratory techniquesChallenge:Systematic and useful clustering algorithms for temporal data,AIDA01-3,Outline of Talk,Example Problem Related work on temporal data clustering Motivation for using Hidden Markov Model (HMM) representation Bayesian HMM clustering methodolog

    5、y Experimental results Synthetic Data Real world ecological data Conclusions and future work,AIDA01-4,Example of Dynamic System,FiO2,RR,MAP,PaO2,Temporal features: FiO2, RR, MAP, PaO2, PIP, PEEP, PaCO2, TV .,ARDS patients response to respiratory therapy:,Goal : To construct profiles of patient respo

    6、nse patterns for better understanding the effects of various therapies,AIDA01-5,Problem Description,Unsupervised Classification (Clustering)Given data objects described by multiple temporal features:(1) to assign category information to individual objects by objectively partitioning the objects into

    7、 homogeneous groups such that the within group object similarity and the between group object dissimilarity are maximized,(2) to form succinct description for each category derived.,AIDA01-6,Related Work on Temporal Data Clustering,AIDA01-7,Motivation for using HMM Representation,What are our modeli

    8、ng objectives ? Why do we choose the HMM representation ? The hidden states valid stages of a dynamic process Direct probabilistic links probabilistic transitions among different stages,Mathematically HMM Model made up of three components : initial state probabilities A: transition matrix B: emissio

    9、n matrix ,AIDA01-8,Continuous Speech Signal,Rabiner and Huang, 1993,Spectral sequence feature extraction statistical pattern recognitionIsolated Word Recognition 1 HMM per word 2-10 states per HMM direct continuous density HMMs produce best resultsContinuous Word Recognition Number of words (?)Utter

    10、ance boundaries (?) Word boundaries: fuzzy/non unique Matching word reference patterns to Words: exponential problem,Whole word models typically not used for continuous speech recognition,AIDA01-9,Continuous Speech Recognizer using sub-words,Recognized Sentence,Sentence (Sw):,W1,W2,silence,silence,W

    11、ord (W1):,U1(W1),U2(W1),UL(W1)(W1),Sub-word Unit (PLU):,Rabiner and Huang, 1993,AIDA01-10,Continuous Speech Recognition Characteristics,Lot of background knowledge available, e.g., Phonemes & syllables for isolated word recognition Sub-word units (Phonelike units (PLUs), Syllable-like units, Acousti

    12、c units) for word segmentation Grammar and semantics for language models As a result, HMM structure usually well-defined Recognition task dependent on learning model parameters Clustering techniques employed to derive efficient state representation Single speaker versus speaker independent recogniti

    13、on Question: What about domains where structure not well known & Data not as well-defined,AIDA01-11,Past Work on HMM Clustering,Past Work: Rabiner et al., 1989 HMM Likelihood Clustering, HMM Threshold Clustering Baum Welch, Viterbi methods for parameter estimation Lee, 1990 Agglomerative procedure B

    14、ahl et al., 1986, Normandin et al., 1994 HMM parameter estimation by Maximal Mutual Information (MMI) Kosaka et al. 1995 HMnet composition and clustering (CCL); Bhattacharyya distance measure Dermatas & Kokkinakis, 1996 Extended Rabiner et al.s work more sophisticated clustering structure clustering

    15、 by recognition error Smyth, 1997 Clustering with finite mixture models Limitations: No objective criterion for cluster partition evaluation and selection Apply uniform, pre-defined HMM model size for clustering,Applied to Speech Recognition,AIDA01-12,Original work on HMM parameter estimation (Rabin

    16、er, et al., 1989),Maximum Likelihood Method (Baum Welch): Parameter estimation by the E(xpectation)-M(aximization) method Assign initial parameters E-step: compute expected values of the necessary statistics M-step: update model parameter values to maximize likelihood Use forward-backward procedure

    17、to cut down on the computational requirements. Baum Welch method very sensitive to initialization: use Viterbi method to find most likely path, and use that to initialize parameters. Extensions to method (estimate model size): state splitting, state merging, etc.,AIDA01-13,AIDA01-14,Our Approach to

    18、HMM Clustering,Four nested search loops:Loop 1: search for the optimal number of clusters in the partition, Loop 2: search for the optimal object distribution to the clusters in the partition, Loop 3: search for the optimal HMM model size for each cluster in the partition, and Loop 4: search for the

    19、 optimal model parameter configuration for each model .Search space:,HMM Learning,AIDA01-15,Introducing Heuristics to HMM Clustering,Four nested levels of search:Loop 1: search for the optimal number of clusters in the partition,Loop 2: search for the optimal object distribution to clusters in the p

    20、artition,Loop 3: search for the optimal HMM model size for each cluster in the partition, andLoop 4: search for the optimal model parameter configuration for each model .,Bayesian Model Selection,HMM Learning,AIDA01-16,Bayesian Model Selection,(Marginal Likelihood),Model Posterior Probability:,Goal:

    21、 Select Model M that maximizes P(M|D),AIDA01-17,Computing Marginal Likelihhods,Issue of complete versus incomplete dataApproximation Techniques Monte Carlo methods (Gibbs sampling) Candidate Method (variation of Gibbs sampling) Chickering Laplace approximation (multivariate Gaussian, quadratic Taylo

    22、r series approx.)Bayesian Information Criteria derived from Laplace approx. Very efficientCheeseman-Stutz approx.,AIDA01-18,Marginal Likelihood Approximations,Bayesian Information Criterion (BIC):Cheeseman-Stutz (CS) Approximation:,Likelihood,Model ComplexityPenalty,Likelihood,Model PriorProbability

    23、,AIDA01-19,Introducing Heuristics to HMM Clustering,Four nested levels of search: Step 1: search for the optimal number of clusters in the partition,Step 2: search for the optimal object distribution to clusters in the partition,Step 3: search for the optimal HMM model size for each cluster in the p

    24、artition, andStep 4: search for the optimal model parameter configuration for each model .,Bayesian Model Selection,AIDA01-20,Bayesian Model Selection for HMM Model Size Selection,Goal: To select the optimal number of states, K, for a HMMthat maximizes,AIDA01-21,BIC and CS for HMM Model Size Selecti

    25、on,(b) CS Approximation,likelihood,(a) BIC approximation,CS,prior,penalty,likelihood,BIC,2 3 4 5 6 7 8 9 10,2 3 4 5 6 7 8 9 10,AIDA01-22,HMM Model Size Selection Process,Sequential search for HMM model size with a heuristic stopping criterion - Bayesian model selection criterion approximated by BIC

    26、and CS Issue 2: State initialization by K-means clustering,Assumptions:Data CompleteData Sufficient otherwise, suboptimal structure,AIDA01-23,Introducing Heuristics to HMM Clustering,Four nested levels of search: Step 1: search for the optimal number of clusters in the partition,Step 2: search for t

    27、he optimal object distribution to clusters in the partition,Step 3: search for the optimal HMM model size for each cluster in the partition, andStep 4: search for the optimal model parameter configuration for each model .,Bayesian Model Selection,AIDA01-24,Cluster Partition Selection : Bayesian Mode

    28、l Selection for Mixture of HMMs,Goal: Select partition with the optimal number of clusters, K,that maximizes P(M|D),Partition Posterior Probability (PPP),Apply BIC approximation,AIDA01-25,BIC and CS for Cluster Partition Selection,(a) BIC approximation,(b) CS approximation,penalty,prior,likelihood,l

    29、ikelihood,BIC,CS,1 2 3 4 5 6 7 8 9 10,1 2 3 4 5 6 7 8 9 10,AIDA01-26,Bayesian Model Selection for Partition Evaluation and Selection,Sequential search for clustering partition model using a heuristic stopping criterion - Bayesian model selection criterion (approximated by BIC and CS),AIDA01-27,Fixed

    30、 Size Cluster Partition Formation,Goal : to maximize:,Hard Data Partitioning,BIC criteria,AIDA01-28,Introducing Heuristics to HMM Clustering,Four nested levels of search:Step 1: search for the optimal number of clusters in the partition,Step 2: search for the optimal object distribution to clusters

    31、in the partition,Step 3: search for the optimal HMM model size for each cluster in the partition, andStep 4: search for the optimal model parameter configuration for each model .,Bayesian Model Selection,EM with hard data partitioning & heuristicmodel initialization,Clustering HMM initialization,AID

    32、A01-29,Bayesian HMM Clustering (BHMMC) with Component Model Size Selection,Complexity O(KNMmaxL),AIDA01-30,Experiments,Exp 1: Study the HMM model size selection methodExp 2: Compare HMM clustering with varying HMM sizes vs. HMM clustering with fixed HMMExp 3: Clustering robustness to data skewnessEx

    33、p 4: Clustering sensitivity to seed selection,AIDA01-31,Experiment 1: Modeling with HMM,2,3,4,5,6,7,8,9,10,1,1,2,3,4,5,6,7,8,9,10,11,12,13,14,X,Y,y=kx+b, k= -1/k,S,1,3,S,2,S,S,4,y=kx+b,S,m,3,m,2,1,1,2,4,m,d,d(S , S )=g( ),d(S , S )=g( ),2,d(S , S ) = f( ),d,d,AIDA01-32,Experiment 1: Modeling with HM

    34、M (continued),d,HMM model sizes rediscovered given 4-state HMMs of different d1 and d2 values: mean,d,d,=0.5,HMM Size Rediscovered,2,3,2,1,0.5,0.1,d,= 1,= 3,= 2,d,1,1,1,1,AIDA01-33,Experiment 2: Clustering with varying vs. fixed HMM model size,Partition Misclassification Count: mean (standard deviat

    35、ion),AIDA01-34,Experiment 2 (continued),Between Partition Similarity,Varying,Fixed 3,Fixed 8,Fixed 15,AIDA01-35,Experiment 2 (continued),Varying,Fixed 3,Fixed 8,Fixed 15,Partition Posterior Probability,AIDA01-36,Experiment 3: BHMMC with regular vs. skewed data,Average Misclassification Count,Cluster

    36、 1 - Cluster 2 %,Model distance = level 1,Model distance = level 2,AIDA01-37,Experiment 4: Clustering Sensitivity to Seed Selection,2,l,l,l,l,l,l,l,l,l,1,1,1,2,2,3,3,3,Data B :,l,l,l,l,l,1,1,1,2,3,l,2,l,3,l,3,2,l,Data A :,AIDA01-38,Experiment 4 (continued),Data B,Data B,Data A,Data A,1,2,3,1,2,3,Ave

    37、rage over 10 runs,Variance on PPP(mean),Variance on misclassificationcount (mean),Level,Level,AIDA01-39,Experiments with Ecological Data collaborators: Dr. Mike Dale and Dr. Pat Dale School of Environmental Studies, Griffith University, Australia,Study the ecological effects of mosquito control by r

    38、unneling, i.e., changing the drainage patterns in an area Data collected from about 30 sites in a salt marsh area south of Brisbane, Australia Data: At each site we have four measures of species performance Height and density of the grass Sporobolous (marine couch) Height and density of succulent Sa

    39、rcocornia (samphire) 11 other environmental variables measured (they have missing values), but they were not used for cluster generation. Data collected in 3 month intervals starting mid 1985, each temporal data sequence is 56 long,AIDA01-40,Results of Experiments (1),Experiment 1: Generate HMM Clus

    40、ters with all four plant variables. Result: Single dynamic process (one cluster partition) with 4 states that form a cycle.,Sporobolous Dense & Tall Sarcocornia Sparse,Sporobolous Sparse & Moderate Sarcocornia Short & Sparse,Sporobolous Sparse Sarcocornia Sparse & Tall,Sporobolous Almost absent Sarc

    41、ocornia Dense & Moderate,Conclusions: Mosquito control treatment has no significant effects on plant ecology Some parts of the marsh shifted got wetter, and shifted more toward Sarcocornia,AIDA01-41,Results of Experiments (2),Experiment 2: Isolate and study the effects on the Sporobolous plantResult

    42、: Two cluster partition: Cluster 1: 5 states with low transition probabilities (more or less static)Cluster 2: Two interacting dynamic cycles,S1 Sporobolous Dense & Tall,S5 Sporobolous Short & Dense,S2 Sporobolous Medium & Dense,S3 Sporobolous Sparse & Medium,S4 Sporobolous Almost Bare,0.9,0.9,0.45,

    43、0.33,0.5,0.33,0.33,0.9,0.1,Loop 1,Loop 2,AIDA01-42,Results of Experiments (3),Summary of Results (Experiment 2) Two paths for getting Sporobolous to bare ground In loop 1 the grass recovers through state 2 In loop 2 the process may stay in state 3 (medium height) with equal chance of going bare (sta

    44、te 4) or becoming short and dense (state 5). Shows modifications (new states) to the overall process Other observations Usually ecological systems have more memory than 3 months. May require higher order Markov processes Feature values may not be independent (consequences?) There was a drought in th

    45、e middle of the observation period Amount of data available for parameter estimation limited may affect statistical validity of estimated parameters How do we incorporate the environmental variables into model? Some are discrete. They are at a different temporal granularity. Missing values.,AIDA01-4

    46、3,Conclusions,Applied the Bayesian clustering methodology to work with mixture of CDHMMs Solved the HMM model size selection problem in the framework of Bayesian model selection Improved the quality of cluster partition and individual cluster models generated by introducing the HMM model size select

    47、ion procedure into the clustering process Proposed a more effective HMM parameter initialization method Identified problems that may be common to other clustering tasks involving graphical models of undetermined sizes,AIDA01-44,Future Work,Improve clustering stability Evaluate partition quality in t

    48、erms of the predictiveness of the cluster models Investigate incremental Bayesian HMM clustering approaches Enhance HMM of dynamic processes (higher order models, non stationary processes) Apply this methodology to real data, perform knowledge based model interpretation and analysis Ecological data Stock data Therapy response data,


    注意事项

    本文(Applying the Hidden Markov Model Methodology for .ppt)为本站会员(orderah291)主动上传,麦多课文档分享仅提供信息存储空间,仅对用户上传内容的表现方式做保护处理,对上载内容本身不做任何修改或编辑。 若此文所含内容侵犯了您的版权或隐私,请立即通知麦多课文档分享(点击联系客服),我们立即给予删除!




    关于我们 - 网站声明 - 网站地图 - 资源地图 - 友情链接 - 网站客服 - 联系我们

    copyright@ 2008-2019 麦多课文库(www.mydoc123.com)网站版权所有
    备案/许可证编号:苏ICP备17064731号-1 

    收起
    展开