SlideShare ist ein Scribd-Unternehmen logo
1 von 49
Conjuntive Formulation of the Random Set Framework for Multiple Instance Learning:Application to Remote Sensing Jeremy Bolton Paul Gader CSI Laboratory  University of Florida
Highlights Conjunctive forms of Random Sets for Multiple Instance Learning: Random Sets can be used to solve MIL problem when multiple concepts are present Previously Developed Formulations assume Disjunctive relationship between concepts learned New formulation provides for a conjunctive relationship  between concepts and its utility is exhibited on a Ground Penetrating Radar (GPR) data set
Outline Multiple Instance Learning MI Problem RSF-MIL Multiple Target Concepts Experimental Results GPR Experiments  Future Work
Multiple Instance Learning
Standard Learning vs. Multiple Instance Learning Standard supervised learning Optimize some model (or learn a target concept) given training samples and corresponding labels MIL Learn a target concept given multiplesets of samples and corresponding labels for the sets. Interpretation: Learning with uncertain labels / noisy teacher
Multiple Instance Learning (MIL)  Given:  Set of I bags  Labeled + or - The ith bag is a set of Ji 	samples in some feature space Interpretation of labels Goal: learn concept What characteristic is common to the positive bags that is not observed in the negative bags
Multiple Instance Learning Traditional Classification Multiple Instance Learning {x1, x2, x3, x4} label = 1 {x1, x2, x3, x4} label = 1 {x1, x2, x3, x4} label = 0 x1 label = 1 x2label = 1 x3label = 0 x4label = 0 x5label = 1
MIL Application: Example GPR EHD: Feature Vector Collaboration: Frigui, Collins, Torrione Construction of bags Collect 15 EHD feature vectors from the 15 depth bins Mine images = + bags FA images = - bags
Standard vs. MI Learning: GPR Example Standard Learning Each training sample (feature vector) must have a label Arduous task  many feature vectors per image and multiple images difficult to label given GPR echoes, ground truthing errors, etc …  label of each vector may not be known EHD: Feature Vector
Standard vs MI Learning: GPR Example EHD: Feature Vector Multiple Instance Learning Each training bag must have a label ,[object Object]
Implicitly accounts for class label uncertainty …,[object Object]
Random Set Brief Random Set
How can we use Random Sets for MIL? It is NOT the case that EACH  element is NOT the  target concept Random set for MIL: Bags are sets (multi-sets) Idea of finding commonality of positive bags inherent in random set formulation Sets have an empty intersection or non-empty intersection relationship Find commonality using intersection operator Random sets governing functional is based on intersection operator Capacity functional : T A.K.A. :  Noisy-OR gate (Pearl 1988)
Random Set Functionals Capacity functionals for intersection calculation Use germ and grain model to model random set Multiple  (J) Concepts Calculate probability of intersection given X and germ and grain pairs: Grains are governed by random radii with assumed cumulative: Random Set model parameters Germ  Grain
RSF-MIL: Germ and Grain Model x T x T T x x x x T T x x x Positive Bags = blue Negative Bags = orange Distinct shapes = distinct bags
Multiple Instance Learning with Multiple Concepts
Multiple Concepts: Disjunction or Conjunction? Disjunction When you have multiple types of concepts When each instance can indicate the presence of a target Conjunction When you have a target type that is composed of multiple (necessary concepts) When each instance can indicate a concept, but not necessary the composite target type
Conjunctive RSF-MIL Previously Developed Disjunctive RSF-MIL (RSF-MIL-d) Conjunctive RSF-MIL (RSF-MIL-c) Noisy-OR combination across concepts and samples Standard noisy-OR for one concept j Noisy-AND combination across concepts
Synthetic Data Experiments Extreme Conjunct data set requires that a target bag exhibits two distinct concepts rather than one or none  AUC (AUC when initialized near solution)
Application to Remote Sensing
Disjunctive Target Concepts Target Concept  Type 1  NoisyOR Target Concept  Type 2  NoisyOR OR … Target Concept  Type n NoisyOR Target Concept Present? Using Large overlapping bins (GROSS Extraction) the target concept can be encapsulated within 1 instance: Therefore a disjunctive relationship exists
What if we want features with finer granularity Constituent Concept 1 (top of hyperbola) NoisyOR Target Concept Present? AND … Constituent Concept 2 (wings of hyperbola) NoisyOR Our features have more granularity, therefore our concepts may be constituents of a target, rather than encapsulating the target concept  Fine Extraction More detail about image and more shape information, but may loose disjunctive nature between (multiple) instances
GPR Experiments Extensive GPR Data set ~800 targets ~ 5,000 non-targets Experimental Design Run RSF-MIL-d (disjunctive) and RSF-MIL-c (conjunctive) Compare both feature extraction methods Gross extraction: large enough to encompass target concept Fine extraction: Non-overlapping bins Hypothesis RSF-MIL will perform well when using gross extraction whereas RSF-MIL-c will perform well using Fine extraction
Experimental Results Highlights RSF-MIL-d using gross extraction performed best  RSF-MIL-c performed better than RSF-MIL-d when using fine extraction Other influencing factors: optimization methods for RSF-MIL-d and RSF-MIL-c are not the same Gross Extraction Fine Extraction
Future Work Implement a general form that can learn disjunction or conjunction relationship from the data Implement a general form that can learn the number of concepts Incorporate spatial information  Develop an improved optimization scheme for RSF-MIL-C
Backup Slides
MIL Example (AHI Imagery) Robust learning tool MIL tools can learn target signature with limited or incomplete ground truth Which spectral signature(s) should we use to train a target model or classifier? Spectral mixing Background signal Ground truth not exact
MI-RVM Addition of set observations and inference using noisy-OR to an RVM model Prior on the weight w
SVM review Classifier structure Optimization
MI-SVM Discussion RVM was altered to fit MIL problem by changing the form of the target variable’s posterior to model a noisy-OR gate. SVM can be altered to fit the MIL problem by changing how the margin is calculated Boost the margin between the bag (rather than samples) and decision surface Look for the MI separating linear discriminant There is at least one sample from each bag in the half space
mi-SVM Enforce MI scenario using extra constraints Mixed integer program: Must find optimal hyperplane and optimal labeling set At least one sample in each positive bag must have a label of 1. All samples in each negative bag must have a label of -1.
Current Applications Multiple Instance Learning MI Problem MI Applications Multiple Instance Learning: Kernel Machines MI-RVM MI-SVM  Current Applications  GPR imagery HSI imagery
HSI: Target Spectra Learning Given labeled areas of interest: learn target signature Given test areas of interest: classify set of samples
Overview of MI-RVM Optimization Two step optimization Estimate optimal w, given posterior of w There is no closed form solution for the parameters of the posterior, so a gradient update method is used Iterate until convergence. Then proceed to step 2. Update parameter on prior of w The distribution on the target variable has no specific parameters. Until system convergence, continue at step 1.
1) Optimization of w Optimize posterior (Bayes’ Rule) of w Update weights using Newton-Raphsonmethod
2) Optimization of Prior Optimization of covariance of prior Making a large number of assumptions, diagonal elements of A can be estimated
Random Sets: Multiple Instance Learning  Random set framework for multiple instance learning Bags are sets Idea of finding commonality of positive bags inherent in random set formulation Find commonality using intersection operator Random sets governing functional is based on intersection operator
MI issues MIL approaches Some approaches are biased to believe only one sample in each bag caused the target concept Some approaches can only label bags It is not clear whether anything is gained over supervised approaches
RSF-MIL x T x T T x x x x T T x x x MIL-like  Positive Bags = blue Negative Bags = orange Distinct shapes = distinct bags
Side Note: Bayesian Networks Noisy-OR Assumption Bayesian Network representation of Noisy-OR Polytree: singly connected DAG
Side Note Full Bayesian network may be intractable Occurrence of causal factors are rare (sparse co-occurrence) So assume polytree So assume result has boolean relationship with causal factors Absorb I, X and A into one node, governed by randomness of I These assumptions greatly simplify inference calculation Calculate Z based on probabilities rather than constructing a distribution using X
Diverse Density (DD) Probabilistic Approach Goal: Standard statistics approaches identify areas in a feature space with high density of target samples and low density of non-target samples DD: identify areas in a feature space with a high “density” of samples from EACH of the postitive bags (“diverse”), and low density of samples from negative bags. Identify attributes or characteristics similar to positive bags, dissimilar with negative bags Assume t is a target characterization Goal: Assuming the bags are conditionally independent
Diverse Density It is NOT the case that EACH  element is NOT the  target concept Calculation (Noisy-OR Model): Optimization
Random Set Brief Random Set
Random Set Functionals Capacity and avoidance functionals  Given a germ and grain model Assumed random radii
When disjunction makes sense Target Concept Present OR Using Large overlapping bins the target concept can be encapsulated within 1 instance: Therefore a disjunctive relationship exists
Theoretical and Developmental Progress Previous Optimization: Did not necessarily promote  	diverse density Current optimization Better for context learning and MIL Previously no feature relevance or selection (hypersphere) Improvement: included learned weights on each feature dimension  ,[object Object]
Improve Existing Code
Develop joint optimization for context learning and MIL

Weitere ähnliche Inhalte

Was ist angesagt?

Topic Models - LDA and Correlated Topic Models
Topic Models - LDA and Correlated Topic ModelsTopic Models - LDA and Correlated Topic Models
Topic Models - LDA and Correlated Topic ModelsClaudia Wagner
 
Latent dirichletallocation presentation
Latent dirichletallocation presentationLatent dirichletallocation presentation
Latent dirichletallocation presentationSoojung Hong
 
Presentation on Machine Learning and Data Mining
Presentation on Machine Learning and Data MiningPresentation on Machine Learning and Data Mining
Presentation on Machine Learning and Data Miningbutest
 
An exact approach to learning Probabilistic Relational Model
An exact approach to learning Probabilistic Relational ModelAn exact approach to learning Probabilistic Relational Model
An exact approach to learning Probabilistic Relational ModelUniversity of Nantes
 
TopicModels_BleiPaper_Summary.pptx
TopicModels_BleiPaper_Summary.pptxTopicModels_BleiPaper_Summary.pptx
TopicModels_BleiPaper_Summary.pptxKalpit Desai
 
Random Generation of Relational Bayesian Networks
Random Generation of Relational Bayesian NetworksRandom Generation of Relational Bayesian Networks
Random Generation of Relational Bayesian NetworksUniversity of Nantes
 
November, 2006 CCKM'06 1
November, 2006 CCKM'06 1 November, 2006 CCKM'06 1
November, 2006 CCKM'06 1 butest
 
Classifiers
ClassifiersClassifiers
ClassifiersAyurdata
 
Learning Probabilistic Relational Models
Learning Probabilistic Relational ModelsLearning Probabilistic Relational Models
Learning Probabilistic Relational ModelsUniversity of Nantes
 
Tree net and_randomforests_2009
Tree net and_randomforests_2009Tree net and_randomforests_2009
Tree net and_randomforests_2009Matthew Magistrado
 

Was ist angesagt? (13)

Explainable AI
Explainable AIExplainable AI
Explainable AI
 
Topic Models - LDA and Correlated Topic Models
Topic Models - LDA and Correlated Topic ModelsTopic Models - LDA and Correlated Topic Models
Topic Models - LDA and Correlated Topic Models
 
Latent dirichletallocation presentation
Latent dirichletallocation presentationLatent dirichletallocation presentation
Latent dirichletallocation presentation
 
Presentation on Machine Learning and Data Mining
Presentation on Machine Learning and Data MiningPresentation on Machine Learning and Data Mining
Presentation on Machine Learning and Data Mining
 
An exact approach to learning Probabilistic Relational Model
An exact approach to learning Probabilistic Relational ModelAn exact approach to learning Probabilistic Relational Model
An exact approach to learning Probabilistic Relational Model
 
TopicModels_BleiPaper_Summary.pptx
TopicModels_BleiPaper_Summary.pptxTopicModels_BleiPaper_Summary.pptx
TopicModels_BleiPaper_Summary.pptx
 
Topicmodels
TopicmodelsTopicmodels
Topicmodels
 
Random Generation of Relational Bayesian Networks
Random Generation of Relational Bayesian NetworksRandom Generation of Relational Bayesian Networks
Random Generation of Relational Bayesian Networks
 
November, 2006 CCKM'06 1
November, 2006 CCKM'06 1 November, 2006 CCKM'06 1
November, 2006 CCKM'06 1
 
Classifiers
ClassifiersClassifiers
Classifiers
 
Learning Probabilistic Relational Models
Learning Probabilistic Relational ModelsLearning Probabilistic Relational Models
Learning Probabilistic Relational Models
 
Tree net and_randomforests_2009
Tree net and_randomforests_2009Tree net and_randomforests_2009
Tree net and_randomforests_2009
 
Canini09a
Canini09aCanini09a
Canini09a
 

Andere mochten auch

Bayesian Efficient Multiple Kernel Learning
Bayesian Efficient Multiple Kernel LearningBayesian Efficient Multiple Kernel Learning
Bayesian Efficient Multiple Kernel LearningJunya Saito
 
TOMOYOで学ぶ Kernel Watchの秘密
TOMOYOで学ぶ Kernel Watchの秘密TOMOYOで学ぶ Kernel Watchの秘密
TOMOYOで学ぶ Kernel Watchの秘密Motohiro KOSAKI
 
ICML2013読み会 Local Deep Kernel Learning for Efficient Non-linear SVM Prediction
ICML2013読み会 Local Deep Kernel Learning for Efficient Non-linear SVM PredictionICML2013読み会 Local Deep Kernel Learning for Efficient Non-linear SVM Prediction
ICML2013読み会 Local Deep Kernel Learning for Efficient Non-linear SVM PredictionSeiya Tokui
 
Kernel overview
Kernel overviewKernel overview
Kernel overviewKai Sasaki
 
Kernel entropy component analysis
Kernel entropy component analysisKernel entropy component analysis
Kernel entropy component analysisKoichiro Suzuki
 
Tokyor45 カーネル多変量解析第2章 カーネル多変量解析の仕組み
Tokyor45 カーネル多変量解析第2章 カーネル多変量解析の仕組みTokyor45 カーネル多変量解析第2章 カーネル多変量解析の仕組み
Tokyor45 カーネル多変量解析第2章 カーネル多変量解析の仕組みYohei Sato
 
関東CV勉強会 Kernel PCA (2011.2.19)
関東CV勉強会 Kernel PCA (2011.2.19)関東CV勉強会 Kernel PCA (2011.2.19)
関東CV勉強会 Kernel PCA (2011.2.19)Akisato Kimura
 
TEDx Manchester: AI & The Future of Work
TEDx Manchester: AI & The Future of WorkTEDx Manchester: AI & The Future of Work
TEDx Manchester: AI & The Future of WorkVolker Hirsch
 

Andere mochten auch (8)

Bayesian Efficient Multiple Kernel Learning
Bayesian Efficient Multiple Kernel LearningBayesian Efficient Multiple Kernel Learning
Bayesian Efficient Multiple Kernel Learning
 
TOMOYOで学ぶ Kernel Watchの秘密
TOMOYOで学ぶ Kernel Watchの秘密TOMOYOで学ぶ Kernel Watchの秘密
TOMOYOで学ぶ Kernel Watchの秘密
 
ICML2013読み会 Local Deep Kernel Learning for Efficient Non-linear SVM Prediction
ICML2013読み会 Local Deep Kernel Learning for Efficient Non-linear SVM PredictionICML2013読み会 Local Deep Kernel Learning for Efficient Non-linear SVM Prediction
ICML2013読み会 Local Deep Kernel Learning for Efficient Non-linear SVM Prediction
 
Kernel overview
Kernel overviewKernel overview
Kernel overview
 
Kernel entropy component analysis
Kernel entropy component analysisKernel entropy component analysis
Kernel entropy component analysis
 
Tokyor45 カーネル多変量解析第2章 カーネル多変量解析の仕組み
Tokyor45 カーネル多変量解析第2章 カーネル多変量解析の仕組みTokyor45 カーネル多変量解析第2章 カーネル多変量解析の仕組み
Tokyor45 カーネル多変量解析第2章 カーネル多変量解析の仕組み
 
関東CV勉強会 Kernel PCA (2011.2.19)
関東CV勉強会 Kernel PCA (2011.2.19)関東CV勉強会 Kernel PCA (2011.2.19)
関東CV勉強会 Kernel PCA (2011.2.19)
 
TEDx Manchester: AI & The Future of Work
TEDx Manchester: AI & The Future of WorkTEDx Manchester: AI & The Future of Work
TEDx Manchester: AI & The Future of Work
 

Ähnlich wie 110726IGARSS_MIL.pptx

Download It
Download ItDownload It
Download Itbutest
 
Introduction to Machine Learning.
Introduction to Machine Learning.Introduction to Machine Learning.
Introduction to Machine Learning.butest
 
Probability density estimation using Product of Conditional Experts
Probability density estimation using Product of Conditional ExpertsProbability density estimation using Product of Conditional Experts
Probability density estimation using Product of Conditional ExpertsChirag Gupta
 
MachineLearning.ppt
MachineLearning.pptMachineLearning.ppt
MachineLearning.pptbutest
 
MachineLearning.ppt
MachineLearning.pptMachineLearning.ppt
MachineLearning.pptbutest
 
MachineLearning.ppt
MachineLearning.pptMachineLearning.ppt
MachineLearning.pptbutest
 
Introduction to Machine Learning Aristotelis Tsirigos
Introduction to Machine Learning Aristotelis Tsirigos Introduction to Machine Learning Aristotelis Tsirigos
Introduction to Machine Learning Aristotelis Tsirigos butest
 
Machine Learning: Foundations Course Number 0368403401
Machine Learning: Foundations Course Number 0368403401Machine Learning: Foundations Course Number 0368403401
Machine Learning: Foundations Course Number 0368403401butest
 
Machine Learning: Foundations Course Number 0368403401
Machine Learning: Foundations Course Number 0368403401Machine Learning: Foundations Course Number 0368403401
Machine Learning: Foundations Course Number 0368403401butest
 
Intro to Model Selection
Intro to Model SelectionIntro to Model Selection
Intro to Model Selectionchenhm
 
Semi-Supervised.pptx
Semi-Supervised.pptxSemi-Supervised.pptx
Semi-Supervised.pptxTamer Nadeem
 
모듈형 패키지를 활용한 나만의 기계학습 모형 만들기 - 회귀나무모형을 중심으로
모듈형 패키지를 활용한 나만의 기계학습 모형 만들기 - 회귀나무모형을 중심으로 모듈형 패키지를 활용한 나만의 기계학습 모형 만들기 - 회귀나무모형을 중심으로
모듈형 패키지를 활용한 나만의 기계학습 모형 만들기 - 회귀나무모형을 중심으로 r-kor
 
3.5 model based clustering
3.5 model based clustering3.5 model based clustering
3.5 model based clusteringKrish_ver2
 
Introduction
IntroductionIntroduction
Introductionbutest
 
Introduction
IntroductionIntroduction
Introductionbutest
 
Introduction
IntroductionIntroduction
Introductionbutest
 
SVM - Functional Verification
SVM - Functional VerificationSVM - Functional Verification
SVM - Functional VerificationSai Kiran Kadam
 
17- Kernels and Clustering.pptx
17- Kernels and Clustering.pptx17- Kernels and Clustering.pptx
17- Kernels and Clustering.pptxssuser2023c6
 

Ähnlich wie 110726IGARSS_MIL.pptx (20)

Download It
Download ItDownload It
Download It
 
Paris Lecture 1
Paris Lecture 1Paris Lecture 1
Paris Lecture 1
 
Introduction to Machine Learning.
Introduction to Machine Learning.Introduction to Machine Learning.
Introduction to Machine Learning.
 
Probability density estimation using Product of Conditional Experts
Probability density estimation using Product of Conditional ExpertsProbability density estimation using Product of Conditional Experts
Probability density estimation using Product of Conditional Experts
 
MachineLearning.ppt
MachineLearning.pptMachineLearning.ppt
MachineLearning.ppt
 
MachineLearning.ppt
MachineLearning.pptMachineLearning.ppt
MachineLearning.ppt
 
MachineLearning.ppt
MachineLearning.pptMachineLearning.ppt
MachineLearning.ppt
 
Introduction to Machine Learning Aristotelis Tsirigos
Introduction to Machine Learning Aristotelis Tsirigos Introduction to Machine Learning Aristotelis Tsirigos
Introduction to Machine Learning Aristotelis Tsirigos
 
Machine Learning: Foundations Course Number 0368403401
Machine Learning: Foundations Course Number 0368403401Machine Learning: Foundations Course Number 0368403401
Machine Learning: Foundations Course Number 0368403401
 
Machine Learning: Foundations Course Number 0368403401
Machine Learning: Foundations Course Number 0368403401Machine Learning: Foundations Course Number 0368403401
Machine Learning: Foundations Course Number 0368403401
 
Intro to Model Selection
Intro to Model SelectionIntro to Model Selection
Intro to Model Selection
 
Semi-Supervised.pptx
Semi-Supervised.pptxSemi-Supervised.pptx
Semi-Supervised.pptx
 
모듈형 패키지를 활용한 나만의 기계학습 모형 만들기 - 회귀나무모형을 중심으로
모듈형 패키지를 활용한 나만의 기계학습 모형 만들기 - 회귀나무모형을 중심으로 모듈형 패키지를 활용한 나만의 기계학습 모형 만들기 - 회귀나무모형을 중심으로
모듈형 패키지를 활용한 나만의 기계학습 모형 만들기 - 회귀나무모형을 중심으로
 
3.5 model based clustering
3.5 model based clustering3.5 model based clustering
3.5 model based clustering
 
Introduction
IntroductionIntroduction
Introduction
 
Introduction
IntroductionIntroduction
Introduction
 
Introduction
IntroductionIntroduction
Introduction
 
SVM - Functional Verification
SVM - Functional VerificationSVM - Functional Verification
SVM - Functional Verification
 
17- Kernels and Clustering.pptx
17- Kernels and Clustering.pptx17- Kernels and Clustering.pptx
17- Kernels and Clustering.pptx
 
Clustering
ClusteringClustering
Clustering
 

Mehr von grssieee

Tangent height accuracy of Superconducting Submillimeter-Wave Limb-Emission S...
Tangent height accuracy of Superconducting Submillimeter-Wave Limb-Emission S...Tangent height accuracy of Superconducting Submillimeter-Wave Limb-Emission S...
Tangent height accuracy of Superconducting Submillimeter-Wave Limb-Emission S...grssieee
 
SEGMENTATION OF POLARIMETRIC SAR DATA WITH A MULTI-TEXTURE PRODUCT MODEL
SEGMENTATION OF POLARIMETRIC SAR DATA WITH A MULTI-TEXTURE PRODUCT MODELSEGMENTATION OF POLARIMETRIC SAR DATA WITH A MULTI-TEXTURE PRODUCT MODEL
SEGMENTATION OF POLARIMETRIC SAR DATA WITH A MULTI-TEXTURE PRODUCT MODELgrssieee
 
TWO-POINT STATISTIC OF POLARIMETRIC SAR DATA TWO-POINT STATISTIC OF POLARIMET...
TWO-POINT STATISTIC OF POLARIMETRIC SAR DATA TWO-POINT STATISTIC OF POLARIMET...TWO-POINT STATISTIC OF POLARIMETRIC SAR DATA TWO-POINT STATISTIC OF POLARIMET...
TWO-POINT STATISTIC OF POLARIMETRIC SAR DATA TWO-POINT STATISTIC OF POLARIMET...grssieee
 
THE SENTINEL-1 MISSION AND ITS APPLICATION CAPABILITIES
THE SENTINEL-1 MISSION AND ITS APPLICATION CAPABILITIESTHE SENTINEL-1 MISSION AND ITS APPLICATION CAPABILITIES
THE SENTINEL-1 MISSION AND ITS APPLICATION CAPABILITIESgrssieee
 
GMES SPACE COMPONENT:PROGRAMMATIC STATUS
GMES SPACE COMPONENT:PROGRAMMATIC STATUSGMES SPACE COMPONENT:PROGRAMMATIC STATUS
GMES SPACE COMPONENT:PROGRAMMATIC STATUSgrssieee
 
PROGRESSES OF DEVELOPMENT OF CFOSAT SCATTEROMETER
PROGRESSES OF DEVELOPMENT OF CFOSAT SCATTEROMETERPROGRESSES OF DEVELOPMENT OF CFOSAT SCATTEROMETER
PROGRESSES OF DEVELOPMENT OF CFOSAT SCATTEROMETERgrssieee
 
DEVELOPMENT OF ALGORITHMS AND PRODUCTS FOR SUPPORTING THE ITALIAN HYPERSPECTR...
DEVELOPMENT OF ALGORITHMS AND PRODUCTS FOR SUPPORTING THE ITALIAN HYPERSPECTR...DEVELOPMENT OF ALGORITHMS AND PRODUCTS FOR SUPPORTING THE ITALIAN HYPERSPECTR...
DEVELOPMENT OF ALGORITHMS AND PRODUCTS FOR SUPPORTING THE ITALIAN HYPERSPECTR...grssieee
 
EO-1/HYPERION: NEARING TWELVE YEARS OF SUCCESSFUL MISSION SCIENCE OPERATION A...
EO-1/HYPERION: NEARING TWELVE YEARS OF SUCCESSFUL MISSION SCIENCE OPERATION A...EO-1/HYPERION: NEARING TWELVE YEARS OF SUCCESSFUL MISSION SCIENCE OPERATION A...
EO-1/HYPERION: NEARING TWELVE YEARS OF SUCCESSFUL MISSION SCIENCE OPERATION A...grssieee
 
EO-1/HYPERION: NEARING TWELVE YEARS OF SUCCESSFUL MISSION SCIENCE OPERATION A...
EO-1/HYPERION: NEARING TWELVE YEARS OF SUCCESSFUL MISSION SCIENCE OPERATION A...EO-1/HYPERION: NEARING TWELVE YEARS OF SUCCESSFUL MISSION SCIENCE OPERATION A...
EO-1/HYPERION: NEARING TWELVE YEARS OF SUCCESSFUL MISSION SCIENCE OPERATION A...grssieee
 
EO-1/HYPERION: NEARING TWELVE YEARS OF SUCCESSFUL MISSION SCIENCE OPERATION A...
EO-1/HYPERION: NEARING TWELVE YEARS OF SUCCESSFUL MISSION SCIENCE OPERATION A...EO-1/HYPERION: NEARING TWELVE YEARS OF SUCCESSFUL MISSION SCIENCE OPERATION A...
EO-1/HYPERION: NEARING TWELVE YEARS OF SUCCESSFUL MISSION SCIENCE OPERATION A...grssieee
 
test 34mb wo animations
test  34mb wo animationstest  34mb wo animations
test 34mb wo animationsgrssieee
 
2011_Fox_Tax_Worksheets.pdf
2011_Fox_Tax_Worksheets.pdf2011_Fox_Tax_Worksheets.pdf
2011_Fox_Tax_Worksheets.pdfgrssieee
 
DLR open house
DLR open houseDLR open house
DLR open housegrssieee
 
DLR open house
DLR open houseDLR open house
DLR open housegrssieee
 
DLR open house
DLR open houseDLR open house
DLR open housegrssieee
 
Tana_IGARSS2011.ppt
Tana_IGARSS2011.pptTana_IGARSS2011.ppt
Tana_IGARSS2011.pptgrssieee
 
Solaro_IGARSS_2011.ppt
Solaro_IGARSS_2011.pptSolaro_IGARSS_2011.ppt
Solaro_IGARSS_2011.pptgrssieee
 

Mehr von grssieee (20)

Tangent height accuracy of Superconducting Submillimeter-Wave Limb-Emission S...
Tangent height accuracy of Superconducting Submillimeter-Wave Limb-Emission S...Tangent height accuracy of Superconducting Submillimeter-Wave Limb-Emission S...
Tangent height accuracy of Superconducting Submillimeter-Wave Limb-Emission S...
 
SEGMENTATION OF POLARIMETRIC SAR DATA WITH A MULTI-TEXTURE PRODUCT MODEL
SEGMENTATION OF POLARIMETRIC SAR DATA WITH A MULTI-TEXTURE PRODUCT MODELSEGMENTATION OF POLARIMETRIC SAR DATA WITH A MULTI-TEXTURE PRODUCT MODEL
SEGMENTATION OF POLARIMETRIC SAR DATA WITH A MULTI-TEXTURE PRODUCT MODEL
 
TWO-POINT STATISTIC OF POLARIMETRIC SAR DATA TWO-POINT STATISTIC OF POLARIMET...
TWO-POINT STATISTIC OF POLARIMETRIC SAR DATA TWO-POINT STATISTIC OF POLARIMET...TWO-POINT STATISTIC OF POLARIMETRIC SAR DATA TWO-POINT STATISTIC OF POLARIMET...
TWO-POINT STATISTIC OF POLARIMETRIC SAR DATA TWO-POINT STATISTIC OF POLARIMET...
 
THE SENTINEL-1 MISSION AND ITS APPLICATION CAPABILITIES
THE SENTINEL-1 MISSION AND ITS APPLICATION CAPABILITIESTHE SENTINEL-1 MISSION AND ITS APPLICATION CAPABILITIES
THE SENTINEL-1 MISSION AND ITS APPLICATION CAPABILITIES
 
GMES SPACE COMPONENT:PROGRAMMATIC STATUS
GMES SPACE COMPONENT:PROGRAMMATIC STATUSGMES SPACE COMPONENT:PROGRAMMATIC STATUS
GMES SPACE COMPONENT:PROGRAMMATIC STATUS
 
PROGRESSES OF DEVELOPMENT OF CFOSAT SCATTEROMETER
PROGRESSES OF DEVELOPMENT OF CFOSAT SCATTEROMETERPROGRESSES OF DEVELOPMENT OF CFOSAT SCATTEROMETER
PROGRESSES OF DEVELOPMENT OF CFOSAT SCATTEROMETER
 
DEVELOPMENT OF ALGORITHMS AND PRODUCTS FOR SUPPORTING THE ITALIAN HYPERSPECTR...
DEVELOPMENT OF ALGORITHMS AND PRODUCTS FOR SUPPORTING THE ITALIAN HYPERSPECTR...DEVELOPMENT OF ALGORITHMS AND PRODUCTS FOR SUPPORTING THE ITALIAN HYPERSPECTR...
DEVELOPMENT OF ALGORITHMS AND PRODUCTS FOR SUPPORTING THE ITALIAN HYPERSPECTR...
 
EO-1/HYPERION: NEARING TWELVE YEARS OF SUCCESSFUL MISSION SCIENCE OPERATION A...
EO-1/HYPERION: NEARING TWELVE YEARS OF SUCCESSFUL MISSION SCIENCE OPERATION A...EO-1/HYPERION: NEARING TWELVE YEARS OF SUCCESSFUL MISSION SCIENCE OPERATION A...
EO-1/HYPERION: NEARING TWELVE YEARS OF SUCCESSFUL MISSION SCIENCE OPERATION A...
 
EO-1/HYPERION: NEARING TWELVE YEARS OF SUCCESSFUL MISSION SCIENCE OPERATION A...
EO-1/HYPERION: NEARING TWELVE YEARS OF SUCCESSFUL MISSION SCIENCE OPERATION A...EO-1/HYPERION: NEARING TWELVE YEARS OF SUCCESSFUL MISSION SCIENCE OPERATION A...
EO-1/HYPERION: NEARING TWELVE YEARS OF SUCCESSFUL MISSION SCIENCE OPERATION A...
 
EO-1/HYPERION: NEARING TWELVE YEARS OF SUCCESSFUL MISSION SCIENCE OPERATION A...
EO-1/HYPERION: NEARING TWELVE YEARS OF SUCCESSFUL MISSION SCIENCE OPERATION A...EO-1/HYPERION: NEARING TWELVE YEARS OF SUCCESSFUL MISSION SCIENCE OPERATION A...
EO-1/HYPERION: NEARING TWELVE YEARS OF SUCCESSFUL MISSION SCIENCE OPERATION A...
 
Test
TestTest
Test
 
test 34mb wo animations
test  34mb wo animationstest  34mb wo animations
test 34mb wo animations
 
Test 70MB
Test 70MBTest 70MB
Test 70MB
 
Test 70MB
Test 70MBTest 70MB
Test 70MB
 
2011_Fox_Tax_Worksheets.pdf
2011_Fox_Tax_Worksheets.pdf2011_Fox_Tax_Worksheets.pdf
2011_Fox_Tax_Worksheets.pdf
 
DLR open house
DLR open houseDLR open house
DLR open house
 
DLR open house
DLR open houseDLR open house
DLR open house
 
DLR open house
DLR open houseDLR open house
DLR open house
 
Tana_IGARSS2011.ppt
Tana_IGARSS2011.pptTana_IGARSS2011.ppt
Tana_IGARSS2011.ppt
 
Solaro_IGARSS_2011.ppt
Solaro_IGARSS_2011.pptSolaro_IGARSS_2011.ppt
Solaro_IGARSS_2011.ppt
 

110726IGARSS_MIL.pptx

  • 1. Conjuntive Formulation of the Random Set Framework for Multiple Instance Learning:Application to Remote Sensing Jeremy Bolton Paul Gader CSI Laboratory University of Florida
  • 2. Highlights Conjunctive forms of Random Sets for Multiple Instance Learning: Random Sets can be used to solve MIL problem when multiple concepts are present Previously Developed Formulations assume Disjunctive relationship between concepts learned New formulation provides for a conjunctive relationship between concepts and its utility is exhibited on a Ground Penetrating Radar (GPR) data set
  • 3. Outline Multiple Instance Learning MI Problem RSF-MIL Multiple Target Concepts Experimental Results GPR Experiments Future Work
  • 5. Standard Learning vs. Multiple Instance Learning Standard supervised learning Optimize some model (or learn a target concept) given training samples and corresponding labels MIL Learn a target concept given multiplesets of samples and corresponding labels for the sets. Interpretation: Learning with uncertain labels / noisy teacher
  • 6. Multiple Instance Learning (MIL) Given: Set of I bags Labeled + or - The ith bag is a set of Ji samples in some feature space Interpretation of labels Goal: learn concept What characteristic is common to the positive bags that is not observed in the negative bags
  • 7. Multiple Instance Learning Traditional Classification Multiple Instance Learning {x1, x2, x3, x4} label = 1 {x1, x2, x3, x4} label = 1 {x1, x2, x3, x4} label = 0 x1 label = 1 x2label = 1 x3label = 0 x4label = 0 x5label = 1
  • 8. MIL Application: Example GPR EHD: Feature Vector Collaboration: Frigui, Collins, Torrione Construction of bags Collect 15 EHD feature vectors from the 15 depth bins Mine images = + bags FA images = - bags
  • 9. Standard vs. MI Learning: GPR Example Standard Learning Each training sample (feature vector) must have a label Arduous task many feature vectors per image and multiple images difficult to label given GPR echoes, ground truthing errors, etc … label of each vector may not be known EHD: Feature Vector
  • 10.
  • 11.
  • 12. Random Set Brief Random Set
  • 13. How can we use Random Sets for MIL? It is NOT the case that EACH element is NOT the target concept Random set for MIL: Bags are sets (multi-sets) Idea of finding commonality of positive bags inherent in random set formulation Sets have an empty intersection or non-empty intersection relationship Find commonality using intersection operator Random sets governing functional is based on intersection operator Capacity functional : T A.K.A. : Noisy-OR gate (Pearl 1988)
  • 14. Random Set Functionals Capacity functionals for intersection calculation Use germ and grain model to model random set Multiple (J) Concepts Calculate probability of intersection given X and germ and grain pairs: Grains are governed by random radii with assumed cumulative: Random Set model parameters Germ Grain
  • 15. RSF-MIL: Germ and Grain Model x T x T T x x x x T T x x x Positive Bags = blue Negative Bags = orange Distinct shapes = distinct bags
  • 16. Multiple Instance Learning with Multiple Concepts
  • 17. Multiple Concepts: Disjunction or Conjunction? Disjunction When you have multiple types of concepts When each instance can indicate the presence of a target Conjunction When you have a target type that is composed of multiple (necessary concepts) When each instance can indicate a concept, but not necessary the composite target type
  • 18. Conjunctive RSF-MIL Previously Developed Disjunctive RSF-MIL (RSF-MIL-d) Conjunctive RSF-MIL (RSF-MIL-c) Noisy-OR combination across concepts and samples Standard noisy-OR for one concept j Noisy-AND combination across concepts
  • 19. Synthetic Data Experiments Extreme Conjunct data set requires that a target bag exhibits two distinct concepts rather than one or none AUC (AUC when initialized near solution)
  • 21. Disjunctive Target Concepts Target Concept Type 1 NoisyOR Target Concept Type 2 NoisyOR OR … Target Concept Type n NoisyOR Target Concept Present? Using Large overlapping bins (GROSS Extraction) the target concept can be encapsulated within 1 instance: Therefore a disjunctive relationship exists
  • 22. What if we want features with finer granularity Constituent Concept 1 (top of hyperbola) NoisyOR Target Concept Present? AND … Constituent Concept 2 (wings of hyperbola) NoisyOR Our features have more granularity, therefore our concepts may be constituents of a target, rather than encapsulating the target concept Fine Extraction More detail about image and more shape information, but may loose disjunctive nature between (multiple) instances
  • 23. GPR Experiments Extensive GPR Data set ~800 targets ~ 5,000 non-targets Experimental Design Run RSF-MIL-d (disjunctive) and RSF-MIL-c (conjunctive) Compare both feature extraction methods Gross extraction: large enough to encompass target concept Fine extraction: Non-overlapping bins Hypothesis RSF-MIL will perform well when using gross extraction whereas RSF-MIL-c will perform well using Fine extraction
  • 24. Experimental Results Highlights RSF-MIL-d using gross extraction performed best RSF-MIL-c performed better than RSF-MIL-d when using fine extraction Other influencing factors: optimization methods for RSF-MIL-d and RSF-MIL-c are not the same Gross Extraction Fine Extraction
  • 25. Future Work Implement a general form that can learn disjunction or conjunction relationship from the data Implement a general form that can learn the number of concepts Incorporate spatial information Develop an improved optimization scheme for RSF-MIL-C
  • 27. MIL Example (AHI Imagery) Robust learning tool MIL tools can learn target signature with limited or incomplete ground truth Which spectral signature(s) should we use to train a target model or classifier? Spectral mixing Background signal Ground truth not exact
  • 28. MI-RVM Addition of set observations and inference using noisy-OR to an RVM model Prior on the weight w
  • 29. SVM review Classifier structure Optimization
  • 30. MI-SVM Discussion RVM was altered to fit MIL problem by changing the form of the target variable’s posterior to model a noisy-OR gate. SVM can be altered to fit the MIL problem by changing how the margin is calculated Boost the margin between the bag (rather than samples) and decision surface Look for the MI separating linear discriminant There is at least one sample from each bag in the half space
  • 31. mi-SVM Enforce MI scenario using extra constraints Mixed integer program: Must find optimal hyperplane and optimal labeling set At least one sample in each positive bag must have a label of 1. All samples in each negative bag must have a label of -1.
  • 32. Current Applications Multiple Instance Learning MI Problem MI Applications Multiple Instance Learning: Kernel Machines MI-RVM MI-SVM Current Applications GPR imagery HSI imagery
  • 33. HSI: Target Spectra Learning Given labeled areas of interest: learn target signature Given test areas of interest: classify set of samples
  • 34. Overview of MI-RVM Optimization Two step optimization Estimate optimal w, given posterior of w There is no closed form solution for the parameters of the posterior, so a gradient update method is used Iterate until convergence. Then proceed to step 2. Update parameter on prior of w The distribution on the target variable has no specific parameters. Until system convergence, continue at step 1.
  • 35. 1) Optimization of w Optimize posterior (Bayes’ Rule) of w Update weights using Newton-Raphsonmethod
  • 36. 2) Optimization of Prior Optimization of covariance of prior Making a large number of assumptions, diagonal elements of A can be estimated
  • 37. Random Sets: Multiple Instance Learning Random set framework for multiple instance learning Bags are sets Idea of finding commonality of positive bags inherent in random set formulation Find commonality using intersection operator Random sets governing functional is based on intersection operator
  • 38. MI issues MIL approaches Some approaches are biased to believe only one sample in each bag caused the target concept Some approaches can only label bags It is not clear whether anything is gained over supervised approaches
  • 39. RSF-MIL x T x T T x x x x T T x x x MIL-like Positive Bags = blue Negative Bags = orange Distinct shapes = distinct bags
  • 40. Side Note: Bayesian Networks Noisy-OR Assumption Bayesian Network representation of Noisy-OR Polytree: singly connected DAG
  • 41. Side Note Full Bayesian network may be intractable Occurrence of causal factors are rare (sparse co-occurrence) So assume polytree So assume result has boolean relationship with causal factors Absorb I, X and A into one node, governed by randomness of I These assumptions greatly simplify inference calculation Calculate Z based on probabilities rather than constructing a distribution using X
  • 42. Diverse Density (DD) Probabilistic Approach Goal: Standard statistics approaches identify areas in a feature space with high density of target samples and low density of non-target samples DD: identify areas in a feature space with a high “density” of samples from EACH of the postitive bags (“diverse”), and low density of samples from negative bags. Identify attributes or characteristics similar to positive bags, dissimilar with negative bags Assume t is a target characterization Goal: Assuming the bags are conditionally independent
  • 43. Diverse Density It is NOT the case that EACH element is NOT the target concept Calculation (Noisy-OR Model): Optimization
  • 44. Random Set Brief Random Set
  • 45. Random Set Functionals Capacity and avoidance functionals Given a germ and grain model Assumed random radii
  • 46. When disjunction makes sense Target Concept Present OR Using Large overlapping bins the target concept can be encapsulated within 1 instance: Therefore a disjunctive relationship exists
  • 47.
  • 49. Develop joint optimization for context learning and MIL
  • 50. Apply MIL approaches (broad scale)
  • 51. Learn similarities between feature sets of mines
  • 52. Aid in training existing algos: find “best” EHD features for training / testing
  • 53.
  • 54. How can we use Random Sets for MIL? Random set for MIL: Bags are sets Idea of finding commonality of positive bags inherent in random set formulation Sets have an empty intersection or non-empty intersection relationship Find commonality using intersection operator Random sets governing functional is based on intersection operator Example: Bags with target {l,a,e,i,o,p,u,f} {f,b,a,e,i,z,o,u} {a,b,c,i,o,u,e,p,f} {a,f,t,e,i,u,o,d,v} Bags without target {s,r,n,m,p,l} {z,s,w,t,g,n,c} {f,p,k,r} {q,x,z,c,v} {p,l,f} intersection union Target concept = br />{a,e,i,o,u,f} {f,s,r,n,m,p,l,z,w,g,n,c,v,q,k} = {a,e,i,o,u}

Hinweis der Redaktion

  1. Explain GPR images and target signatures.Given a GPR image, typically multiple features vectors are calculated at each depth bin or image subsets. Note that some feature vectors exhibit the target concept and some do not, which ones exhibit it can be considered uncertain, unless an expert is used label each feature vector. Note that this is exactly the multiple instance scenario – when optimizing a classifier for landmine detection we are learning in conditions of uncertainty: we know that there is a target in this image, but we don’t know which features vectors contain the target and which do not.
  2. After producing multiple sets for multiple GPR images, the multiple instance learner will 1) identify the commonalities (common patterns) shared by the positives bags that are not observed in the negative bags – it will learn the target concept. 2) given the classifier/model chosen, it will aid in the optimization of classifier or model parameters.Some supervised, semi-supervised, or active learning methods may attempt to assign labels to all training samples, such that some expert is aiding, some criterion is satisfied, or some objective is optimized. With multiple instance learning, we say, FORGET ABOUT IT. The multiple instance learner will figure it out.