You searched for subject:(unsupervised learning)
.
Showing records 1 – 30 of
488 total matches.
◁ [1] [2] [3] [4] [5] … [17] ▶
1.
Kocher, MIrco.
Text clustering with styles.
Degree: 2017, Université de Neuchâtel
URL: http://doc.rero.ch/record/306696
► Cette thèse présente le problème du regroupement d'auteurs formulé de la manière suivante : en partant d'un ensemble composé de <i>n</i> textes, le but est…
(more)
▼ Cette thèse présente le problème du regroupement
d'auteurs formulé de la manière suivante : en partant d'un ensemble
composé de <i>n</i> textes, le but est de déterminer le
nombre <i>k</i> d'auteurs distincts, pour regrouper les
textes en <i>k</i> classes. De manière itérative, nous
construisons un système stable et simple qui est capable de
regrouper automatiquement les documents selon leurs
thèmes. Dans notre étude, nous commençons par proposer
une mesure capable d'estimer l'(in-)certitude de la décision
proposée, dans le but d'obtenir un indicateur de confiance en lieu
et place d'une simple réponse. Ensuite, nous combinons les paires
de textes pour lesquelles une même affectation apparaît, et dont
nous sommes suffisamment confiants pour affirmer qu'ils sont
rédigés par le même auteur. Enfin, après avoir vérifié chaque tuple
de textes, nous construisons les classes en nous basant sur une
stratégie utilisant une distance entre distributions probabilistes.
Grâce à l'utilisation d'une limite dynamique, nous sommes à même de
choisir les plus petites distances relatives pour détecter une
origine commune entre textes. Bien que notre étude se
concentre principalement sur la création de méthodes simples, des
schémas plus complexes mènent à des résultats plus performants.
Ainsi, nous avons opté pour une représentation distribuée et nous
avons comparé son efficacité à plusieurs méthodes d'attribution
d'auteurs. Cette évaluation nous permet de démontrer que toutes les
approches n'excellent pas dans toutes les situations, et que des
méthodes d'apprentissage profond peuvent être sensibles au choix
des paramètres. Les observations les plus proches des
exemples en question (ou la catégorie ayant la plus petite
distance) déterminent généralement les réponses proposées. Nous
avons testé plusieurs fonctions de distance inter-textuelle sur des
critères théoriques et empiriques. Nous démontrons que les
distances dites de Tanimoto et de Matusita respectent toutes les
propriétés théoriques. Toutes deux obtiennent également de bons
résultats dans le cadre de tests empiriques. Toutefois, les mesures
de Canberra et de Clark sont encore mieux adaptées, bien qu'elles
ne remplissent pas toutes les caractéristiques théoriques
demandées. De manière générale, l'on constate que la fonction
Cosinus ne répond pas à toutes les conditions, et se comporte de
façon suboptimale. Enfin, nous observons que la réduction des
traits stylistiques retenues diminue le temps d'exécution et peut
également améliorer les performances en ignorant les
redondantes. Nous testons nos systèmes pour différentes
langues naturelles appartenant à une variété de familles de langues
et pour plusieurs genres de textes. Grâce à la sélection flexible
des attributs, nos systèmes sont capables de produire des résultats
fiables dans toutes les conditions testées.
Advisors/Committee Members: Jacques (Dir.).
Subjects/Keywords: unsupervised learning
Record Details
Similar Records
Cite
Share »
Record Details
Similar Records
Cite
« Share





❌
APA ·
Chicago ·
MLA ·
Vancouver ·
CSE |
Export
to Zotero / EndNote / Reference
Manager
APA (6th Edition):
Kocher, M. (2017). Text clustering with styles. (Thesis). Université de Neuchâtel. Retrieved from http://doc.rero.ch/record/306696
Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation
Chicago Manual of Style (16th Edition):
Kocher, MIrco. “Text clustering with styles.” 2017. Thesis, Université de Neuchâtel. Accessed February 27, 2021.
http://doc.rero.ch/record/306696.
Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation
MLA Handbook (7th Edition):
Kocher, MIrco. “Text clustering with styles.” 2017. Web. 27 Feb 2021.
Vancouver:
Kocher M. Text clustering with styles. [Internet] [Thesis]. Université de Neuchâtel; 2017. [cited 2021 Feb 27].
Available from: http://doc.rero.ch/record/306696.
Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation
Council of Science Editors:
Kocher M. Text clustering with styles. [Thesis]. Université de Neuchâtel; 2017. Available from: http://doc.rero.ch/record/306696
Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation

Victoria University of Wellington
2.
Butler-Yeoman, Tony.
Learning to Disentangle the Complex Causes of Data.
Degree: 2017, Victoria University of Wellington
URL: http://hdl.handle.net/10063/6951
► The ability to extract and model the meaning in data has been key to the success of modern machine learning. Typically, data reflects a combination…
(more)
▼ The ability to extract and model the meaning in data has been key to the success of modern machine
learning. Typically, data reflects a combination of multiple sources that are mixed together. For example, photographs of people’s faces reflect the
subject of the photograph, lighting conditions, angle, and background scene. It is therefore natural to wish to extract these multiple, largely independent, sources, which is known as disentangling in the literature. Additional benefits of disentangling arise from the fact that the data is then simpler, meaning that there are fewer free parameters, which reduces the curse of dimensionality and aids
learning.
While there has been a lot of research into finding disentangled representations, it remains an open problem. This thesis considers a number of approaches to a particularly difficult version of this task: we wish to disentangle the complex causes of data in an entirely
unsupervised setting. That is, given access only to unlabeled, entangled data, we search for algorithms that can identify the generative factors of that data, which we call causes. Further, we assume that causes can themselves be complex and require a high-dimensional representation.
We consider three approaches to this challenge: as an inference problem, as an extension of independent components analysis, and as a
learning problem. Each method is motivated, described, and tested on a set of datasets build from entangled combinations of images, most commonly MNIST digits. Where the results fall short of disentangling, the reasons for this are dissected and analysed. The last method that we describe, which is based on combinations of autoencoders that learn to predict each other’s output, shows some promise on this extremely challenging problem.
Advisors/Committee Members: Frean, Marcus, Marsland, Stephen.
Subjects/Keywords: Unsupervised; Machine; Learning
Record Details
Similar Records
Cite
Share »
Record Details
Similar Records
Cite
« Share





❌
APA ·
Chicago ·
MLA ·
Vancouver ·
CSE |
Export
to Zotero / EndNote / Reference
Manager
APA (6th Edition):
Butler-Yeoman, T. (2017). Learning to Disentangle the Complex Causes of Data. (Masters Thesis). Victoria University of Wellington. Retrieved from http://hdl.handle.net/10063/6951
Chicago Manual of Style (16th Edition):
Butler-Yeoman, Tony. “Learning to Disentangle the Complex Causes of Data.” 2017. Masters Thesis, Victoria University of Wellington. Accessed February 27, 2021.
http://hdl.handle.net/10063/6951.
MLA Handbook (7th Edition):
Butler-Yeoman, Tony. “Learning to Disentangle the Complex Causes of Data.” 2017. Web. 27 Feb 2021.
Vancouver:
Butler-Yeoman T. Learning to Disentangle the Complex Causes of Data. [Internet] [Masters thesis]. Victoria University of Wellington; 2017. [cited 2021 Feb 27].
Available from: http://hdl.handle.net/10063/6951.
Council of Science Editors:
Butler-Yeoman T. Learning to Disentangle the Complex Causes of Data. [Masters Thesis]. Victoria University of Wellington; 2017. Available from: http://hdl.handle.net/10063/6951
3.
Aversano, Gianmarco.
Development of physics-based reduced-order models for reacting flow applications : Développement de modèles d’ordre réduit basés sur la physique pour les applications d’écoulement réactif.
Degree: Docteur es, Combustion, 2019, Université Paris-Saclay (ComUE); Université libre de Bruxelles (1970-....)
URL: http://www.theses.fr/2019SACLC095
► L’objectif final étant de développer des modèles d’ordre réduit pour les applications de combustion, des techniques d’apprentissage automatique non supervisées et supervisées ont été testées…
(more)
▼ L’objectif final étant de développer des modèles d’ordre réduit pour les applications de combustion, des techniques d’apprentissage automatique non supervisées et supervisées ont été testées et combinées dans les travaux de la présente thèse pour l’extraction de caractéristiques et la construction de modèles d’ordre réduit. Ainsi, l’application de techniques pilotées par les données pour la détection des caractéristiques d’ensembles de données de combustion turbulente (simulation numérique directe) a été étudiée sur deux flammes H2 / CO: une évolution spatiale (DNS1) et une jet à évolution temporelle (DNS2). Des méthodes telles que l’analyse en composantes principales (ACP), l’analyse en composantes principales locales (LPCA), la factorisation matricielle non négative (NMF) et les autoencodeurs ont été explorées à cette fin. Il a été démontré que divers facteurs pouvaient affecter les performances de ces méthodes, tels que les critères utilisés pour le centrage et la mise à l’échelle des données d’origine ou le choix du nombre de dimensions dans les approximations de rang inférieur. Un ensemble de lignes directrices a été présenté qui peut aider le processus d’identification de caractéristiques physiques significatives à partir de données de flux réactifs turbulents. Des méthodes de compression de données telles que l’analyse en composantes principales (ACP) et les variations ont été combinées à des méthodes d’interpolation telles que le krigeage, pour la construction de modèles ordonnées à prix réduits et calculables pour la prédiction de l’état d’un système de combustion dans des conditions de fonctionnement inconnues ou des combinaisons de modèles valeurs de paramètre d’entrée. La méthodologie a d’abord été testée pour la prévision des flammes 1D avec un nombre croissant de paramètres d’entrée (rapport d’équivalence, composition du carburant et température d’entrée), avec des variantes de l’approche PCA classique, à savoir PCA contrainte et PCA locale, appliquée aux cas de combustion la première fois en combinaison avec une technique d’interpolation. Les résultats positifs de l’étude ont conduit à l’application de la méthodologie proposée aux flammes 2D avec deux paramètres d’entrée, à savoir la composition du combustible et la vitesse d’entrée, qui ont donné des résultats satisfaisants. Des alternatives aux méthodes non supervisées et supervisées choisies ont également été testées sur les mêmes données 2D. L’utilisation de la factorisation matricielle non négative (FNM) pour l’approximation de bas rang a été étudiée en raison de la capacité de la méthode à représenter des données à valeur positive, ce qui permet de ne pas enfreindre des lois physiques importantes telles que la positivité des fractions de masse d’espèces chimiques et comparée à la PCA. Comme méthodes supervisées alternatives, la combinaison de l’expansion du chaos polynomial (PCE) et du Kriging et l’utilisation de réseaux de neurones artificiels (RNA) ont été testées. Les résultats des travaux susmentionnés ont ouvert la voie au développement d’un…
Advisors/Committee Members: Gicquel, Olivier (thesis director), Parente, Alessandro (thesis director).
Subjects/Keywords: Combustion; Unsupervised learning; Supervised learning; Combustion; Unsupervised learning; Supervised learning
Record Details
Similar Records
Cite
Share »
Record Details
Similar Records
Cite
« Share





❌
APA ·
Chicago ·
MLA ·
Vancouver ·
CSE |
Export
to Zotero / EndNote / Reference
Manager
APA (6th Edition):
Aversano, G. (2019). Development of physics-based reduced-order models for reacting flow applications : Développement de modèles d’ordre réduit basés sur la physique pour les applications d’écoulement réactif. (Doctoral Dissertation). Université Paris-Saclay (ComUE); Université libre de Bruxelles (1970-....). Retrieved from http://www.theses.fr/2019SACLC095
Chicago Manual of Style (16th Edition):
Aversano, Gianmarco. “Development of physics-based reduced-order models for reacting flow applications : Développement de modèles d’ordre réduit basés sur la physique pour les applications d’écoulement réactif.” 2019. Doctoral Dissertation, Université Paris-Saclay (ComUE); Université libre de Bruxelles (1970-....). Accessed February 27, 2021.
http://www.theses.fr/2019SACLC095.
MLA Handbook (7th Edition):
Aversano, Gianmarco. “Development of physics-based reduced-order models for reacting flow applications : Développement de modèles d’ordre réduit basés sur la physique pour les applications d’écoulement réactif.” 2019. Web. 27 Feb 2021.
Vancouver:
Aversano G. Development of physics-based reduced-order models for reacting flow applications : Développement de modèles d’ordre réduit basés sur la physique pour les applications d’écoulement réactif. [Internet] [Doctoral dissertation]. Université Paris-Saclay (ComUE); Université libre de Bruxelles (1970-....); 2019. [cited 2021 Feb 27].
Available from: http://www.theses.fr/2019SACLC095.
Council of Science Editors:
Aversano G. Development of physics-based reduced-order models for reacting flow applications : Développement de modèles d’ordre réduit basés sur la physique pour les applications d’écoulement réactif. [Doctoral Dissertation]. Université Paris-Saclay (ComUE); Université libre de Bruxelles (1970-....); 2019. Available from: http://www.theses.fr/2019SACLC095
4.
Hirayama, Jun-ichiro.
Probabilistic approach to unsupervised representation learning in dynamic environments : 動的環境における教師なし表現学習への確率的アプローチ; ドウテキ カンキョウ ニオケル キョウシ ナシ ヒョウゲン ガクシュウ エノ カクリツテキ アプローチ.
Degree: Nara Institute of Science and Technology / 奈良先端科学技術大学院大学
URL: http://hdl.handle.net/10061/4366
Subjects/Keywords: unsupervised learning
Record Details
Similar Records
Cite
Share »
Record Details
Similar Records
Cite
« Share





❌
APA ·
Chicago ·
MLA ·
Vancouver ·
CSE |
Export
to Zotero / EndNote / Reference
Manager
APA (6th Edition):
Hirayama, J. (n.d.). Probabilistic approach to unsupervised representation learning in dynamic environments : 動的環境における教師なし表現学習への確率的アプローチ; ドウテキ カンキョウ ニオケル キョウシ ナシ ヒョウゲン ガクシュウ エノ カクリツテキ アプローチ. (Thesis). Nara Institute of Science and Technology / 奈良先端科学技術大学院大学. Retrieved from http://hdl.handle.net/10061/4366
Note: this citation may be lacking information needed for this citation format:
No year of publication.
Not specified: Masters Thesis or Doctoral Dissertation
Chicago Manual of Style (16th Edition):
Hirayama, Jun-ichiro. “Probabilistic approach to unsupervised representation learning in dynamic environments : 動的環境における教師なし表現学習への確率的アプローチ; ドウテキ カンキョウ ニオケル キョウシ ナシ ヒョウゲン ガクシュウ エノ カクリツテキ アプローチ.” Thesis, Nara Institute of Science and Technology / 奈良先端科学技術大学院大学. Accessed February 27, 2021.
http://hdl.handle.net/10061/4366.
Note: this citation may be lacking information needed for this citation format:
No year of publication.
Not specified: Masters Thesis or Doctoral Dissertation
MLA Handbook (7th Edition):
Hirayama, Jun-ichiro. “Probabilistic approach to unsupervised representation learning in dynamic environments : 動的環境における教師なし表現学習への確率的アプローチ; ドウテキ カンキョウ ニオケル キョウシ ナシ ヒョウゲン ガクシュウ エノ カクリツテキ アプローチ.” Web. 27 Feb 2021.
Note: this citation may be lacking information needed for this citation format:
No year of publication.
Vancouver:
Hirayama J. Probabilistic approach to unsupervised representation learning in dynamic environments : 動的環境における教師なし表現学習への確率的アプローチ; ドウテキ カンキョウ ニオケル キョウシ ナシ ヒョウゲン ガクシュウ エノ カクリツテキ アプローチ. [Internet] [Thesis]. Nara Institute of Science and Technology / 奈良先端科学技術大学院大学; [cited 2021 Feb 27].
Available from: http://hdl.handle.net/10061/4366.
Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation
No year of publication.
Council of Science Editors:
Hirayama J. Probabilistic approach to unsupervised representation learning in dynamic environments : 動的環境における教師なし表現学習への確率的アプローチ; ドウテキ カンキョウ ニオケル キョウシ ナシ ヒョウゲン ガクシュウ エノ カクリツテキ アプローチ. [Thesis]. Nara Institute of Science and Technology / 奈良先端科学技術大学院大学; Available from: http://hdl.handle.net/10061/4366
Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation
No year of publication.
5.
Madokoro, Hirokazu.
Unsupervised Category Formation and Its Applications to Robot Vision : 教師なしカテゴリ形成とロボットビジョンへの応用; キョウシ ナシ カテゴリ ケイセイ ト ロボット ビジョン エノ オウヨウ.
Degree: Nara Institute of Science and Technology / 奈良先端科学技術大学院大学
URL: http://hdl.handle.net/10061/6019
Subjects/Keywords: Unsupervised learning
Record Details
Similar Records
Cite
Share »
Record Details
Similar Records
Cite
« Share





❌
APA ·
Chicago ·
MLA ·
Vancouver ·
CSE |
Export
to Zotero / EndNote / Reference
Manager
APA (6th Edition):
Madokoro, H. (n.d.). Unsupervised Category Formation and Its Applications to Robot Vision : 教師なしカテゴリ形成とロボットビジョンへの応用; キョウシ ナシ カテゴリ ケイセイ ト ロボット ビジョン エノ オウヨウ. (Thesis). Nara Institute of Science and Technology / 奈良先端科学技術大学院大学. Retrieved from http://hdl.handle.net/10061/6019
Note: this citation may be lacking information needed for this citation format:
No year of publication.
Not specified: Masters Thesis or Doctoral Dissertation
Chicago Manual of Style (16th Edition):
Madokoro, Hirokazu. “Unsupervised Category Formation and Its Applications to Robot Vision : 教師なしカテゴリ形成とロボットビジョンへの応用; キョウシ ナシ カテゴリ ケイセイ ト ロボット ビジョン エノ オウヨウ.” Thesis, Nara Institute of Science and Technology / 奈良先端科学技術大学院大学. Accessed February 27, 2021.
http://hdl.handle.net/10061/6019.
Note: this citation may be lacking information needed for this citation format:
No year of publication.
Not specified: Masters Thesis or Doctoral Dissertation
MLA Handbook (7th Edition):
Madokoro, Hirokazu. “Unsupervised Category Formation and Its Applications to Robot Vision : 教師なしカテゴリ形成とロボットビジョンへの応用; キョウシ ナシ カテゴリ ケイセイ ト ロボット ビジョン エノ オウヨウ.” Web. 27 Feb 2021.
Note: this citation may be lacking information needed for this citation format:
No year of publication.
Vancouver:
Madokoro H. Unsupervised Category Formation and Its Applications to Robot Vision : 教師なしカテゴリ形成とロボットビジョンへの応用; キョウシ ナシ カテゴリ ケイセイ ト ロボット ビジョン エノ オウヨウ. [Internet] [Thesis]. Nara Institute of Science and Technology / 奈良先端科学技術大学院大学; [cited 2021 Feb 27].
Available from: http://hdl.handle.net/10061/6019.
Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation
No year of publication.
Council of Science Editors:
Madokoro H. Unsupervised Category Formation and Its Applications to Robot Vision : 教師なしカテゴリ形成とロボットビジョンへの応用; キョウシ ナシ カテゴリ ケイセイ ト ロボット ビジョン エノ オウヨウ. [Thesis]. Nara Institute of Science and Technology / 奈良先端科学技術大学院大学; Available from: http://hdl.handle.net/10061/6019
Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation
No year of publication.

University of Guelph
6.
Im, Jiwoong.
Analyzing Unsupervised Representation Learning Models Under the View of Dynamical Systems.
Degree: Master of Applied Science, School of Engineering, 2015, University of Guelph
URL: https://atrium.lib.uoguelph.ca/xmlui/handle/10214/8809
► The objective of this thesis is to take the dynamical systems approach to understand the unsupervised learning models and learning algorithms. Gated auto-encoders (GAEs) are…
(more)
▼ The objective of this thesis is to take the dynamical systems approach to understand the
unsupervised learning models and
learning algorithms. Gated auto-encoders (GAEs) are an interesting and flexible extension of auto-encoders which can learn transformations among different images or pixel covariances within images. We examine the GAEs' ability to represent different functions or data distributions. We apply a dynamical systems view to GAEs, deriving a scoring function, and drawing connections to RBMs. In the second part of our study, we investigate the performance of Minimum Probability Flow (MPF)
learning for training restricted Boltzmann machines (RBMs). MPF proposes a tractable, consistent, objective function defined in terms of a Taylor expansion of the KL divergence with respect to sampling dynamics. We propose a more general form for the sampling dynamics in MPF, and explore the consequences of different choices for these dynamics for training RBMs.
Advisors/Committee Members: Taylor, Graham W (advisor).
Subjects/Keywords: Machine learning; Deep Learning; unsupervised learning
Record Details
Similar Records
Cite
Share »
Record Details
Similar Records
Cite
« Share





❌
APA ·
Chicago ·
MLA ·
Vancouver ·
CSE |
Export
to Zotero / EndNote / Reference
Manager
APA (6th Edition):
Im, J. (2015). Analyzing Unsupervised Representation Learning Models Under the View of Dynamical Systems. (Masters Thesis). University of Guelph. Retrieved from https://atrium.lib.uoguelph.ca/xmlui/handle/10214/8809
Chicago Manual of Style (16th Edition):
Im, Jiwoong. “Analyzing Unsupervised Representation Learning Models Under the View of Dynamical Systems.” 2015. Masters Thesis, University of Guelph. Accessed February 27, 2021.
https://atrium.lib.uoguelph.ca/xmlui/handle/10214/8809.
MLA Handbook (7th Edition):
Im, Jiwoong. “Analyzing Unsupervised Representation Learning Models Under the View of Dynamical Systems.” 2015. Web. 27 Feb 2021.
Vancouver:
Im J. Analyzing Unsupervised Representation Learning Models Under the View of Dynamical Systems. [Internet] [Masters thesis]. University of Guelph; 2015. [cited 2021 Feb 27].
Available from: https://atrium.lib.uoguelph.ca/xmlui/handle/10214/8809.
Council of Science Editors:
Im J. Analyzing Unsupervised Representation Learning Models Under the View of Dynamical Systems. [Masters Thesis]. University of Guelph; 2015. Available from: https://atrium.lib.uoguelph.ca/xmlui/handle/10214/8809

University of Toronto
7.
Makhzani, Alireza.
Unsupervised Representation Learning with Autoencoders.
Degree: PhD, 2018, University of Toronto
URL: http://hdl.handle.net/1807/89800
► Despite the recent progress in machine learning and deep learning, unsupervised learning still remains a largely unsolved problem. It is widely recognized that unsupervised learning…
(more)
▼ Despite the recent progress in machine
learning and deep
learning,
unsupervised learning still remains a largely unsolved problem. It is widely recognized that
unsupervised learning algorithms that can learn useful representations are needed for solving problems with limited label information. In this thesis, we study the problem of
learning unsupervised representations using autoencoders, and propose regularization techniques that enable autoencoders to learn useful representations of data in
unsupervised and semi-supervised settings. First, we exploit sparsity as a generic prior on the representations of autoencoders and propose sparse autoencoders that can learn sparse representations with very fast inference processes, making them well-suited to large problem sizes where conventional sparse coding algorithms cannot be applied. Next, we study autoencoders from a probabilistic perspective and propose generative autoencoders that use a generative adversarial network (GAN) to match the distribution of the latent code of the autoencoder with a pre-defined prior. We show that these generative autoencoders can learn posterior approximations that are more expressive than tractable densities often used in variational inference. We demonstrate the performance of developed methods of this thesis on real world image datasets and show their applications in generative modeling, clustering, semi-supervised classification and dimensionality reduction.
Advisors/Committee Members: Frey, Brendan, Electrical and Computer Engineering.
Subjects/Keywords: Deep Learning; Machine Learning; Unsupervised Learning; 0984
Record Details
Similar Records
Cite
Share »
Record Details
Similar Records
Cite
« Share





❌
APA ·
Chicago ·
MLA ·
Vancouver ·
CSE |
Export
to Zotero / EndNote / Reference
Manager
APA (6th Edition):
Makhzani, A. (2018). Unsupervised Representation Learning with Autoencoders. (Doctoral Dissertation). University of Toronto. Retrieved from http://hdl.handle.net/1807/89800
Chicago Manual of Style (16th Edition):
Makhzani, Alireza. “Unsupervised Representation Learning with Autoencoders.” 2018. Doctoral Dissertation, University of Toronto. Accessed February 27, 2021.
http://hdl.handle.net/1807/89800.
MLA Handbook (7th Edition):
Makhzani, Alireza. “Unsupervised Representation Learning with Autoencoders.” 2018. Web. 27 Feb 2021.
Vancouver:
Makhzani A. Unsupervised Representation Learning with Autoencoders. [Internet] [Doctoral dissertation]. University of Toronto; 2018. [cited 2021 Feb 27].
Available from: http://hdl.handle.net/1807/89800.
Council of Science Editors:
Makhzani A. Unsupervised Representation Learning with Autoencoders. [Doctoral Dissertation]. University of Toronto; 2018. Available from: http://hdl.handle.net/1807/89800

University of Alberta
8.
White, Martha.
Regularized factor models.
Degree: PhD, Department of Computing Science, 2014, University of Alberta
URL: https://era.library.ualberta.ca/files/gq67jt70x
► This dissertation explores regularized factor models as a simple unification of machine learn- ing problems, with a focus on algorithmic development within this known formalism.…
(more)
▼ This dissertation explores regularized factor models
as a simple unification of machine learn- ing problems, with a
focus on algorithmic development within this known formalism. The
main contributions are (1) the development of generic, efficient
algorithms for a subclass of regularized factorizations and (2) new
unifications that facilitate application of these algorithms to
problems previously without known tractable algorithms.
Concurrently, the generality of the formalism is further
demonstrated with a thorough summary of known, but often scattered,
connections between supervised and unsupervised learning problems
and algorithms. The dissertation first presents the main
algorithmic advances: convex reformulations of non- convex
regularized factorization objectives. A convex reformulation is
developed for a general subset of regularized factor models, with
an efficiently computable optimization for five different
regularization choices. The thesis then describes advances using
these generic convex reformulation techniques in three important
problems: multi-view subspace learning, semi-supervised learn- ing
and estimating autoregressive moving average models. These novel
settings are unified under regularized factor models by
incorporating problem properties in terms of regularization. Once
ex- pressed as regularized factor models, we can take advantage of
the convex reformulation techniques to obtain novel algorithms that
produce global solutions. These advances include the first global
estimation procedure for two-view subspace learning and for
autoregressive moving average models. The simple algorithms
obtained from these general convex reformulation techniques are
empirically shown to be effective across these three problems on a
variety of datasets. This dissertation illustrates that many
problems can be specified as a simple regularized factorization,
that this class is amenable to global optimization and that it is
advantageous to represent machine learning problems as regularized
factor models.
Subjects/Keywords: machine learning; artificial intelligence; unsupervised learning
Record Details
Similar Records
Cite
Share »
Record Details
Similar Records
Cite
« Share





❌
APA ·
Chicago ·
MLA ·
Vancouver ·
CSE |
Export
to Zotero / EndNote / Reference
Manager
APA (6th Edition):
White, M. (2014). Regularized factor models. (Doctoral Dissertation). University of Alberta. Retrieved from https://era.library.ualberta.ca/files/gq67jt70x
Chicago Manual of Style (16th Edition):
White, Martha. “Regularized factor models.” 2014. Doctoral Dissertation, University of Alberta. Accessed February 27, 2021.
https://era.library.ualberta.ca/files/gq67jt70x.
MLA Handbook (7th Edition):
White, Martha. “Regularized factor models.” 2014. Web. 27 Feb 2021.
Vancouver:
White M. Regularized factor models. [Internet] [Doctoral dissertation]. University of Alberta; 2014. [cited 2021 Feb 27].
Available from: https://era.library.ualberta.ca/files/gq67jt70x.
Council of Science Editors:
White M. Regularized factor models. [Doctoral Dissertation]. University of Alberta; 2014. Available from: https://era.library.ualberta.ca/files/gq67jt70x

University of Debrecen
9.
Bod, Gergely.
Self-taught learning: Implementation using MATLAB
.
Degree: DE – TEK – Informatikai Kar, 2014, University of Debrecen
URL: http://hdl.handle.net/2437/178637
► Self-taught learning is a new framework in the domain of machine learning. It has the potential that by using unsupervised learning strategies to automatically learn…
(more)
▼ Self-taught
learning is a new framework in the domain of machine
learning. It has the potential that by using
unsupervised learning strategies to automatically learn and extract information and consequently create a succint representation of the given input. I have presented the theoretical underpinnings of the idea and then proceeded to show a possible implementation of the algorithm in the MATLAB programming language. I have also included my experimental results obtained during my research.
Advisors/Committee Members: Antal, Bálint (advisor).
Subjects/Keywords: unsupervised learning;
machine learning;
neural network;
autoencoder
Record Details
Similar Records
Cite
Share »
Record Details
Similar Records
Cite
« Share





❌
APA ·
Chicago ·
MLA ·
Vancouver ·
CSE |
Export
to Zotero / EndNote / Reference
Manager
APA (6th Edition):
Bod, G. (2014). Self-taught learning: Implementation using MATLAB
. (Thesis). University of Debrecen. Retrieved from http://hdl.handle.net/2437/178637
Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation
Chicago Manual of Style (16th Edition):
Bod, Gergely. “Self-taught learning: Implementation using MATLAB
.” 2014. Thesis, University of Debrecen. Accessed February 27, 2021.
http://hdl.handle.net/2437/178637.
Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation
MLA Handbook (7th Edition):
Bod, Gergely. “Self-taught learning: Implementation using MATLAB
.” 2014. Web. 27 Feb 2021.
Vancouver:
Bod G. Self-taught learning: Implementation using MATLAB
. [Internet] [Thesis]. University of Debrecen; 2014. [cited 2021 Feb 27].
Available from: http://hdl.handle.net/2437/178637.
Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation
Council of Science Editors:
Bod G. Self-taught learning: Implementation using MATLAB
. [Thesis]. University of Debrecen; 2014. Available from: http://hdl.handle.net/2437/178637
Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation

University of Connecticut
10.
Yankee, Tara N.
Rank Aggregation of Feature Scoring Methods for Unsupervised Learning.
Degree: M. Eng., Biomedical Engineering, 2017, University of Connecticut
URL: https://opencommons.uconn.edu/gs_theses/1123
► The ability to collect and store large amounts of data is transforming data-driven discovery; recent technological advances in biology allow systematic data production and…
(more)
▼ The ability to collect and store large amounts of data is transforming data-driven discovery; recent technological advances in biology allow systematic data production and storage at a previously unattainable scale. It is common for biological Big Data to have an order of magnitude or more features than samples. Feature scoring with selection is therefore an essential pre-processing step to finding meaningful clusters in these data. Many feature scoring algorithms have been proposed; they are based on dramatically different ideas about what constitutes a “good” or “important” feature. Motivated by studies in data classification, we use a rank aggregation (RANKAGG) method to combine estimates of feature importance from multiple sources and use a subset of the highest scoring features for subsequent clustering. We demonstrate the performance of RANKAGG on five real-world biological data-sets, and compare the clustering performance of RANKAGG to the thirteen individual feature scoring methods comprising RANKAGG. The rank aggregated features have a mean perfor- mance across the five data-sets equal to the best individual feature scoring method but with lower variance, indicating robust performance across a variety of data. We carefully consider if there is any systematic way to remove rankers from RANKAGG to improve clustering performance. We demonstrate that rank aggregated feature selection yields excellent performance in clustering problems and possibly more im- portantly, greatly limits the risk of choosing a method that is sub-optimal for a given data-set.
Advisors/Committee Members: Kevin Brown, Ph.D., Ion Mandiou, Ph.D., Yong-Jun Shin, M.D., Ph.D., Kevin Brown, Ph.D..
Subjects/Keywords: clustering; ensemble learning; feature selection; unsupervised learning
Record Details
Similar Records
Cite
Share »
Record Details
Similar Records
Cite
« Share





❌
APA ·
Chicago ·
MLA ·
Vancouver ·
CSE |
Export
to Zotero / EndNote / Reference
Manager
APA (6th Edition):
Yankee, T. N. (2017). Rank Aggregation of Feature Scoring Methods for Unsupervised Learning. (Masters Thesis). University of Connecticut. Retrieved from https://opencommons.uconn.edu/gs_theses/1123
Chicago Manual of Style (16th Edition):
Yankee, Tara N. “Rank Aggregation of Feature Scoring Methods for Unsupervised Learning.” 2017. Masters Thesis, University of Connecticut. Accessed February 27, 2021.
https://opencommons.uconn.edu/gs_theses/1123.
MLA Handbook (7th Edition):
Yankee, Tara N. “Rank Aggregation of Feature Scoring Methods for Unsupervised Learning.” 2017. Web. 27 Feb 2021.
Vancouver:
Yankee TN. Rank Aggregation of Feature Scoring Methods for Unsupervised Learning. [Internet] [Masters thesis]. University of Connecticut; 2017. [cited 2021 Feb 27].
Available from: https://opencommons.uconn.edu/gs_theses/1123.
Council of Science Editors:
Yankee TN. Rank Aggregation of Feature Scoring Methods for Unsupervised Learning. [Masters Thesis]. University of Connecticut; 2017. Available from: https://opencommons.uconn.edu/gs_theses/1123
11.
-6888-3095.
Embodied learning for visual recognition.
Degree: PhD, Electrical and Computer Engineering, 2017, University of Texas – Austin
URL: http://hdl.handle.net/2152/63489
► The field of visual recognition in recent years has come to rely on large expensively curated and manually labeled "bags of disembodied images". In the…
(more)
▼ The field of visual recognition in recent years has come to rely on large expensively curated and manually labeled "bags of disembodied images". In the wake of this, my focus has been on understanding and exploiting alternate "free" sources of supervision available to visual
learning agents that are situated within real environments. For example, even simply moving from orderless image collections to continuous visual observations offers opportunities to understand the dynamics and other physical properties of the visual world. Further, embodied agents may have the abilities to move around their environment and/or effect changes within it, in which case these abilities offer new means to acquire useful supervision. In this dissertation, I present my work along this and related directions.
Advisors/Committee Members: Grauman, Kristen Lorraine, 1979- (advisor), Efros, Alexei (committee member), Ghosh, Joydeep (committee member), Niekum, Scott (committee member), Thomaz, Andrea (committee member).
Subjects/Keywords: Computer vision; Unsupervised learning; Embodied learning
Record Details
Similar Records
Cite
Share »
Record Details
Similar Records
Cite
« Share





❌
APA ·
Chicago ·
MLA ·
Vancouver ·
CSE |
Export
to Zotero / EndNote / Reference
Manager
APA (6th Edition):
-6888-3095. (2017). Embodied learning for visual recognition. (Doctoral Dissertation). University of Texas – Austin. Retrieved from http://hdl.handle.net/2152/63489
Note: this citation may be lacking information needed for this citation format:
Author name may be incomplete
Chicago Manual of Style (16th Edition):
-6888-3095. “Embodied learning for visual recognition.” 2017. Doctoral Dissertation, University of Texas – Austin. Accessed February 27, 2021.
http://hdl.handle.net/2152/63489.
Note: this citation may be lacking information needed for this citation format:
Author name may be incomplete
MLA Handbook (7th Edition):
-6888-3095. “Embodied learning for visual recognition.” 2017. Web. 27 Feb 2021.
Note: this citation may be lacking information needed for this citation format:
Author name may be incomplete
Vancouver:
-6888-3095. Embodied learning for visual recognition. [Internet] [Doctoral dissertation]. University of Texas – Austin; 2017. [cited 2021 Feb 27].
Available from: http://hdl.handle.net/2152/63489.
Note: this citation may be lacking information needed for this citation format:
Author name may be incomplete
Council of Science Editors:
-6888-3095. Embodied learning for visual recognition. [Doctoral Dissertation]. University of Texas – Austin; 2017. Available from: http://hdl.handle.net/2152/63489
Note: this citation may be lacking information needed for this citation format:
Author name may be incomplete

University of New South Wales
12.
Xu, Jie.
On-line and unsupervised learning for codebook based visual recognition.
Degree: Computer Science & Engineering, 2011, University of New South Wales
URL: http://handle.unsw.edu.au/1959.4/51513
;
https://unsworks.unsw.edu.au/fapi/datastream/unsworks:10200/SOURCE02?view=true
► In this thesis we develop unsupervised and on-line learning algorithmsfor codebook based visual recognition tasks. First, we study the Prob-abilistic Latent Semantic Analysis (PLSA), which…
(more)
▼ In this thesis we develop
unsupervised and on-line
learning algorithmsfor codebook based visual recognition tasks. First, we study the Prob-abilistic Latent Semantic Analysis (PLSA), which is one instance ofcodebook based recognition models. It has been successfully appliedto visual recognition tasks, such as image categorization, action recog-nition, etc. However it has been learned mainly in batch mode, andtherefore it cannot handle the data that arrives sequentially. We pro-pose a novel on-line
learning algorithm for
learning the parameters ofthe PLSA under that situation. Our contributions are two-fold: (i)an on-line
learning algorithm that learns the parameters of the PLSAmodel from incoming data; (ii) a codebook adaptation algorithm thatcan capture the full characteristics of all features during the learn-ing. Experimental results demonstrate that the proposed algorithmcan handle sequentially arriving data that the batch PLSA learningcannot cope with.We then look at the Implicit Shape Model (ISM) for object detec-tion. ISM is a codebook based model in which object information isretained in codebooks. Existing ISM based methods require manuallabeling of training data. We propose an algorithm that can label thetraining data automatically. We also propose a method for identify-ing moving edges in video frames so that object hypotheses can begenerated only from the moving edges. We compare the proposed al-gorithm with a background subtraction based moving object detectionalgorithm. The experimental results demonstrate that the proposedalgorithm achieves comparable performance to the background sub-traction based counterpart, and it even outperforms the counterpartin complex situations.We then extend the aforementioned batch algorithm for on-line learn-ing. We propose an on-line training data collection algorithm and alsoan on-line codebook based object detector. We evaluate the algorithmon three video datasets. The experimental results demonstrate thatour algorithm outperforms the state-of-the-art on-line conservativelearning algorithm.
Advisors/Committee Members: Wang, Yang, Computer Science & Engineering, Faculty of Engineering, UNSW, Wang, Wei, Computer Science & Engineering, Faculty of Engineering, UNSW, Ye, Getian, Computer Science & Engineering, Faculty of Engineering, UNSW.
Subjects/Keywords: Visual recognition; Online learning; Unsupervised learning
Record Details
Similar Records
Cite
Share »
Record Details
Similar Records
Cite
« Share





❌
APA ·
Chicago ·
MLA ·
Vancouver ·
CSE |
Export
to Zotero / EndNote / Reference
Manager
APA (6th Edition):
Xu, J. (2011). On-line and unsupervised learning for codebook based visual recognition. (Doctoral Dissertation). University of New South Wales. Retrieved from http://handle.unsw.edu.au/1959.4/51513 ; https://unsworks.unsw.edu.au/fapi/datastream/unsworks:10200/SOURCE02?view=true
Chicago Manual of Style (16th Edition):
Xu, Jie. “On-line and unsupervised learning for codebook based visual recognition.” 2011. Doctoral Dissertation, University of New South Wales. Accessed February 27, 2021.
http://handle.unsw.edu.au/1959.4/51513 ; https://unsworks.unsw.edu.au/fapi/datastream/unsworks:10200/SOURCE02?view=true.
MLA Handbook (7th Edition):
Xu, Jie. “On-line and unsupervised learning for codebook based visual recognition.” 2011. Web. 27 Feb 2021.
Vancouver:
Xu J. On-line and unsupervised learning for codebook based visual recognition. [Internet] [Doctoral dissertation]. University of New South Wales; 2011. [cited 2021 Feb 27].
Available from: http://handle.unsw.edu.au/1959.4/51513 ; https://unsworks.unsw.edu.au/fapi/datastream/unsworks:10200/SOURCE02?view=true.
Council of Science Editors:
Xu J. On-line and unsupervised learning for codebook based visual recognition. [Doctoral Dissertation]. University of New South Wales; 2011. Available from: http://handle.unsw.edu.au/1959.4/51513 ; https://unsworks.unsw.edu.au/fapi/datastream/unsworks:10200/SOURCE02?view=true

Georgia Tech
13.
Haresamudram, Harish.
The role of representations in human activity recognition.
Degree: MS, Electrical and Computer Engineering, 2019, Georgia Tech
URL: http://hdl.handle.net/1853/62706
► We investigate the role of representations in sensor based human activity recognition (HAR). In particular, we develop convolutional and recurrent autoencoder architectures for feature learning…
(more)
▼ We investigate the role of representations in sensor based human activity recognition (HAR). In particular, we develop convolutional and recurrent autoencoder architectures for feature
learning and compare their performance to a distribution-based representation as well as a supervised deep
learning representation based on the DeepConvLSTM architecture. This is
motivated by the promises deep
learning methods offer – they learn end-to-end, eliminate the necessity for hand crafting features and generalize well across tasks and datasets. The
choice of studying
unsupervised learning methods is motivated by the fact that they afford the possibility of
learning meaningful representations without the need for labeled data. Such representations allow for leveraging large, unlabeled datasets for performing feature and transfer
learning. The study is performed on five datasets which are diverse in terms of the number of subjects, activities, and settings. The analysis is performed from a wearables standpoint, considering factors such as memory footprint, the effect of dimensionality, and computation time. We find that the convolutional and recurrent autoencoder based representations outperform the distribution-based representation on all datasets. Additionally, we conclude that autoencoder based representations offer comparable performance to supervised Deep-ConvLSTM based representation. On the larger datasets with multiple sensors such as Opportunity and PAMAP2, the convolutional and recurrent autoencoder based representations are observed to be highly effective. Resource-constrained scenarios justify the utilization of the distribution-based representation, which has low computational costs and memory requirements. Finally, when the number of sensors is low, we observe that the vanilla autoencoder based representations produce good performance.
Advisors/Committee Members: Ploetz, Thomas (advisor), Anderson, David V. (advisor), Essa, Irfan (committee member), Vela, Patricio (committee member).
Subjects/Keywords: Unsupervised learning; Human activity recognition; Autoencoder models
Record Details
Similar Records
Cite
Share »
Record Details
Similar Records
Cite
« Share





❌
APA ·
Chicago ·
MLA ·
Vancouver ·
CSE |
Export
to Zotero / EndNote / Reference
Manager
APA (6th Edition):
Haresamudram, H. (2019). The role of representations in human activity recognition. (Masters Thesis). Georgia Tech. Retrieved from http://hdl.handle.net/1853/62706
Chicago Manual of Style (16th Edition):
Haresamudram, Harish. “The role of representations in human activity recognition.” 2019. Masters Thesis, Georgia Tech. Accessed February 27, 2021.
http://hdl.handle.net/1853/62706.
MLA Handbook (7th Edition):
Haresamudram, Harish. “The role of representations in human activity recognition.” 2019. Web. 27 Feb 2021.
Vancouver:
Haresamudram H. The role of representations in human activity recognition. [Internet] [Masters thesis]. Georgia Tech; 2019. [cited 2021 Feb 27].
Available from: http://hdl.handle.net/1853/62706.
Council of Science Editors:
Haresamudram H. The role of representations in human activity recognition. [Masters Thesis]. Georgia Tech; 2019. Available from: http://hdl.handle.net/1853/62706

Universidade Nova
14.
Madsen, Jacob Hastrup.
Outlier detection for improved clustering : empirical research for unsupervised data mining.
Degree: 2018, Universidade Nova
URL: https://www.rcaap.pt/detail.jsp?id=oai:run.unl.pt:10362/34464
► Many clustering algorithms are sensitive to noise disturbing the results when trying to identify and characterize clusters in data. Due to the multidimensional nature of…
(more)
▼ Many clustering algorithms are sensitive to noise disturbing the results when trying to identify and
characterize clusters in data. Due to the multidimensional nature of clustering, the discipline of outlier
detection is a complex task as statistical approaches are not adequate.
In this research work, we contend that for clustering, outliers should be perceived as observations with
deviating characteristics worsening the ratio of intra-cluster and inter-cluster distance. We present a
research question that deals with improving clustering results specifically for the two clustering
algorithms, k-means and hierarchical clustering, by the means of outlier detection. To improve clustering
results, we identify and discuss the literature of outlier detection, and undertake on 11 algorithms and 2
statistical test to the process of treating data prior to clustering. To evaluate the results of applied
clustering, six evaluation metrics are applied, of which one metric is introduced in this study.
Using real world datasets, we demonstrate that outlier detection does improve clustering results with
respect to clustering objectives, but only to an extent where data allows it. That is, if data contains ‘real’
clusters and actual outliers, proper use of outlier algorithms improves clustering significantly. Advantages
and disadvantages for outlier algorithms, when dealing with different types of data, are discussed along
with the different properties of evaluation metrics describing the fulfillment of clustering objectives.
Finally, it is demonstrated that the main challenge of improving clustering results for users, with regards
to outlier detection, is the lack of tools to understand data structures prior to clustering. Future research
is emphasized for tools such as dimension reduction, to help users avoid applying every tool in the toolbox.
Advisors/Committee Members: Costa, Ana Cristina Marinho da.
Subjects/Keywords: Outlier Detection; Unsupervised Learning; Clustering; Data Mining
Record Details
Similar Records
Cite
Share »
Record Details
Similar Records
Cite
« Share





❌
APA ·
Chicago ·
MLA ·
Vancouver ·
CSE |
Export
to Zotero / EndNote / Reference
Manager
APA (6th Edition):
Madsen, J. H. (2018). Outlier detection for improved clustering : empirical research for unsupervised data mining. (Thesis). Universidade Nova. Retrieved from https://www.rcaap.pt/detail.jsp?id=oai:run.unl.pt:10362/34464
Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation
Chicago Manual of Style (16th Edition):
Madsen, Jacob Hastrup. “Outlier detection for improved clustering : empirical research for unsupervised data mining.” 2018. Thesis, Universidade Nova. Accessed February 27, 2021.
https://www.rcaap.pt/detail.jsp?id=oai:run.unl.pt:10362/34464.
Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation
MLA Handbook (7th Edition):
Madsen, Jacob Hastrup. “Outlier detection for improved clustering : empirical research for unsupervised data mining.” 2018. Web. 27 Feb 2021.
Vancouver:
Madsen JH. Outlier detection for improved clustering : empirical research for unsupervised data mining. [Internet] [Thesis]. Universidade Nova; 2018. [cited 2021 Feb 27].
Available from: https://www.rcaap.pt/detail.jsp?id=oai:run.unl.pt:10362/34464.
Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation
Council of Science Editors:
Madsen JH. Outlier detection for improved clustering : empirical research for unsupervised data mining. [Thesis]. Universidade Nova; 2018. Available from: https://www.rcaap.pt/detail.jsp?id=oai:run.unl.pt:10362/34464
Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation

University of Manchester
15.
Rostamniakankalhori, Sharareh.
Integrated Supervised and Unsupervised Learning Method to
Predict the Outcome of Tuberculosis Treatment Course.
Degree: 2011, University of Manchester
URL: http://www.manchester.ac.uk/escholar/uk-ac-man-scw:132404
► Tuberculosis (TB) is an infectious disease which is a global public health problem with over 9 million new cases annually. Tuberculosis treatment, with patient supervision…
(more)
▼ Tuberculosis (TB) is an infectious disease which is
a global public health problem with over 9 million new cases
annually. Tuberculosis treatment, with patient supervision and
support is an element of the global plan to stop TB designed by the
World Health Organization in 2006. The plan requires prediction of
patient treatment course destination. The prediction outcome can be
used to determine how intensive the level of supplying services and
supports in frame of DOTS therapy should be. No predictive model
for the outcome has been developed yet and only limited reports of
influential factors for considered outcome are available.To fill
this gap, this thesis develops a machine
learning approach to
predict the outcome of tuberculosis treatment course, which
includes, firstly, data of 6,450 Iranian TB patients under DOTS
(directly observed treatment, short course ) therapy were analysed
to initially diagnose the significant predictors by correlation
analysis; secondly, these significant features were applied to find
the best classification approach from six examined algorithms
including decision tree, Bayesian network, logistic regression,
multilayer perceptron, radial basis function, and support vector
machine; thirdly, the prediction accuracy of these existing
techniques was improved by proposing and developing a new
integrated method of k-mean clustering and classification
algorithms. Finally, a cluster-based simplified decision tree
(CSDT) was developed through an innovative hierarchical clustering
and classification algorithm. CSDT was built by k-mean partitioning
and the decision tree
learning. This innovative method not only
improves the prediction accuracy significantly but also leads to a
much simpler and interpretative decision tree.The main results of
this study included, firstly, finding seventeen significantly
correlated features which were: age, sex, weight, nationality, area
of residency, current stay in prison, low body weight, TB type,
treatment category, length of disease, TB case type, recent TB
infection, diabetic or HIV positive, and social risk factors like
history of imprisonment, IV drug usage, and unprotected sex ;
secondly, the results by applying and comparing six applied
supervised machine
learning tools on the testing set revealed that
decision trees gave the best prediction accuracy (74.21%) compared
with other methods; thirdly, by using testing set, the new
integrated approach to combine the clustering and classification
approach leads to the prediction accuracy improvement for all
applied classifiers; the most and least improvement for prediction
accuracy were shown by logistic regression (10%) and support vector
machine (4%) respectively. Finally, by applying the proposed and
developed CSDT, cluster-based simplified decision trees were
optioned, which reduced the size of the resulting decision tree and
further improved the prediction accuracy.Data type and having
normal distribution have created an opportunity for the decision
tree to outperform other algorithms. Pre-
learning by k-mean…
Advisors/Committee Members: Zeng, Xiaojun.
Subjects/Keywords: Integrated Supervised and Unsupervised Learning;
Tuberculosis; plediction
Record Details
Similar Records
Cite
Share »
Record Details
Similar Records
Cite
« Share





❌
APA ·
Chicago ·
MLA ·
Vancouver ·
CSE |
Export
to Zotero / EndNote / Reference
Manager
APA (6th Edition):
Rostamniakankalhori, S. (2011). Integrated Supervised and Unsupervised Learning Method to
Predict the Outcome of Tuberculosis Treatment Course. (Doctoral Dissertation). University of Manchester. Retrieved from http://www.manchester.ac.uk/escholar/uk-ac-man-scw:132404
Chicago Manual of Style (16th Edition):
Rostamniakankalhori, Sharareh. “Integrated Supervised and Unsupervised Learning Method to
Predict the Outcome of Tuberculosis Treatment Course.” 2011. Doctoral Dissertation, University of Manchester. Accessed February 27, 2021.
http://www.manchester.ac.uk/escholar/uk-ac-man-scw:132404.
MLA Handbook (7th Edition):
Rostamniakankalhori, Sharareh. “Integrated Supervised and Unsupervised Learning Method to
Predict the Outcome of Tuberculosis Treatment Course.” 2011. Web. 27 Feb 2021.
Vancouver:
Rostamniakankalhori S. Integrated Supervised and Unsupervised Learning Method to
Predict the Outcome of Tuberculosis Treatment Course. [Internet] [Doctoral dissertation]. University of Manchester; 2011. [cited 2021 Feb 27].
Available from: http://www.manchester.ac.uk/escholar/uk-ac-man-scw:132404.
Council of Science Editors:
Rostamniakankalhori S. Integrated Supervised and Unsupervised Learning Method to
Predict the Outcome of Tuberculosis Treatment Course. [Doctoral Dissertation]. University of Manchester; 2011. Available from: http://www.manchester.ac.uk/escholar/uk-ac-man-scw:132404

George Mason University
16.
Wang, Pu.
Nonparametric Bayesian Models for Unsupervised Learning
.
Degree: 2011, George Mason University
URL: http://hdl.handle.net/1920/6360
► Unsupervised learning is an important topic in machine learning. In particular, clustering is an unsupervised learning problem that arises in a variety of applications for…
(more)
▼ Unsupervised learning is an important topic in machine
learning. In particular, clustering
is an
unsupervised learning problem that arises in a variety of applications for data analysis
and mining. Unfortunately, clustering is an ill-posed problem and, as such, a challenging
one: no ground-truth that can be used to validate clustering results is available. Two issues
arise as a consequence. Various clustering algorithms embed their own bias resulting from
di erent optimization criteria. As a result, each algorithm may discover di erent patterns
in a given dataset. The second issue concerns the setting of parameters. In clustering,
parameter setting controls the characterization of individual clusters, and the total number
of clusters in the data.
Clustering ensembles have been proposed to address the issue of di erent biases induced
by various algorithms. Clustering ensembles combine di erent clustering results, and can
provide solutions that are robust against spurious elements in the data. Although clustering
ensembles provide a signi cant advance, they do not address satisfactorily the model selection
and the parameter tuning problem.
Bayesian approaches have been applied to clustering to address the parameter tuning
and model selection issues. Bayesian methods provide a principled way to address these
problems by assuming prior distributions on model parameters. Prior distributions assign
low probabilities to parameter values which are unlikely. Therefore they serve as regularizers
for modeling parameters, and can help avoid over- tting. In addition, the marginal likelihood
is used by Bayesian approaches as the criterion for model selection. Although Bayesian
methods provide a principled way to perform parameter tuning and model selection, the
key question \How many clusters?" is still open. This is a fundamental question for model
selection. A special kind of Bayesian methods, nonparametric Bayesian approaches, have
been proposed to address this important model selection issue. Unlike parametric Bayesian
models, for which the number of parameters is nite and xed, nonparametric Bayesian
models allow the number of parameters to grow with the number of observations. After
observing the data, nonparametric Bayesian models t the data with nite dimensional
parameters.
An additional issue with clustering is high dimensionality. High-dimensional data pose
a di cult challenge to the clustering process. A common scenario with high-dimensional
data is that clusters may exist in di erent subspaces comprised of di erent combinations of
features (dimensions). In other words, data points in a cluster may be similar to each other
along a subset of dimensions, but not in all dimensions. People have proposed subspace
clustering techniques, a.k.a. co-clustering or bi-clustering, to address the dimensionality
issue (here, I use the term co-clustering). Like clustering, also co-clustering su ers from the
ill-posed nature and the lack of ground-truth to validate the results.
Although attempts…
Advisors/Committee Members: Domeniconi, Carlotta (advisor).
Subjects/Keywords: Unsupervised Learning;
Clustering;
Bayesian Nonparametrics;
Clustering Ensembles
Record Details
Similar Records
Cite
Share »
Record Details
Similar Records
Cite
« Share





❌
APA ·
Chicago ·
MLA ·
Vancouver ·
CSE |
Export
to Zotero / EndNote / Reference
Manager
APA (6th Edition):
Wang, P. (2011). Nonparametric Bayesian Models for Unsupervised Learning
. (Thesis). George Mason University. Retrieved from http://hdl.handle.net/1920/6360
Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation
Chicago Manual of Style (16th Edition):
Wang, Pu. “Nonparametric Bayesian Models for Unsupervised Learning
.” 2011. Thesis, George Mason University. Accessed February 27, 2021.
http://hdl.handle.net/1920/6360.
Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation
MLA Handbook (7th Edition):
Wang, Pu. “Nonparametric Bayesian Models for Unsupervised Learning
.” 2011. Web. 27 Feb 2021.
Vancouver:
Wang P. Nonparametric Bayesian Models for Unsupervised Learning
. [Internet] [Thesis]. George Mason University; 2011. [cited 2021 Feb 27].
Available from: http://hdl.handle.net/1920/6360.
Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation
Council of Science Editors:
Wang P. Nonparametric Bayesian Models for Unsupervised Learning
. [Thesis]. George Mason University; 2011. Available from: http://hdl.handle.net/1920/6360
Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation
17.
ALAKKARI, SALAHEDDIN.
Modelling Large-scale Datasets Using Principal Component Analysis.
Degree: School of Computer Science & Statistics. Discipline of Computer Science, 2020, Trinity College Dublin
URL: http://hdl.handle.net/2262/92525
► Principal Component Analysis (PCA) is one of the most well-known unsupervised learning techniques used for dimensionality reduction and feature extraction. The main task of PCA…
(more)
▼ Principal Component Analysis (PCA) is one of the most well-known
unsupervised learning techniques used for dimensionality reduction and feature extraction. The main task of PCA is to compute a low-dimensional space that captures the maximal variability in the input dataset. Such a holistic linear representation is optimal in terms of the mean-squared-error. The basis vectors that form such a space correspond to the k most significant eigenvectors of the sample covariance matrix. However, computing such eigenvectors is computationally expensive with quadratic computational dependence on the data size. The ever-increasing size of datasets necessitates investigating reduced-complexity methods to find such eigenvectors.
A common treatment is to apply streaming PCA methods which aim to approximate the eigenvectors based on a single data pass with a linear computational cost. However, state-of-the-art streaming approaches are highly sequential and assume that samples are independent and identically distributed. In the first part of this thesis, we investigate the convergence of such methods when extended to the mini-batch mode, which is superior to the traditional fully online mode in terms of computation run-time. Furthermore, we propose an acceleration scheme for mini-batch streaming methods that are based on the Stochastic Gradient Approximation (SGA). Such methods provide the cheapest computational cost compared to other streaming algorithms. Based on empirical evaluation using the spiked covariance model and benchmark datasets, we show that applying our scheme significantly enhances the convergence of the original techniques in addition to outperforming other state-of-the-art methods.
In the second part, we investigate the performance of PCA when applied in a partitioning manner in which attributes are divided into a number of subsets, and then the standard approach is performed on each subset separately. We study two strategies for mapping attributes to different sets, namely, Cell-based PCA (CPCA) where samples are spatially divided into smaller blocks and Band-based PCA (BPCA) where attributes are partitioned based on their values distribution. We show that such models have several advantages over the holistic approach, including enhanced reconstruction quality and increased scalability. We also find that the baseline model, obtained when randomly mapping attributes, is analogous to the holistic PCA which entails a more practical and parallel alternative to streaming PCA paradigms. Not only are these methods beneficial for data compression but they also provide lightweight representations that would enhance the accuracy and training time of deep
learning models.
Time-varying datasets of various physical observations are also addressed. We theoretically draw the analogy between many analytic physical models and the PCA eigenvalue problem. It is shown that, for a wide range of physical phenomena, the eigenvectors derived using PCA are analogous to the analytic physical model. Since time-varying datasets are no…
Advisors/Committee Members: Dingliana, John.
Subjects/Keywords: dimensionality reduction; unsupervised machine learning; data analysis
Record Details
Similar Records
Cite
Share »
Record Details
Similar Records
Cite
« Share





❌
APA ·
Chicago ·
MLA ·
Vancouver ·
CSE |
Export
to Zotero / EndNote / Reference
Manager
APA (6th Edition):
ALAKKARI, S. (2020). Modelling Large-scale Datasets Using Principal Component Analysis. (Thesis). Trinity College Dublin. Retrieved from http://hdl.handle.net/2262/92525
Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation
Chicago Manual of Style (16th Edition):
ALAKKARI, SALAHEDDIN. “Modelling Large-scale Datasets Using Principal Component Analysis.” 2020. Thesis, Trinity College Dublin. Accessed February 27, 2021.
http://hdl.handle.net/2262/92525.
Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation
MLA Handbook (7th Edition):
ALAKKARI, SALAHEDDIN. “Modelling Large-scale Datasets Using Principal Component Analysis.” 2020. Web. 27 Feb 2021.
Vancouver:
ALAKKARI S. Modelling Large-scale Datasets Using Principal Component Analysis. [Internet] [Thesis]. Trinity College Dublin; 2020. [cited 2021 Feb 27].
Available from: http://hdl.handle.net/2262/92525.
Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation
Council of Science Editors:
ALAKKARI S. Modelling Large-scale Datasets Using Principal Component Analysis. [Thesis]. Trinity College Dublin; 2020. Available from: http://hdl.handle.net/2262/92525
Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation

University of Toronto
18.
Miasnikof, Pierre.
Subgraph Density and Graph Clustering.
Degree: PhD, 2019, University of Toronto
URL: http://hdl.handle.net/1807/97609
► Graph clustering, also often referred to as network community detection, is an unsupervised learning task. It is the process of grouping vertices into sets of…
(more)
▼ Graph clustering, also often referred to as network community detection, is an
unsupervised learning task. It is the process of grouping vertices into sets of vertices that share common relations. Graph clustering is a pivotal topic in the field of network science. It has applications in various areas such as the analysis of chemical, biological and social networks and any other area where a macroscopic systems point of view is important.
Currently, three open problems emerge from the literature: determining if a graph is clusterable before any clustering effort is undertaken, assessing the quality of clusterings identified by an algorithm and the development of fast accurate clustering techniques. In this dissertation, we attempt to answer these three open questions. We use heterogeneity in local densities to determine whether a graph is suited for clustering, if it has a clustered structure to begin with. We also compare global, mean intra-cluster and mean inter-cluster densities to assess the quality and strength of the clusters identified by any clustering algorithm. Finally, as a minor contribution, we examine the possibility of using intra-cluster density maximization as a clustering algorithm.
Our clusterability test is shown to be more accurate than techniques proposed in the very recent literature and to rely on fewer assumptions. We demonstrate how heterogeneous densities can be used to identify graphs that display a clustered structure.
On the topic of clustering quality assessment, we offer an alternative to the very widely used and problematic modularity quality function. This alternative is consistent with the sets of axioms defining good clustering quality functions, which have been presented by numerous authors. We base our measures on our own definition of clustering quality and accompanying necessary and sufficient conditions. Our approach is shown to be more responsive to connectivity patterns than both modularity and conductance measures.
Finally, we also explore intra-cluster density as an objective function to be maximized, for the purpose of clustering. We show how an ensemble optimization meta-heuristic may offer a solution to the clustering problem.
Advisors/Committee Members: Lawryshyn, Yuri, Pardalos, Panos M., Chemical Engineering Applied Chemistry.
Subjects/Keywords: Data Science; Graph Clustering; Unsupervised Learning; 0463
Record Details
Similar Records
Cite
Share »
Record Details
Similar Records
Cite
« Share





❌
APA ·
Chicago ·
MLA ·
Vancouver ·
CSE |
Export
to Zotero / EndNote / Reference
Manager
APA (6th Edition):
Miasnikof, P. (2019). Subgraph Density and Graph Clustering. (Doctoral Dissertation). University of Toronto. Retrieved from http://hdl.handle.net/1807/97609
Chicago Manual of Style (16th Edition):
Miasnikof, Pierre. “Subgraph Density and Graph Clustering.” 2019. Doctoral Dissertation, University of Toronto. Accessed February 27, 2021.
http://hdl.handle.net/1807/97609.
MLA Handbook (7th Edition):
Miasnikof, Pierre. “Subgraph Density and Graph Clustering.” 2019. Web. 27 Feb 2021.
Vancouver:
Miasnikof P. Subgraph Density and Graph Clustering. [Internet] [Doctoral dissertation]. University of Toronto; 2019. [cited 2021 Feb 27].
Available from: http://hdl.handle.net/1807/97609.
Council of Science Editors:
Miasnikof P. Subgraph Density and Graph Clustering. [Doctoral Dissertation]. University of Toronto; 2019. Available from: http://hdl.handle.net/1807/97609

University of Minnesota
19.
Traganitis, Panagiotis.
Scalable and Ensemble Learning for Big Data.
Degree: PhD, Electrical/Computer Engineering, 2019, University of Minnesota
URL: http://hdl.handle.net/11299/206358
► The turn of the decade has trademarked society and computing research with a ``data deluge.'' As the number of smart, highly accurate and Internet-capable devices…
(more)
▼ The turn of the decade has trademarked society and computing research with a ``data deluge.'' As the number of smart, highly accurate and Internet-capable devices increases, so does the amount of data that is generated and collected. While this sheer amount of data has the potential to enable high quality inference, and mining of information, it introduces numerous challenges in the processing and pattern analysis, since available statistical inference and machine learning approaches do not necessarily scale well with the number of data and their dimensionality. In addition to the challenges related to scalability, data gathered are often noisy, dynamic, contaminated by outliers or corrupted to specifically inhibit the inference task. Moreover, many machine learning approaches have been shown to be susceptible to adversarial attacks. At the same time, the cost of cloud and distributed computing is rapidly declining. Therefore, there is a pressing need for statistical inference and machine learning tools that are robust to attacks and scale with the volume and dimensionality of the data, by harnessing efficiently the available computational resources. This thesis is centered on analytical and algorithmic foundations that aim to enable statistical inference and data analytics from large volumes of high-dimensional data. The vision is to establish a comprehensive framework based on state-of-the-art machine learning, optimization and statistical inference tools to enable truly large-scale inference, which can tap on the available (possibly distributed) computational resources, and be resilient to adversarial attacks. The ultimate goal is to both analytically and numerically demonstrate how valuable insights from signal processing can lead to markedly improved and accelerated learning tools. To this end, the present thesis investigates two main research thrusts: i) Large-scale subspace clustering; and ii) unsupervised ensemble learning. The aforementioned research thrusts introduce novel algorithms that aim to tackle the issues of large-scale learning. The potential of the proposed algorithms is showcased by rigorous theoretical results and extensive numerical tests.
Subjects/Keywords: Big Data; clustering; Ensemble; learning; subspace; unsupervised
Record Details
Similar Records
Cite
Share »
Record Details
Similar Records
Cite
« Share





❌
APA ·
Chicago ·
MLA ·
Vancouver ·
CSE |
Export
to Zotero / EndNote / Reference
Manager
APA (6th Edition):
Traganitis, P. (2019). Scalable and Ensemble Learning for Big Data. (Doctoral Dissertation). University of Minnesota. Retrieved from http://hdl.handle.net/11299/206358
Chicago Manual of Style (16th Edition):
Traganitis, Panagiotis. “Scalable and Ensemble Learning for Big Data.” 2019. Doctoral Dissertation, University of Minnesota. Accessed February 27, 2021.
http://hdl.handle.net/11299/206358.
MLA Handbook (7th Edition):
Traganitis, Panagiotis. “Scalable and Ensemble Learning for Big Data.” 2019. Web. 27 Feb 2021.
Vancouver:
Traganitis P. Scalable and Ensemble Learning for Big Data. [Internet] [Doctoral dissertation]. University of Minnesota; 2019. [cited 2021 Feb 27].
Available from: http://hdl.handle.net/11299/206358.
Council of Science Editors:
Traganitis P. Scalable and Ensemble Learning for Big Data. [Doctoral Dissertation]. University of Minnesota; 2019. Available from: http://hdl.handle.net/11299/206358

Texas Tech University
20.
Hill, Jason E.
Application of Information Theoretic Unsupervised Learning to Medical Image Analysis.
Degree: 2013, Texas Tech University
URL: http://hdl.handle.net/2346/48865
► Automated segmentation of medical images is a challenging problem. The number of segments in a medical image may be unknown a priori, due to the…
(more)
▼ Automated segmentation of medical images is a challenging problem. The number of segments in a medical image may be unknown a priori, due to the presence or absence of pathological anomalies. Some
unsupervised learning techniques that take advantage of information theory concepts may provide a solid approach to the solution of this problem. To this end, there has been the recent development of the Improved “Jump” Method (IJM), a technique that efficiently finds a suitable number of clusters representing different tissue characteristics in a medical image. The IJM works by optimizing an objective function, the margin, that quantifies the quality of particular cluster configurations. Recent developments involving interesting relationships between Spectral Clustering (SC) and kernel Principal Component Analysis (kPCA) are used by the implementation of the IJM to cover the non-linear domain. In this novel SC approach the data is mapped to a new space where the points belonging to the same cluster are collinear if the parameters of a Radial Basis Function (RBF) kernel are adequately selected. After projecting these points onto the unit sphere, IJM measures the quality of different cluster configurations, yielding an algorithm that simultaneously selects the number of clusters, and the RBF kernel parameter.
Validation of this method is sought via segmentation of MR brain images in a combination of all major modalities. Such labeled MRI datasets serve as benchmarks for any segmentation algorithm. The effectiveness of the nonlinear IJM is demonstrated in the segmentation of uterine cervix color images for early identification of cervical neoplasia, as an aid to cervical cancer diagnosis. Limitations of the current implementation of IJM are encountered when attempting to segment and MR brain images with multiple sclerosis (MS) lesions. These limitations and a strategy to overcome them are discussed. Finally, an outlook to applying this method to the segmentation of cells in Pap smear test micrographs is laid out.
Advisors/Committee Members: Mitra, Sunanda (Committee Chair), Nutter, Brian (committee member).
Subjects/Keywords: Unsupervised learning; Medical images; Spectral clustering.
Record Details
Similar Records
Cite
Share »
Record Details
Similar Records
Cite
« Share





❌
APA ·
Chicago ·
MLA ·
Vancouver ·
CSE |
Export
to Zotero / EndNote / Reference
Manager
APA (6th Edition):
Hill, J. E. (2013). Application of Information Theoretic Unsupervised Learning to Medical Image Analysis. (Thesis). Texas Tech University. Retrieved from http://hdl.handle.net/2346/48865
Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation
Chicago Manual of Style (16th Edition):
Hill, Jason E. “Application of Information Theoretic Unsupervised Learning to Medical Image Analysis.” 2013. Thesis, Texas Tech University. Accessed February 27, 2021.
http://hdl.handle.net/2346/48865.
Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation
MLA Handbook (7th Edition):
Hill, Jason E. “Application of Information Theoretic Unsupervised Learning to Medical Image Analysis.” 2013. Web. 27 Feb 2021.
Vancouver:
Hill JE. Application of Information Theoretic Unsupervised Learning to Medical Image Analysis. [Internet] [Thesis]. Texas Tech University; 2013. [cited 2021 Feb 27].
Available from: http://hdl.handle.net/2346/48865.
Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation
Council of Science Editors:
Hill JE. Application of Information Theoretic Unsupervised Learning to Medical Image Analysis. [Thesis]. Texas Tech University; 2013. Available from: http://hdl.handle.net/2346/48865
Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation

University of Southern California
21.
Deutsch, Shay.
Learning the geometric structure of high dimensional data
using the Tensor Voting Graph.
Degree: PhD, Computer Science, 2017, University of Southern California
URL: http://digitallibrary.usc.edu/cdm/compoundobject/collection/p15799coll3/id/640839/rec/3793
► This study addresses a range of fundamental problems in unsupervised manifold learning. Given a set of noisy points in a high dimensional space that lie…
(more)
▼ This study addresses a range of fundamental problems
in
unsupervised manifold
learning. Given a set of noisy points in a
high dimensional space that lie near one or more possibly
intersecting smooth manifolds, different challenges include
learning the local geometric structure at each point, geodesic
distance estimation, and clustering. These challenges are
ubiquitous in
unsupervised manifold
learning, and many applications
in computer vision as well as other scientific applications would
benefit from a principled approach to these problems. ❧ In the
first part of this thesis we present a hybrid local-global method
that leverages the algorithmic capabilities of the Tensor Voting
framework. However, unlike Tensor Voting, which can learn complex
structures reliably only locally, our method is capable of reliably
inferring the global structure of complex manifolds using a unique
graph construction called the Tensor Voting Graph (TVG). This graph
provides an efficient tool to perform the desired global manifold
learning tasks such as geodesic distance estimation and clustering
on complex manifolds, thus overcoming one of one of the main
limitations of Tensor Voting as a strictly local approach.
Moreover, we propose to explicitly and directly resolve the
ambiguities near the intersections with a novel algorithm, which
uses the TVG and the positions of the points near the manifold
intersections. ❧ In the second part of this thesis we propose a new
framework for manifold denoising based processing in the graph
Fourier frequency domain, derived from the spectral decomposition
of the discrete graph Laplacian. The suggested approach, called
MFD, uses the Spectral Graph Wavelet transform in order to perform
non-iterative denoising directly in the graph frequency domain. To
the best of our knowledge, MFD is the first attempt to use graph
signal processing [55] tools for manifold denoising on unstructured
domains. We provide theoretical justification for our Manifold
Frequency Denoising approach on unstructured graphs and demonstrate
that for smooth manifolds the coordinate signals also exhibit
smoothness. This is first demonstrated in the case of noiseless
observations, by proving that manifolds with smoother
characteristics create more energy in the lower frequencies.
Moreover, it is shown that higher frequency wavelet coefficients
decay in a way that depends on the smoothness properties of the
manifold, which is explicitly tied to the curvature properties. We
then provide an analysis for the case of noisy points and a noisy
graph, establishing results which tie the noisy graph Laplacian to
the noiseless graph Laplacian characteristics that are induced by
the smoothness properties of the manifold. The suggested MFD
framework holds attractive features such as robustness to k nearest
neighbors parameter selection on the graph, and it is
computationally efficient. ❧ Finally, the last part of this
research merges the Manifold Frequency Denoising and the Tensor
Voting Graph methods into a uniform framework, which allows us to…
Advisors/Committee Members: Medioni, Gerard (Committee Chair), Ortega, Antonio (Committee Member), Nakano, Aiichiro (Committee Member).
Subjects/Keywords: manifold learning; unsupervised denoising; Tensor Voting Graph
Record Details
Similar Records
Cite
Share »
Record Details
Similar Records
Cite
« Share





❌
APA ·
Chicago ·
MLA ·
Vancouver ·
CSE |
Export
to Zotero / EndNote / Reference
Manager
APA (6th Edition):
Deutsch, S. (2017). Learning the geometric structure of high dimensional data
using the Tensor Voting Graph. (Doctoral Dissertation). University of Southern California. Retrieved from http://digitallibrary.usc.edu/cdm/compoundobject/collection/p15799coll3/id/640839/rec/3793
Chicago Manual of Style (16th Edition):
Deutsch, Shay. “Learning the geometric structure of high dimensional data
using the Tensor Voting Graph.” 2017. Doctoral Dissertation, University of Southern California. Accessed February 27, 2021.
http://digitallibrary.usc.edu/cdm/compoundobject/collection/p15799coll3/id/640839/rec/3793.
MLA Handbook (7th Edition):
Deutsch, Shay. “Learning the geometric structure of high dimensional data
using the Tensor Voting Graph.” 2017. Web. 27 Feb 2021.
Vancouver:
Deutsch S. Learning the geometric structure of high dimensional data
using the Tensor Voting Graph. [Internet] [Doctoral dissertation]. University of Southern California; 2017. [cited 2021 Feb 27].
Available from: http://digitallibrary.usc.edu/cdm/compoundobject/collection/p15799coll3/id/640839/rec/3793.
Council of Science Editors:
Deutsch S. Learning the geometric structure of high dimensional data
using the Tensor Voting Graph. [Doctoral Dissertation]. University of Southern California; 2017. Available from: http://digitallibrary.usc.edu/cdm/compoundobject/collection/p15799coll3/id/640839/rec/3793

University of California – Merced
22.
Vladymyrov, Maksym.
Large-Scale Methods for Nonlinear Manifold Learning.
Degree: Electrical Engineering and Computer Science, 2014, University of California – Merced
URL: http://www.escholarship.org/uc/item/9hj5v8z2
► High-dimensional data representation is an important problem in many different areas of science. Nowadays, it is becoming crucial to interpret the data of varying dimensionality…
(more)
▼ High-dimensional data representation is an important problem in many different areas of science. Nowadays, it is becoming crucial to interpret the data of varying dimensionality correctly. Dimensionality reduction methods process the data in order to help visualize the data, reduce its complexity, or find latent representation of the original problem. The algorithms of nonlinear dimensionality reduction (also known as manifold learning) are used to decrease the dimensionality of the problem while preserving the general structure of the data. Both spectral methods (such as Laplacian Eigenmaps or ISOMAP) and nonlinear embedding algorithms (NLE, such as t-SNE or Elastic Embedding) have shown to provide very good nonlinear embedding of high-dimensional data sets. However, those methods are notorious for very slow optimization, practically preventing them from being used when a data set is bigger than few thousand points. In my thesis we investigate several techniques to improve different stages of nonlinear dimensionally algorithms. First, we analyze the entropic affinities as a better way to build a similarity matrix. We explore its properties and propose a nearly-optimal algorithm to construct them. Second, we present a novel faster method to optimize NLE by using second-order information during the optimization. Third, for spectral methods, we investigate landmark-based optimization that cleverly substitutes original large-scale problem with a much smaller easy-to-solve subproblem. Finally, we apply Fast Multipole Methods approximation that allows fast computation of the gradient and the objective function of NLE and reduces their computational complexity from O(N2) to O(N). Each of the proposed methods accelerate the optimization dramatically by one or two orders of magnitude compared to the existing techniques, effectively allowing corresponding methods to run on a dataset with millions of points.
Subjects/Keywords: Computer science; dimensionality reduction; machine learning; manifold learning; unsupervised learning
Record Details
Similar Records
Cite
Share »
Record Details
Similar Records
Cite
« Share





❌
APA ·
Chicago ·
MLA ·
Vancouver ·
CSE |
Export
to Zotero / EndNote / Reference
Manager
APA (6th Edition):
Vladymyrov, M. (2014). Large-Scale Methods for Nonlinear Manifold Learning. (Thesis). University of California – Merced. Retrieved from http://www.escholarship.org/uc/item/9hj5v8z2
Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation
Chicago Manual of Style (16th Edition):
Vladymyrov, Maksym. “Large-Scale Methods for Nonlinear Manifold Learning.” 2014. Thesis, University of California – Merced. Accessed February 27, 2021.
http://www.escholarship.org/uc/item/9hj5v8z2.
Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation
MLA Handbook (7th Edition):
Vladymyrov, Maksym. “Large-Scale Methods for Nonlinear Manifold Learning.” 2014. Web. 27 Feb 2021.
Vancouver:
Vladymyrov M. Large-Scale Methods for Nonlinear Manifold Learning. [Internet] [Thesis]. University of California – Merced; 2014. [cited 2021 Feb 27].
Available from: http://www.escholarship.org/uc/item/9hj5v8z2.
Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation
Council of Science Editors:
Vladymyrov M. Large-Scale Methods for Nonlinear Manifold Learning. [Thesis]. University of California – Merced; 2014. Available from: http://www.escholarship.org/uc/item/9hj5v8z2
Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation

Penn State University
23.
Shen, Sida.
Sepsis Data Analytics.
Degree: 2020, Penn State University
URL: https://submit-etda.libraries.psu.edu/catalog/17780sps5425
► Sepsis is a potentially life-threatening condition caused by the body’s response to infection. Body releases chemicals into the blood stream to fight infection. However, sepsis…
(more)
▼ Sepsis is a potentially life-threatening condition caused by the body’s response to infection.
Body releases chemicals into the blood stream to fight infection. However, sepsis occurs
when the body’s response to the chemicals go out of balance. Despite the use of antibiotics
and modern treatments, sepsis is still one of the main causes of ICU mortality rate.
The current broad definition of sepsis is not suitable for the heterogeneous nature of
this disease; it is necessary to discover novel phenotypes of sepsis and design custom
treatment plans. In this thesis, two novel phenotype discovery methods have been
successfully developed and tested on MIMIC-III database. The first method utilizes
first lab result for each patient, after feature imputation to resolve missing values, 11
features are included (heart rate, respiratory rate, systolic blood pressure noninvasive
(sbp-noninvasive), temperature, sodium, white blood cells (WBC), creatinine, glucose
and all 3 scores on the Glasgow Coma Scale (GCS)). With dimensionality reduction using
Principal Component Analysis and clustering using K-means algorithm, three phenotypes
are discovered; the first group patients (population: 44.9%, mortality: 14.88%) have high
possibility of respiratory and renal failures; the second group patients (population: 23.8%,
mortality: 9.15%) have high possibility of liver and coagulation failures; the third group
patients (population: 31.4%, mortality: 20.9%) have high possibility of cardiovascular
and central nervous system (CNS) failures. In the second model, we adapted deep embedding
clustering to cluster sepsis patient into novel phenotypes. We included 7 measurements (heart rate,
respiratory rate, hemoglobin, white blood cell, creatinine, glucose and sodium) combined
with 12 time-steps with 4-hour intervals (48 hours span). For each patient a sample with
84 features is constructed. A multi-layer fully connected auto-encoder is trained with
20 latent units; after 300 epochs, auto-encoder reconstruction loss (mean square error)
converges. The encoder and soft assignment clustering layer are trained jointly using
stochastic gradient descent; the auxiliary target distribution is updated every 200 steps.
The network converges after 7800 steps with Kullback-Leibler divergence loss of 0.028.
The derived 4 phenotypes present a clearly separated patient outcome with mortality
standard deviation of 4.97%. However, by comparing bio-markers’ statistics with patient
outcome across the derived phenotypes, we can not see reasonable pattern that connect
the two. There are some high dimensional features that the deep clustering model has
captured, which we believe can lead to the discovery of the true cause of sepsis mortality.
Advisors/Committee Members: Soundar Kumara, Thesis Advisor/Co-Advisor, Robert Carl Voigt, Program Head/Chair, Kamesh Madduri, Committee Member.
Subjects/Keywords: Machine learning; Deep learning; Sepsis; Healthcare; Unsupervised Learning; Data Mining
Record Details
Similar Records
Cite
Share »
Record Details
Similar Records
Cite
« Share





❌
APA ·
Chicago ·
MLA ·
Vancouver ·
CSE |
Export
to Zotero / EndNote / Reference
Manager
APA (6th Edition):
Shen, S. (2020). Sepsis Data Analytics. (Thesis). Penn State University. Retrieved from https://submit-etda.libraries.psu.edu/catalog/17780sps5425
Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation
Chicago Manual of Style (16th Edition):
Shen, Sida. “Sepsis Data Analytics.” 2020. Thesis, Penn State University. Accessed February 27, 2021.
https://submit-etda.libraries.psu.edu/catalog/17780sps5425.
Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation
MLA Handbook (7th Edition):
Shen, Sida. “Sepsis Data Analytics.” 2020. Web. 27 Feb 2021.
Vancouver:
Shen S. Sepsis Data Analytics. [Internet] [Thesis]. Penn State University; 2020. [cited 2021 Feb 27].
Available from: https://submit-etda.libraries.psu.edu/catalog/17780sps5425.
Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation
Council of Science Editors:
Shen S. Sepsis Data Analytics. [Thesis]. Penn State University; 2020. Available from: https://submit-etda.libraries.psu.edu/catalog/17780sps5425
Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation

University of Waterloo
24.
Jaini, Priyank.
Likelihood-based Density Estimation using Deep Architectures.
Degree: 2019, University of Waterloo
URL: http://hdl.handle.net/10012/15356
► Multivariate density estimation is a central problem in unsupervised machine learning that has been studied immensely in both statistics and machine learning. Several methods have…
(more)
▼ Multivariate density estimation is a central problem in unsupervised machine learning that has been studied immensely in both statistics and machine learning. Several methods have thus been proposed for density estimation including classical techniques like histograms, kernel density estimation methods, mixture models, and more recently neural density estimation that leverages the recent advances in deep learning and neural networks to tractably represent a density function. In today's age, when large amounts of data are being generated in almost every field, it is of paramount importance to develop density estimation methods that are cheap both computationally and in memory cost. The main contribution of this thesis is in providing a principled study of parametric density estimation methods using mixture models and triangular maps for neural density estimation.
The first part of the thesis focuses on the compact representation of mixture models using deep architectures like latent tree models, hidden Markov models, tensorial mixture models, hierarchical tensor formats and sum-product networks. It provides a unifying view of possible representations of mixture models using such deep architectures. The unifying view allows us to prove exponential separation between deep mixture models and mixture models represented using shallow architectures, demonstrating the benefits of depth in their representation. In a surprising result thereafter, we prove that a deep mixture model can be approximated using the conditional gradient algorithm by a shallow architecture of polynomial size w.r.t. the inverse of the approximation accuracy.
Next, we address the more practical problem of density estimation of mixture models for streaming data by proposing an online Bayesian Moment Matching algorithm for Gaussian mixture models that can be distributed over several processors for fast computation. Exact Bayesian learning of mixture models is intractable because the number of terms in the posterior grows exponentially w.r.t. to the number of observations. We circumvent this problem by projecting the exact posterior on to a simple family of densities by matching a set of sufficient moments. Subsequently, we extend this algorithm for sequential data modeling using transfer learning by learning a hidden Markov model over the observations with Gaussian mixtures. We apply this algorithm on three diverse applications of activity recognition based on smartphone sensors, sleep stage classification for predicting neurological disorders using electroencephalography data and network size prediction for telecommunication networks.
In the second part, we focus on neural density estimation methods where we provide a unified framework for estimating densities using monotone and bijective triangular maps represented using deep neural networks. Using this unified framework we study the limitations and representation power of recent flow based and autoregressive methods. Based on this framework, we subsequently propose a novel Sum-of-Squares…
Subjects/Keywords: machine learning; unsupervised learning; deep learning; probabilitic graphical models
Record Details
Similar Records
Cite
Share »
Record Details
Similar Records
Cite
« Share





❌
APA ·
Chicago ·
MLA ·
Vancouver ·
CSE |
Export
to Zotero / EndNote / Reference
Manager
APA (6th Edition):
Jaini, P. (2019). Likelihood-based Density Estimation using Deep Architectures. (Thesis). University of Waterloo. Retrieved from http://hdl.handle.net/10012/15356
Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation
Chicago Manual of Style (16th Edition):
Jaini, Priyank. “Likelihood-based Density Estimation using Deep Architectures.” 2019. Thesis, University of Waterloo. Accessed February 27, 2021.
http://hdl.handle.net/10012/15356.
Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation
MLA Handbook (7th Edition):
Jaini, Priyank. “Likelihood-based Density Estimation using Deep Architectures.” 2019. Web. 27 Feb 2021.
Vancouver:
Jaini P. Likelihood-based Density Estimation using Deep Architectures. [Internet] [Thesis]. University of Waterloo; 2019. [cited 2021 Feb 27].
Available from: http://hdl.handle.net/10012/15356.
Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation
Council of Science Editors:
Jaini P. Likelihood-based Density Estimation using Deep Architectures. [Thesis]. University of Waterloo; 2019. Available from: http://hdl.handle.net/10012/15356
Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation

Linköping University
25.
Alirezaie, Marjan.
Semantic Analysis Of Multi Meaning Words Using Machine Learning And Knowledge Representation.
Degree: Computer and Information Science, 2011, Linköping University
URL: http://urn.kb.se/resolve?urn=urn:nbn:se:liu:diva-70086
► The present thesis addresses machine learning in a domain of naturallanguage phrases that are names of universities. It describes two approaches to this problem…
(more)
▼ The present thesis addresses machine learning in a domain of naturallanguage phrases that are names of universities. It describes two approaches to this problem and a software implementation that has made it possible to evaluate them and to compare them.
In general terms, the system's task is to learn to 'understand' the significance of the various components of a university name, such as the city or region where the university is located, the scienti c disciplines that are studied there, or the name of a famous person which may be part of the university name. A concrete test for whether the system has acquired this understanding is when it is able to compose a plausible university name given some components that should occur in the name.
In order to achieve this capability, our system learns the structure of available names of some universities in a given data set, i.e. it acquires a grammar for the microlanguage of university names. One of the challenges is that the system may encounter ambiguities due to multi meaning words. This problem is addressed using a small ontology that is created during the training phase.
Both domain knowledge and grammatical knowledge is represented using decision trees, which is an ecient method for concept learning. Besides for inductive inference, their role is to partition the data set into a hierarchical structure which is used for resolving ambiguities.
The present report also de nes some modi cations in the de nitions of parameters, for example a parameter for entropy, which enable the system to deal with cognitive uncertainties. Our method for automatic syntax acquisition, ADIOS, is an unsupervised learning method. This method is described and discussed here, including a report on the outcome of the tests using our data set.
The software that has been implemented and used in this project has been implemented in C.
Subjects/Keywords: Machine Learning; Supervised Learning; Unsupervised Learning; Computer Sciences; Datavetenskap (datalogi)
Record Details
Similar Records
Cite
Share »
Record Details
Similar Records
Cite
« Share





❌
APA ·
Chicago ·
MLA ·
Vancouver ·
CSE |
Export
to Zotero / EndNote / Reference
Manager
APA (6th Edition):
Alirezaie, M. (2011). Semantic Analysis Of Multi Meaning Words Using Machine Learning And Knowledge Representation. (Thesis). Linköping University. Retrieved from http://urn.kb.se/resolve?urn=urn:nbn:se:liu:diva-70086
Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation
Chicago Manual of Style (16th Edition):
Alirezaie, Marjan. “Semantic Analysis Of Multi Meaning Words Using Machine Learning And Knowledge Representation.” 2011. Thesis, Linköping University. Accessed February 27, 2021.
http://urn.kb.se/resolve?urn=urn:nbn:se:liu:diva-70086.
Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation
MLA Handbook (7th Edition):
Alirezaie, Marjan. “Semantic Analysis Of Multi Meaning Words Using Machine Learning And Knowledge Representation.” 2011. Web. 27 Feb 2021.
Vancouver:
Alirezaie M. Semantic Analysis Of Multi Meaning Words Using Machine Learning And Knowledge Representation. [Internet] [Thesis]. Linköping University; 2011. [cited 2021 Feb 27].
Available from: http://urn.kb.se/resolve?urn=urn:nbn:se:liu:diva-70086.
Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation
Council of Science Editors:
Alirezaie M. Semantic Analysis Of Multi Meaning Words Using Machine Learning And Knowledge Representation. [Thesis]. Linköping University; 2011. Available from: http://urn.kb.se/resolve?urn=urn:nbn:se:liu:diva-70086
Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation

University of Toronto
26.
Srivastava, Nitish.
Deep Learning Models for Unsupervised and Transfer Learning.
Degree: PhD, 2017, University of Toronto
URL: http://hdl.handle.net/1807/80672
► This thesis is a compilation of five research contributions whose goal is to do unsupervised and transfer learning by designing models that learn distributed representations…
(more)
▼ This thesis is a compilation of five research contributions whose goal is to do
unsupervised and transfer
learning by designing models that learn distributed representations using deep neural networks. First, we describe a Deep Boltzmann Machine model applied to image-text and audio-video multi-modal data. We show that the learned generative probabilistic model can jointly model both modalities and also produce good conditional distributions on each modality given the other. We use this model to infer fused high-level representations and evaluate them using retrieval and classification tasks.
Second, we propose a Boltzmann Machine based topic model for modeling bag-of-words documents. This model augments the Replicated Softmax Model with a second hidden layer of latent words without sacrificing RBM-like inference and training. We describe how this can be viewed as a beneficial modification of the otherwise rigid, complementary prior that is implicit in RBM-like models.
Third, we describe an RNN-based encoder-decoder model that learns to represent video sequences. This model is inspired by sequence-to-sequence
learning for machine translation. We train an RNN encoder to come up with a representation of the input sequence that can be used to both decode the input back, and predict the future sequence. This representation is evaluated using action recognition benchmarks.
Fourth, we develop a theory of directional units and use them to construct Boltzmann Machines and Autoencoders. A directional unit is a structured, vector-valued hidden unit which represents a continuous space of features. The magnitude and direction of a directional unit represent the strength and pose of a feature within this space, respectively. Networks of these units can potentially do better coincidence detection and learn general equivariance classes. Temporal coherence based
learning can be used with these units to factor out the dynamic properties of a feature, part, or object from static properties such as identity.
Last, we describe a contribution to transfer
learning. We show how a deep convolutional net trained to classify among a given set of categories can transfer its knowledge to new categories even when very few labelled examples are available for the new categories.
Advisors/Committee Members: Hinton, Geoffrey E, Salakhutdinov, Ruslan R, Computer Science.
Subjects/Keywords: Boltzmann Machines; Deep Learning; Machine Learning; Neural Networks; Unsupervised Learning; 0984
Record Details
Similar Records
Cite
Share »
Record Details
Similar Records
Cite
« Share





❌
APA ·
Chicago ·
MLA ·
Vancouver ·
CSE |
Export
to Zotero / EndNote / Reference
Manager
APA (6th Edition):
Srivastava, N. (2017). Deep Learning Models for Unsupervised and Transfer Learning. (Doctoral Dissertation). University of Toronto. Retrieved from http://hdl.handle.net/1807/80672
Chicago Manual of Style (16th Edition):
Srivastava, Nitish. “Deep Learning Models for Unsupervised and Transfer Learning.” 2017. Doctoral Dissertation, University of Toronto. Accessed February 27, 2021.
http://hdl.handle.net/1807/80672.
MLA Handbook (7th Edition):
Srivastava, Nitish. “Deep Learning Models for Unsupervised and Transfer Learning.” 2017. Web. 27 Feb 2021.
Vancouver:
Srivastava N. Deep Learning Models for Unsupervised and Transfer Learning. [Internet] [Doctoral dissertation]. University of Toronto; 2017. [cited 2021 Feb 27].
Available from: http://hdl.handle.net/1807/80672.
Council of Science Editors:
Srivastava N. Deep Learning Models for Unsupervised and Transfer Learning. [Doctoral Dissertation]. University of Toronto; 2017. Available from: http://hdl.handle.net/1807/80672
27.
STUART, CHRISTOFFER.
Applying Machine Learning to Identify Maintenance Level for Software Releases
.
Degree: Chalmers tekniska högskola / Institutionen för data och informationsteknik, 2020, Chalmers University of Technology
URL: http://hdl.handle.net/20.500.12380/300701
► Maintenance is the single largest cost in software development. Therefore it is important to understand what causes maintenance, and if it can be predicted. Many…
(more)
▼ Maintenance is the single largest cost in software development. Therefore it is important
to understand what causes maintenance, and if it can be predicted. Many
studies have shown that certain ways of measuring the complexity of developed
programs can create decent prediction models to determine the likelihood of maintenance
due to failures in the software. Most have been prior to release and often
requires specific, object-oriented, metrics of the software to set up the models. These
metrics are not always available in the software development companies. This study
determines that cumulative software failure levels after release can be determined
using available data at a software development company and machine learning algorithms.
Subjects/Keywords: Machine learning;
supervised learning;
unsupervised learning;
defect prediction;
cumulative failure prediction
Record Details
Similar Records
Cite
Share »
Record Details
Similar Records
Cite
« Share





❌
APA ·
Chicago ·
MLA ·
Vancouver ·
CSE |
Export
to Zotero / EndNote / Reference
Manager
APA (6th Edition):
STUART, C. (2020). Applying Machine Learning to Identify Maintenance Level for Software Releases
. (Thesis). Chalmers University of Technology. Retrieved from http://hdl.handle.net/20.500.12380/300701
Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation
Chicago Manual of Style (16th Edition):
STUART, CHRISTOFFER. “Applying Machine Learning to Identify Maintenance Level for Software Releases
.” 2020. Thesis, Chalmers University of Technology. Accessed February 27, 2021.
http://hdl.handle.net/20.500.12380/300701.
Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation
MLA Handbook (7th Edition):
STUART, CHRISTOFFER. “Applying Machine Learning to Identify Maintenance Level for Software Releases
.” 2020. Web. 27 Feb 2021.
Vancouver:
STUART C. Applying Machine Learning to Identify Maintenance Level for Software Releases
. [Internet] [Thesis]. Chalmers University of Technology; 2020. [cited 2021 Feb 27].
Available from: http://hdl.handle.net/20.500.12380/300701.
Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation
Council of Science Editors:
STUART C. Applying Machine Learning to Identify Maintenance Level for Software Releases
. [Thesis]. Chalmers University of Technology; 2020. Available from: http://hdl.handle.net/20.500.12380/300701
Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation

Rochester Institute of Technology
28.
Karnam, Srivallabha.
Self-Supervised Learning for Segmentation using Image Reconstruction.
Degree: MS, Computer Engineering, 2020, Rochester Institute of Technology
URL: https://scholarworks.rit.edu/theses/10532
► Deep learning is the engine that is piloting tremendous growth in various segments of the industry by consuming valuable fuel called data. We are…
(more)
▼ Deep
learning is the engine that is piloting tremendous growth in various segments of the industry by consuming valuable fuel called data. We are witnessing many businesses adopting this technology be it healthcare, transportation, defense, semiconductor, or retail. But most of the accomplishments that we see now rely on supervised
learning. Supervised
learning needs a substantial volume of labeled data which are usually annotated by humans- an arduous and expensive task often leading to datasets that are insufficient in size or human labeling errors. The performance of deep
learning models is only as good as the data. Self-supervised
learning minimizes the need for labeled data as it extracts the pertinent context and inherited data content. We are inspired by image interpolation where we resize an image from a one-pixel grid to another. We introduce a novel self-supervised
learning method specialized for semantic segmentation tasks. We use Image reconstruction as a pre-text task where pixels and or pixel channel (R or G or B pixel channel) in the input images are dropped in a defined or random manner and the original image serves as ground truth. We use the ImageNet dataset for a pretext
learning task, and PASCAL V0C to evaluate efficacy of proposed methods. In segmentation tasks decoder is equally important as the encoder, since our proposed method learns both the encoder and decoder as a part of a pretext task, our method outperforms existing self-supervised segmentation methods.
Advisors/Committee Members: Raymond Ptucha.
Subjects/Keywords: Classification; Computer vision; Self-supervised learning; Semantic segmentation; Unsupervised learning
Record Details
Similar Records
Cite
Share »
Record Details
Similar Records
Cite
« Share





❌
APA ·
Chicago ·
MLA ·
Vancouver ·
CSE |
Export
to Zotero / EndNote / Reference
Manager
APA (6th Edition):
Karnam, S. (2020). Self-Supervised Learning for Segmentation using Image Reconstruction. (Masters Thesis). Rochester Institute of Technology. Retrieved from https://scholarworks.rit.edu/theses/10532
Chicago Manual of Style (16th Edition):
Karnam, Srivallabha. “Self-Supervised Learning for Segmentation using Image Reconstruction.” 2020. Masters Thesis, Rochester Institute of Technology. Accessed February 27, 2021.
https://scholarworks.rit.edu/theses/10532.
MLA Handbook (7th Edition):
Karnam, Srivallabha. “Self-Supervised Learning for Segmentation using Image Reconstruction.” 2020. Web. 27 Feb 2021.
Vancouver:
Karnam S. Self-Supervised Learning for Segmentation using Image Reconstruction. [Internet] [Masters thesis]. Rochester Institute of Technology; 2020. [cited 2021 Feb 27].
Available from: https://scholarworks.rit.edu/theses/10532.
Council of Science Editors:
Karnam S. Self-Supervised Learning for Segmentation using Image Reconstruction. [Masters Thesis]. Rochester Institute of Technology; 2020. Available from: https://scholarworks.rit.edu/theses/10532

Rochester Institute of Technology
29.
Shembekar, Parikshit Prashant.
Anomaly Detection in Videos through Deep Unsupervised Techniques.
Degree: MS, Computer Science (GCCIS), 2020, Rochester Institute of Technology
URL: https://scholarworks.rit.edu/theses/10572
► Identifying abnormality in videos is an area of active research. Most of the work makes extensive use of supervised approaches, even though these methods…
(more)
▼ Identifying abnormality in videos is an area of active research. Most of the work makes extensive use of supervised approaches, even though these methods often give superior performances the major drawback being abnormalities cannot be conformed to select classes, thus the need for
unsupervised models to approach this task. We introduce Dirichlet Process Mixture Models (DPMM) along with Autoencoders to learn the normality in the data. Autoencoders have been extensively used in the literature for feature extraction and enable us to capture rich features into a small dimensional space. We use the Stick Breaking formulation of the DPMM which is a non-parametric version of the Gaussian mixture model and it can create new clusters as more and more data is observed. We exploit this property of the stick-breaking model to incorporate online
learning and prediction of data in an
unsupervised manner. We first introduce a two-phase model with feature extraction through autoencoders in the first step and then model inference through the DPMM in the second step. We seek to improve upon this model by introducing a model that does both the feature extraction and model inference in an end-to-end fashion by modeling the stick-breaking formulation to the Variational Autoencoder (VAE) setting.
Advisors/Committee Members: Ifeoma Nwogu.
Subjects/Keywords: Anomaly detection; Deep learning; Probabilistic graphical models; Unsupervised learning
Record Details
Similar Records
Cite
Share »
Record Details
Similar Records
Cite
« Share





❌
APA ·
Chicago ·
MLA ·
Vancouver ·
CSE |
Export
to Zotero / EndNote / Reference
Manager
APA (6th Edition):
Shembekar, P. P. (2020). Anomaly Detection in Videos through Deep Unsupervised Techniques. (Masters Thesis). Rochester Institute of Technology. Retrieved from https://scholarworks.rit.edu/theses/10572
Chicago Manual of Style (16th Edition):
Shembekar, Parikshit Prashant. “Anomaly Detection in Videos through Deep Unsupervised Techniques.” 2020. Masters Thesis, Rochester Institute of Technology. Accessed February 27, 2021.
https://scholarworks.rit.edu/theses/10572.
MLA Handbook (7th Edition):
Shembekar, Parikshit Prashant. “Anomaly Detection in Videos through Deep Unsupervised Techniques.” 2020. Web. 27 Feb 2021.
Vancouver:
Shembekar PP. Anomaly Detection in Videos through Deep Unsupervised Techniques. [Internet] [Masters thesis]. Rochester Institute of Technology; 2020. [cited 2021 Feb 27].
Available from: https://scholarworks.rit.edu/theses/10572.
Council of Science Editors:
Shembekar PP. Anomaly Detection in Videos through Deep Unsupervised Techniques. [Masters Thesis]. Rochester Institute of Technology; 2020. Available from: https://scholarworks.rit.edu/theses/10572

Cornell University
30.
Sener, Ozan.
Learning From Large-Scale Visual Data For Robots.
Degree: PhD, Electrical Engineering, 2016, Cornell University
URL: http://hdl.handle.net/1813/45306
Subjects/Keywords: Robotics; Unsupervised Learning; Deep Learning
Record Details
Similar Records
Cite
Share »
Record Details
Similar Records
Cite
« Share





❌
APA ·
Chicago ·
MLA ·
Vancouver ·
CSE |
Export
to Zotero / EndNote / Reference
Manager
APA (6th Edition):
Sener, O. (2016). Learning From Large-Scale Visual Data For Robots. (Doctoral Dissertation). Cornell University. Retrieved from http://hdl.handle.net/1813/45306
Chicago Manual of Style (16th Edition):
Sener, Ozan. “Learning From Large-Scale Visual Data For Robots.” 2016. Doctoral Dissertation, Cornell University. Accessed February 27, 2021.
http://hdl.handle.net/1813/45306.
MLA Handbook (7th Edition):
Sener, Ozan. “Learning From Large-Scale Visual Data For Robots.” 2016. Web. 27 Feb 2021.
Vancouver:
Sener O. Learning From Large-Scale Visual Data For Robots. [Internet] [Doctoral dissertation]. Cornell University; 2016. [cited 2021 Feb 27].
Available from: http://hdl.handle.net/1813/45306.
Council of Science Editors:
Sener O. Learning From Large-Scale Visual Data For Robots. [Doctoral Dissertation]. Cornell University; 2016. Available from: http://hdl.handle.net/1813/45306
◁ [1] [2] [3] [4] [5] … [17] ▶
.