Advanced search options

Advanced Search Options 🞨

Browse by author name (“Author name starts with…”).

Find ETDs with:

in
/  
in
/  
in
/  
in

Written in Published in Earliest date Latest date

Sorted by

Results per page:

Sorted by: relevance · author · university · dateNew search

You searched for subject:( Q learning algorithm). Showing records 1 – 30 of 56624 total matches.

[1] [2] [3] [4] [5] … [1888]

Search Limiters

Last 2 Years | English Only

Degrees

Languages

Country

▼ Search Limiters


Universidade do Rio Grande do Norte

1. Santos, João Paulo Queiroz dos. Uma implementação paralela híbrida para o problema do caixeiro viajante usando algoritmos genéticos, GRASP e aprendizagem por reforço .

Degree: 2009, Universidade do Rio Grande do Norte

 The metaheuristics techiniques are known to solve optimization problems classified as NP-complete and are successful in obtaining good quality solutions. They use non-deterministic approaches to… (more)

Subjects/Keywords: Metaheurísticas GRASP; Algoritmos genéticos; Q-learning; Sistemas paralelos e distribuídos; GRASP metaheuristics; Genetic algorithm; Q-learning; Parallel and distributed systems

Record DetailsSimilar RecordsGoogle PlusoneFacebookTwitterCiteULikeMendeleyreddit

APA · Chicago · MLA · Vancouver · CSE | Export to Zotero / EndNote / Reference Manager

APA (6th Edition):

Santos, J. P. Q. d. (2009). Uma implementação paralela híbrida para o problema do caixeiro viajante usando algoritmos genéticos, GRASP e aprendizagem por reforço . (Masters Thesis). Universidade do Rio Grande do Norte. Retrieved from http://repositorio.ufrn.br/handle/123456789/15221

Chicago Manual of Style (16th Edition):

Santos, João Paulo Queiroz dos. “Uma implementação paralela híbrida para o problema do caixeiro viajante usando algoritmos genéticos, GRASP e aprendizagem por reforço .” 2009. Masters Thesis, Universidade do Rio Grande do Norte. Accessed October 22, 2019. http://repositorio.ufrn.br/handle/123456789/15221.

MLA Handbook (7th Edition):

Santos, João Paulo Queiroz dos. “Uma implementação paralela híbrida para o problema do caixeiro viajante usando algoritmos genéticos, GRASP e aprendizagem por reforço .” 2009. Web. 22 Oct 2019.

Vancouver:

Santos JPQd. Uma implementação paralela híbrida para o problema do caixeiro viajante usando algoritmos genéticos, GRASP e aprendizagem por reforço . [Internet] [Masters thesis]. Universidade do Rio Grande do Norte; 2009. [cited 2019 Oct 22]. Available from: http://repositorio.ufrn.br/handle/123456789/15221.

Council of Science Editors:

Santos JPQd. Uma implementação paralela híbrida para o problema do caixeiro viajante usando algoritmos genéticos, GRASP e aprendizagem por reforço . [Masters Thesis]. Universidade do Rio Grande do Norte; 2009. Available from: http://repositorio.ufrn.br/handle/123456789/15221


Universidade do Rio Grande do Norte

2. Santos, João Paulo Queiroz dos. Uma implementação paralela híbrida para o problema do caixeiro viajante usando algoritmos genéticos, GRASP e aprendizagem por reforço .

Degree: 2009, Universidade do Rio Grande do Norte

 The metaheuristics techiniques are known to solve optimization problems classified as NP-complete and are successful in obtaining good quality solutions. They use non-deterministic approaches to… (more)

Subjects/Keywords: Metaheurísticas GRASP; Algoritmos genéticos; Q-learning; Sistemas paralelos e distribuídos; GRASP metaheuristics; Genetic algorithm; Q-learning; Parallel and distributed systems

Record DetailsSimilar RecordsGoogle PlusoneFacebookTwitterCiteULikeMendeleyreddit

APA · Chicago · MLA · Vancouver · CSE | Export to Zotero / EndNote / Reference Manager

APA (6th Edition):

Santos, J. P. Q. d. (2009). Uma implementação paralela híbrida para o problema do caixeiro viajante usando algoritmos genéticos, GRASP e aprendizagem por reforço . (Thesis). Universidade do Rio Grande do Norte. Retrieved from http://repositorio.ufrn.br/handle/123456789/15221

Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation

Chicago Manual of Style (16th Edition):

Santos, João Paulo Queiroz dos. “Uma implementação paralela híbrida para o problema do caixeiro viajante usando algoritmos genéticos, GRASP e aprendizagem por reforço .” 2009. Thesis, Universidade do Rio Grande do Norte. Accessed October 22, 2019. http://repositorio.ufrn.br/handle/123456789/15221.

Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation

MLA Handbook (7th Edition):

Santos, João Paulo Queiroz dos. “Uma implementação paralela híbrida para o problema do caixeiro viajante usando algoritmos genéticos, GRASP e aprendizagem por reforço .” 2009. Web. 22 Oct 2019.

Vancouver:

Santos JPQd. Uma implementação paralela híbrida para o problema do caixeiro viajante usando algoritmos genéticos, GRASP e aprendizagem por reforço . [Internet] [Thesis]. Universidade do Rio Grande do Norte; 2009. [cited 2019 Oct 22]. Available from: http://repositorio.ufrn.br/handle/123456789/15221.

Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation

Council of Science Editors:

Santos JPQd. Uma implementação paralela híbrida para o problema do caixeiro viajante usando algoritmos genéticos, GRASP e aprendizagem por reforço . [Thesis]. Universidade do Rio Grande do Norte; 2009. Available from: http://repositorio.ufrn.br/handle/123456789/15221

Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation

3. Francisco Chagas de Lima Júnior. Algoritmo Q-learning como estratégia de exploração e/ou explotação para metaheurísticas GRASP e algoritmo genético.

Degree: 2009, Universidade Federal do Rio Grande do Norte

Técnicas de otimização conhecidas como metaheurísticas têm obtido sucesso na resolução de problemas classificados como NP - Árduos. Estes métodos utilizam abordagens não determinísticas que… (more)

Subjects/Keywords: MetaheurísticaGRASP; Algoritmos genéticos; AlgoritmoQ-learning; Problema do caixeiro viajante; ENGENHARIA ELETRICA; GRASP metaheuristic; Genetic algorithm; Q-learning algorithm; Travelling salesman problem

Record DetailsSimilar RecordsGoogle PlusoneFacebookTwitterCiteULikeMendeleyreddit

APA · Chicago · MLA · Vancouver · CSE | Export to Zotero / EndNote / Reference Manager

APA (6th Edition):

Júnior, F. C. d. L. (2009). Algoritmo Q-learning como estratégia de exploração e/ou explotação para metaheurísticas GRASP e algoritmo genético. (Thesis). Universidade Federal do Rio Grande do Norte. Retrieved from http://bdtd.bczm.ufrn.br/tedesimplificado//tde_busca/arquivo.php?codArquivo=2428

Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation

Chicago Manual of Style (16th Edition):

Júnior, Francisco Chagas de Lima. “Algoritmo Q-learning como estratégia de exploração e/ou explotação para metaheurísticas GRASP e algoritmo genético.” 2009. Thesis, Universidade Federal do Rio Grande do Norte. Accessed October 22, 2019. http://bdtd.bczm.ufrn.br/tedesimplificado//tde_busca/arquivo.php?codArquivo=2428.

Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation

MLA Handbook (7th Edition):

Júnior, Francisco Chagas de Lima. “Algoritmo Q-learning como estratégia de exploração e/ou explotação para metaheurísticas GRASP e algoritmo genético.” 2009. Web. 22 Oct 2019.

Vancouver:

Júnior FCdL. Algoritmo Q-learning como estratégia de exploração e/ou explotação para metaheurísticas GRASP e algoritmo genético. [Internet] [Thesis]. Universidade Federal do Rio Grande do Norte; 2009. [cited 2019 Oct 22]. Available from: http://bdtd.bczm.ufrn.br/tedesimplificado//tde_busca/arquivo.php?codArquivo=2428.

Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation

Council of Science Editors:

Júnior FCdL. Algoritmo Q-learning como estratégia de exploração e/ou explotação para metaheurísticas GRASP e algoritmo genético. [Thesis]. Universidade Federal do Rio Grande do Norte; 2009. Available from: http://bdtd.bczm.ufrn.br/tedesimplificado//tde_busca/arquivo.php?codArquivo=2428

Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation


Universidade do Rio Grande do Norte

4. Lima Júnior, Francisco Chagas de. Algoritmo Q-learning como estratégia de exploração e/ou explotação para metaheurísticas GRASP e algoritmo genético .

Degree: 2009, Universidade do Rio Grande do Norte

 Techniques of optimization known as metaheuristics have achieved success in the resolution of many problems classified as NP-Hard. These methods use non deterministic approaches that… (more)

Subjects/Keywords: MetaheurísticaGRASP; Algoritmos genéticos; AlgoritmoQ-learning; Problema do caixeiro viajante; GRASP metaheuristic; Genetic algorithm; Q-learning algorithm; Travelling salesman problem

Record DetailsSimilar RecordsGoogle PlusoneFacebookTwitterCiteULikeMendeleyreddit

APA · Chicago · MLA · Vancouver · CSE | Export to Zotero / EndNote / Reference Manager

APA (6th Edition):

Lima Júnior, F. C. d. (2009). Algoritmo Q-learning como estratégia de exploração e/ou explotação para metaheurísticas GRASP e algoritmo genético . (Doctoral Dissertation). Universidade do Rio Grande do Norte. Retrieved from http://repositorio.ufrn.br/handle/123456789/15129

Chicago Manual of Style (16th Edition):

Lima Júnior, Francisco Chagas de. “Algoritmo Q-learning como estratégia de exploração e/ou explotação para metaheurísticas GRASP e algoritmo genético .” 2009. Doctoral Dissertation, Universidade do Rio Grande do Norte. Accessed October 22, 2019. http://repositorio.ufrn.br/handle/123456789/15129.

MLA Handbook (7th Edition):

Lima Júnior, Francisco Chagas de. “Algoritmo Q-learning como estratégia de exploração e/ou explotação para metaheurísticas GRASP e algoritmo genético .” 2009. Web. 22 Oct 2019.

Vancouver:

Lima Júnior FCd. Algoritmo Q-learning como estratégia de exploração e/ou explotação para metaheurísticas GRASP e algoritmo genético . [Internet] [Doctoral dissertation]. Universidade do Rio Grande do Norte; 2009. [cited 2019 Oct 22]. Available from: http://repositorio.ufrn.br/handle/123456789/15129.

Council of Science Editors:

Lima Júnior FCd. Algoritmo Q-learning como estratégia de exploração e/ou explotação para metaheurísticas GRASP e algoritmo genético . [Doctoral Dissertation]. Universidade do Rio Grande do Norte; 2009. Available from: http://repositorio.ufrn.br/handle/123456789/15129


Universidade do Rio Grande do Norte

5. Lima Júnior, Francisco Chagas de. Algoritmo Q-learning como estratégia de exploração e/ou explotação para metaheurísticas GRASP e algoritmo genético .

Degree: 2009, Universidade do Rio Grande do Norte

 Techniques of optimization known as metaheuristics have achieved success in the resolution of many problems classified as NP-Hard. These methods use non deterministic approaches that… (more)

Subjects/Keywords: MetaheurísticaGRASP; Algoritmos genéticos; AlgoritmoQ-learning; Problema do caixeiro viajante; GRASP metaheuristic; Genetic algorithm; Q-learning algorithm; Travelling salesman problem

Record DetailsSimilar RecordsGoogle PlusoneFacebookTwitterCiteULikeMendeleyreddit

APA · Chicago · MLA · Vancouver · CSE | Export to Zotero / EndNote / Reference Manager

APA (6th Edition):

Lima Júnior, F. C. d. (2009). Algoritmo Q-learning como estratégia de exploração e/ou explotação para metaheurísticas GRASP e algoritmo genético . (Thesis). Universidade do Rio Grande do Norte. Retrieved from http://repositorio.ufrn.br/handle/123456789/15129

Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation

Chicago Manual of Style (16th Edition):

Lima Júnior, Francisco Chagas de. “Algoritmo Q-learning como estratégia de exploração e/ou explotação para metaheurísticas GRASP e algoritmo genético .” 2009. Thesis, Universidade do Rio Grande do Norte. Accessed October 22, 2019. http://repositorio.ufrn.br/handle/123456789/15129.

Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation

MLA Handbook (7th Edition):

Lima Júnior, Francisco Chagas de. “Algoritmo Q-learning como estratégia de exploração e/ou explotação para metaheurísticas GRASP e algoritmo genético .” 2009. Web. 22 Oct 2019.

Vancouver:

Lima Júnior FCd. Algoritmo Q-learning como estratégia de exploração e/ou explotação para metaheurísticas GRASP e algoritmo genético . [Internet] [Thesis]. Universidade do Rio Grande do Norte; 2009. [cited 2019 Oct 22]. Available from: http://repositorio.ufrn.br/handle/123456789/15129.

Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation

Council of Science Editors:

Lima Júnior FCd. Algoritmo Q-learning como estratégia de exploração e/ou explotação para metaheurísticas GRASP e algoritmo genético . [Thesis]. Universidade do Rio Grande do Norte; 2009. Available from: http://repositorio.ufrn.br/handle/123456789/15129

Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation


Indian Institute of Science

6. Lakshmanan, K. Online Learning and Simulation Based Algorithms for Stochastic Optimization.

Degree: 2012, Indian Institute of Science

 In many optimization problems, the relationship between the objective and parameters is not known. The objective function itself may be stochastic such as a long-run… (more)

Subjects/Keywords: Stochastic Approximation Algorithms; Stochastic Optimization; Markov Decision Process; Reinforcement Learning Algorithm; Queueing Networks; Queuing Theory; Quasi-Newton Stochastic Approximation Algorithm; Online Q-Learning Algorithm; Online Actor-Critic Algorithm; Markov Decision Processes; Q-learning Algorithm; Linear Function Approximation; Quasi-Newton Smoothed Functional Algorithms; Computer Science

Record DetailsSimilar RecordsGoogle PlusoneFacebookTwitterCiteULikeMendeleyreddit

APA · Chicago · MLA · Vancouver · CSE | Export to Zotero / EndNote / Reference Manager

APA (6th Edition):

Lakshmanan, K. (2012). Online Learning and Simulation Based Algorithms for Stochastic Optimization. (Thesis). Indian Institute of Science. Retrieved from http://hdl.handle.net/2005/3245

Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation

Chicago Manual of Style (16th Edition):

Lakshmanan, K. “Online Learning and Simulation Based Algorithms for Stochastic Optimization.” 2012. Thesis, Indian Institute of Science. Accessed October 22, 2019. http://hdl.handle.net/2005/3245.

Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation

MLA Handbook (7th Edition):

Lakshmanan, K. “Online Learning and Simulation Based Algorithms for Stochastic Optimization.” 2012. Web. 22 Oct 2019.

Vancouver:

Lakshmanan K. Online Learning and Simulation Based Algorithms for Stochastic Optimization. [Internet] [Thesis]. Indian Institute of Science; 2012. [cited 2019 Oct 22]. Available from: http://hdl.handle.net/2005/3245.

Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation

Council of Science Editors:

Lakshmanan K. Online Learning and Simulation Based Algorithms for Stochastic Optimization. [Thesis]. Indian Institute of Science; 2012. Available from: http://hdl.handle.net/2005/3245

Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation


Mississippi State University

7. WANG, YI-CHI. APPLICATION OF REINFORCEMENT LEARNING TO MULTI-AGENT PRODUCTION SCHEDULING.

Degree: PhD, Industrial Engineering, 2003, Mississippi State University

 Reinforcement learning (RL) has received attention in recent years from agent-based researchers because it can be applied to problems where autonomous agents learn to select… (more)

Subjects/Keywords: Q-LEARNING ALGORITHM; PRODUCTION SCHEDULING; MULTI-AGENT; REINFORCEMENT LEARNING

Record DetailsSimilar RecordsGoogle PlusoneFacebookTwitterCiteULikeMendeleyreddit

APA · Chicago · MLA · Vancouver · CSE | Export to Zotero / EndNote / Reference Manager

APA (6th Edition):

WANG, Y. (2003). APPLICATION OF REINFORCEMENT LEARNING TO MULTI-AGENT PRODUCTION SCHEDULING. (Doctoral Dissertation). Mississippi State University. Retrieved from http://sun.library.msstate.edu/ETD-db/theses/available/etd-10212003-094739/ ;

Chicago Manual of Style (16th Edition):

WANG, YI-CHI. “APPLICATION OF REINFORCEMENT LEARNING TO MULTI-AGENT PRODUCTION SCHEDULING.” 2003. Doctoral Dissertation, Mississippi State University. Accessed October 22, 2019. http://sun.library.msstate.edu/ETD-db/theses/available/etd-10212003-094739/ ;.

MLA Handbook (7th Edition):

WANG, YI-CHI. “APPLICATION OF REINFORCEMENT LEARNING TO MULTI-AGENT PRODUCTION SCHEDULING.” 2003. Web. 22 Oct 2019.

Vancouver:

WANG Y. APPLICATION OF REINFORCEMENT LEARNING TO MULTI-AGENT PRODUCTION SCHEDULING. [Internet] [Doctoral dissertation]. Mississippi State University; 2003. [cited 2019 Oct 22]. Available from: http://sun.library.msstate.edu/ETD-db/theses/available/etd-10212003-094739/ ;.

Council of Science Editors:

WANG Y. APPLICATION OF REINFORCEMENT LEARNING TO MULTI-AGENT PRODUCTION SCHEDULING. [Doctoral Dissertation]. Mississippi State University; 2003. Available from: http://sun.library.msstate.edu/ETD-db/theses/available/etd-10212003-094739/ ;

8. João Paulo Queiroz dos Santos. Uma implementação paralela híbrida para o problema do caixeiro viajante usando algoritmos genéticos, GRASP e aprendizagem por reforço.

Degree: 2009, Universidade Federal do Rio Grande do Norte

As metaheurísticas são técnicas conhecidas para a resolução de problemas de otimização, classificados como NP-Completos e vêm obtendo sucesso em soluções aproximadas de boa qualidade.… (more)

Subjects/Keywords: Genetic algorithm; Parallel and distributed systems; Metaheurísticas GRASP; Algoritmos genéticos; Q-learning; Sistemas paralelos e distribuídos; ENGENHARIA ELETRICA; GRASP metaheuristics; Q-learning

Record DetailsSimilar RecordsGoogle PlusoneFacebookTwitterCiteULikeMendeleyreddit

APA · Chicago · MLA · Vancouver · CSE | Export to Zotero / EndNote / Reference Manager

APA (6th Edition):

Santos, J. P. Q. d. (2009). Uma implementação paralela híbrida para o problema do caixeiro viajante usando algoritmos genéticos, GRASP e aprendizagem por reforço. (Thesis). Universidade Federal do Rio Grande do Norte. Retrieved from http://bdtd.bczm.ufrn.br/tedesimplificado//tde_busca/arquivo.php?codArquivo=2365

Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation

Chicago Manual of Style (16th Edition):

Santos, João Paulo Queiroz dos. “Uma implementação paralela híbrida para o problema do caixeiro viajante usando algoritmos genéticos, GRASP e aprendizagem por reforço.” 2009. Thesis, Universidade Federal do Rio Grande do Norte. Accessed October 22, 2019. http://bdtd.bczm.ufrn.br/tedesimplificado//tde_busca/arquivo.php?codArquivo=2365.

Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation

MLA Handbook (7th Edition):

Santos, João Paulo Queiroz dos. “Uma implementação paralela híbrida para o problema do caixeiro viajante usando algoritmos genéticos, GRASP e aprendizagem por reforço.” 2009. Web. 22 Oct 2019.

Vancouver:

Santos JPQd. Uma implementação paralela híbrida para o problema do caixeiro viajante usando algoritmos genéticos, GRASP e aprendizagem por reforço. [Internet] [Thesis]. Universidade Federal do Rio Grande do Norte; 2009. [cited 2019 Oct 22]. Available from: http://bdtd.bczm.ufrn.br/tedesimplificado//tde_busca/arquivo.php?codArquivo=2365.

Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation

Council of Science Editors:

Santos JPQd. Uma implementação paralela híbrida para o problema do caixeiro viajante usando algoritmos genéticos, GRASP e aprendizagem por reforço. [Thesis]. Universidade Federal do Rio Grande do Norte; 2009. Available from: http://bdtd.bczm.ufrn.br/tedesimplificado//tde_busca/arquivo.php?codArquivo=2365

Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation


Brno University of Technology

9. Kefurt, Pavel. Strojové učení - aplikace pro demonstraci základních přístupů .

Degree: 2013, Brno University of Technology

 Tato práce se zabývá především základními algoritmy strojového učení. V první části práce jsou vybrané algoritmy popsány. Zbývající část se následně věnuje implementaci těchto algoritmů… (more)

Subjects/Keywords: Umělá inteligence; strojové učení; učení s učitelem; učení bez učitele; posilované učení; ID3 algoritmus; Back Propagation; K-Means clustering; Self-Organizing Maps; SOM; Q-Learning; State-Action-Reward-State-Action; SARSA; Artificial learning; machine learning; supervised learning; unsupervised learning; reinforcement learning; ID3 algorithm; Back Propagation; K-Means clustering; Self-Organizing Maps; SOM; Q-Learning; State-Action-Reward-State-Action; SARSA

Record DetailsSimilar RecordsGoogle PlusoneFacebookTwitterCiteULikeMendeleyreddit

APA · Chicago · MLA · Vancouver · CSE | Export to Zotero / EndNote / Reference Manager

APA (6th Edition):

Kefurt, P. (2013). Strojové učení - aplikace pro demonstraci základních přístupů . (Thesis). Brno University of Technology. Retrieved from http://hdl.handle.net/11012/54905

Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation

Chicago Manual of Style (16th Edition):

Kefurt, Pavel. “Strojové učení - aplikace pro demonstraci základních přístupů .” 2013. Thesis, Brno University of Technology. Accessed October 22, 2019. http://hdl.handle.net/11012/54905.

Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation

MLA Handbook (7th Edition):

Kefurt, Pavel. “Strojové učení - aplikace pro demonstraci základních přístupů .” 2013. Web. 22 Oct 2019.

Vancouver:

Kefurt P. Strojové učení - aplikace pro demonstraci základních přístupů . [Internet] [Thesis]. Brno University of Technology; 2013. [cited 2019 Oct 22]. Available from: http://hdl.handle.net/11012/54905.

Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation

Council of Science Editors:

Kefurt P. Strojové učení - aplikace pro demonstraci základních přístupů . [Thesis]. Brno University of Technology; 2013. Available from: http://hdl.handle.net/11012/54905

Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation


Delft University of Technology

10. Chaeron, S.L.A. Ranking the Level of Fear from Voice using Nominal Classification Methods:.

Degree: 2009, Delft University of Technology

 To investigate human emotion, which is conveyed in human speech, methods which can achieve this need to be developed. One way to respond to it… (more)

Subjects/Keywords: Machine learning algorithm

Record DetailsSimilar RecordsGoogle PlusoneFacebookTwitterCiteULikeMendeleyreddit

APA · Chicago · MLA · Vancouver · CSE | Export to Zotero / EndNote / Reference Manager

APA (6th Edition):

Chaeron, S. L. A. (2009). Ranking the Level of Fear from Voice using Nominal Classification Methods:. (Masters Thesis). Delft University of Technology. Retrieved from http://resolver.tudelft.nl/uuid:d537fb74-e1c6-4efc-a661-18704cc170d9

Chicago Manual of Style (16th Edition):

Chaeron, S L A. “Ranking the Level of Fear from Voice using Nominal Classification Methods:.” 2009. Masters Thesis, Delft University of Technology. Accessed October 22, 2019. http://resolver.tudelft.nl/uuid:d537fb74-e1c6-4efc-a661-18704cc170d9.

MLA Handbook (7th Edition):

Chaeron, S L A. “Ranking the Level of Fear from Voice using Nominal Classification Methods:.” 2009. Web. 22 Oct 2019.

Vancouver:

Chaeron SLA. Ranking the Level of Fear from Voice using Nominal Classification Methods:. [Internet] [Masters thesis]. Delft University of Technology; 2009. [cited 2019 Oct 22]. Available from: http://resolver.tudelft.nl/uuid:d537fb74-e1c6-4efc-a661-18704cc170d9.

Council of Science Editors:

Chaeron SLA. Ranking the Level of Fear from Voice using Nominal Classification Methods:. [Masters Thesis]. Delft University of Technology; 2009. Available from: http://resolver.tudelft.nl/uuid:d537fb74-e1c6-4efc-a661-18704cc170d9


University of Arizona

11. Soto Santibanez, Miguel Angel. BUILDING AN ARTIFICIAL CEREBELLUM USING A SYSTEM OF DISTRIBUTED Q-LEARNING AGENTS .

Degree: 2010, University of Arizona

 About 400 million years ago sharks developed a separate co-processor in their brains that not only made them faster but also more precisely coordinated. This… (more)

Subjects/Keywords: Artificial; Cerebellum; Distributed; Q-learning

Record DetailsSimilar RecordsGoogle PlusoneFacebookTwitterCiteULikeMendeleyreddit

APA · Chicago · MLA · Vancouver · CSE | Export to Zotero / EndNote / Reference Manager

APA (6th Edition):

Soto Santibanez, M. A. (2010). BUILDING AN ARTIFICIAL CEREBELLUM USING A SYSTEM OF DISTRIBUTED Q-LEARNING AGENTS . (Doctoral Dissertation). University of Arizona. Retrieved from http://hdl.handle.net/10150/194811

Chicago Manual of Style (16th Edition):

Soto Santibanez, Miguel Angel. “BUILDING AN ARTIFICIAL CEREBELLUM USING A SYSTEM OF DISTRIBUTED Q-LEARNING AGENTS .” 2010. Doctoral Dissertation, University of Arizona. Accessed October 22, 2019. http://hdl.handle.net/10150/194811.

MLA Handbook (7th Edition):

Soto Santibanez, Miguel Angel. “BUILDING AN ARTIFICIAL CEREBELLUM USING A SYSTEM OF DISTRIBUTED Q-LEARNING AGENTS .” 2010. Web. 22 Oct 2019.

Vancouver:

Soto Santibanez MA. BUILDING AN ARTIFICIAL CEREBELLUM USING A SYSTEM OF DISTRIBUTED Q-LEARNING AGENTS . [Internet] [Doctoral dissertation]. University of Arizona; 2010. [cited 2019 Oct 22]. Available from: http://hdl.handle.net/10150/194811.

Council of Science Editors:

Soto Santibanez MA. BUILDING AN ARTIFICIAL CEREBELLUM USING A SYSTEM OF DISTRIBUTED Q-LEARNING AGENTS . [Doctoral Dissertation]. University of Arizona; 2010. Available from: http://hdl.handle.net/10150/194811


Brigham Young University

12. White, Spencer Kesson. Reinforcement Programming: A New Technique in Automatic Algorithm Development.

Degree: MS, 2006, Brigham Young University

 Reinforcement programming is a new technique for using computers to automatically create algorithms. By using the principles of reinforcement learning and Q-learning, reinforcement programming learns… (more)

Subjects/Keywords: genetic reinforcement programming Q-learning RPSort algorithm automatic; Computer Sciences

Record DetailsSimilar RecordsGoogle PlusoneFacebookTwitterCiteULikeMendeleyreddit

APA · Chicago · MLA · Vancouver · CSE | Export to Zotero / EndNote / Reference Manager

APA (6th Edition):

White, S. K. (2006). Reinforcement Programming: A New Technique in Automatic Algorithm Development. (Masters Thesis). Brigham Young University. Retrieved from https://scholarsarchive.byu.edu/cgi/viewcontent.cgi?article=1479&context=etd

Chicago Manual of Style (16th Edition):

White, Spencer Kesson. “Reinforcement Programming: A New Technique in Automatic Algorithm Development.” 2006. Masters Thesis, Brigham Young University. Accessed October 22, 2019. https://scholarsarchive.byu.edu/cgi/viewcontent.cgi?article=1479&context=etd.

MLA Handbook (7th Edition):

White, Spencer Kesson. “Reinforcement Programming: A New Technique in Automatic Algorithm Development.” 2006. Web. 22 Oct 2019.

Vancouver:

White SK. Reinforcement Programming: A New Technique in Automatic Algorithm Development. [Internet] [Masters thesis]. Brigham Young University; 2006. [cited 2019 Oct 22]. Available from: https://scholarsarchive.byu.edu/cgi/viewcontent.cgi?article=1479&context=etd.

Council of Science Editors:

White SK. Reinforcement Programming: A New Technique in Automatic Algorithm Development. [Masters Thesis]. Brigham Young University; 2006. Available from: https://scholarsarchive.byu.edu/cgi/viewcontent.cgi?article=1479&context=etd


Université Catholique de Louvain

13. Degryse, Baptiste. Deep Q-Learning for Robocode.

Degree: 2017, Université Catholique de Louvain

Q-learning can be used to find an optimal action-selection policy for any given finite Markov Decision Process. The Q-network is a neural network that approximate… (more)

Subjects/Keywords: Deep Q-Learning; Robocode; Q-Learning; LSTM; Convolutional; Neural; Network; Q-Network

Record DetailsSimilar RecordsGoogle PlusoneFacebookTwitterCiteULikeMendeleyreddit

APA · Chicago · MLA · Vancouver · CSE | Export to Zotero / EndNote / Reference Manager

APA (6th Edition):

Degryse, B. (2017). Deep Q-Learning for Robocode. (Thesis). Université Catholique de Louvain. Retrieved from http://hdl.handle.net/2078.1/thesis:10589

Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation

Chicago Manual of Style (16th Edition):

Degryse, Baptiste. “Deep Q-Learning for Robocode.” 2017. Thesis, Université Catholique de Louvain. Accessed October 22, 2019. http://hdl.handle.net/2078.1/thesis:10589.

Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation

MLA Handbook (7th Edition):

Degryse, Baptiste. “Deep Q-Learning for Robocode.” 2017. Web. 22 Oct 2019.

Vancouver:

Degryse B. Deep Q-Learning for Robocode. [Internet] [Thesis]. Université Catholique de Louvain; 2017. [cited 2019 Oct 22]. Available from: http://hdl.handle.net/2078.1/thesis:10589.

Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation

Council of Science Editors:

Degryse B. Deep Q-Learning for Robocode. [Thesis]. Université Catholique de Louvain; 2017. Available from: http://hdl.handle.net/2078.1/thesis:10589

Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation


Delft University of Technology

14. Molenkamp, D. Intelligent Controller Selection for Aggressive Quadrotor Manoeuvring: A reinforcement learning approach:.

Degree: 2016, Delft University of Technology

 A novel intelligent controller selection method for quadrotor attitude and altitude control is presented that maintains performance in different regimes of the flight envelope. Conventional… (more)

Subjects/Keywords: reinforcement learning; q-learning; quadcopter; adaptive control

Record DetailsSimilar RecordsGoogle PlusoneFacebookTwitterCiteULikeMendeleyreddit

APA · Chicago · MLA · Vancouver · CSE | Export to Zotero / EndNote / Reference Manager

APA (6th Edition):

Molenkamp, D. (2016). Intelligent Controller Selection for Aggressive Quadrotor Manoeuvring: A reinforcement learning approach:. (Masters Thesis). Delft University of Technology. Retrieved from http://resolver.tudelft.nl/uuid:70dd7e02-d234-4774-9e54-94ba8d881de6

Chicago Manual of Style (16th Edition):

Molenkamp, D. “Intelligent Controller Selection for Aggressive Quadrotor Manoeuvring: A reinforcement learning approach:.” 2016. Masters Thesis, Delft University of Technology. Accessed October 22, 2019. http://resolver.tudelft.nl/uuid:70dd7e02-d234-4774-9e54-94ba8d881de6.

MLA Handbook (7th Edition):

Molenkamp, D. “Intelligent Controller Selection for Aggressive Quadrotor Manoeuvring: A reinforcement learning approach:.” 2016. Web. 22 Oct 2019.

Vancouver:

Molenkamp D. Intelligent Controller Selection for Aggressive Quadrotor Manoeuvring: A reinforcement learning approach:. [Internet] [Masters thesis]. Delft University of Technology; 2016. [cited 2019 Oct 22]. Available from: http://resolver.tudelft.nl/uuid:70dd7e02-d234-4774-9e54-94ba8d881de6.

Council of Science Editors:

Molenkamp D. Intelligent Controller Selection for Aggressive Quadrotor Manoeuvring: A reinforcement learning approach:. [Masters Thesis]. Delft University of Technology; 2016. Available from: http://resolver.tudelft.nl/uuid:70dd7e02-d234-4774-9e54-94ba8d881de6


University of KwaZulu-Natal

15. Waltham, Michael. Q-Cog: a Q-Learning based cognitive agent architecture for complex 3D virtual worlds.

Degree: 2017, University of KwaZulu-Natal

 Intelligent cognitive agents should be able to autonomously gather new knowledge and learn from their own experiences in order to adapt to a changing environment.… (more)

Subjects/Keywords: 3D.; Virtual worlds.; Q learning.; Virtual learning.

Record DetailsSimilar RecordsGoogle PlusoneFacebookTwitterCiteULikeMendeleyreddit

APA · Chicago · MLA · Vancouver · CSE | Export to Zotero / EndNote / Reference Manager

APA (6th Edition):

Waltham, M. (2017). Q-Cog: a Q-Learning based cognitive agent architecture for complex 3D virtual worlds. (Thesis). University of KwaZulu-Natal. Retrieved from http://hdl.handle.net/10413/15654

Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation

Chicago Manual of Style (16th Edition):

Waltham, Michael. “Q-Cog: a Q-Learning based cognitive agent architecture for complex 3D virtual worlds.” 2017. Thesis, University of KwaZulu-Natal. Accessed October 22, 2019. http://hdl.handle.net/10413/15654.

Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation

MLA Handbook (7th Edition):

Waltham, Michael. “Q-Cog: a Q-Learning based cognitive agent architecture for complex 3D virtual worlds.” 2017. Web. 22 Oct 2019.

Vancouver:

Waltham M. Q-Cog: a Q-Learning based cognitive agent architecture for complex 3D virtual worlds. [Internet] [Thesis]. University of KwaZulu-Natal; 2017. [cited 2019 Oct 22]. Available from: http://hdl.handle.net/10413/15654.

Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation

Council of Science Editors:

Waltham M. Q-Cog: a Q-Learning based cognitive agent architecture for complex 3D virtual worlds. [Thesis]. University of KwaZulu-Natal; 2017. Available from: http://hdl.handle.net/10413/15654

Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation


University of Texas – Austin

16. -5370-2810. Enhancing performance of reinforcement learning models in the presence of noisy rewards.

Degree: MSin Engineering, Operations Research & Industrial Engineering, 2019, University of Texas – Austin

 Reinforcement learning models learn the optimal policy by interacting with the environment and observing the states and rewards. If the rewards that the model observes… (more)

Subjects/Keywords: Reinforcement learning; Q-learning; Noisy rewards

Record DetailsSimilar RecordsGoogle PlusoneFacebookTwitterCiteULikeMendeleyreddit

APA · Chicago · MLA · Vancouver · CSE | Export to Zotero / EndNote / Reference Manager

APA (6th Edition):

-5370-2810. (2019). Enhancing performance of reinforcement learning models in the presence of noisy rewards. (Masters Thesis). University of Texas – Austin. Retrieved from http://dx.doi.org/10.26153/tsw/2860

Note: this citation may be lacking information needed for this citation format:
Author name may be incomplete

Chicago Manual of Style (16th Edition):

-5370-2810. “Enhancing performance of reinforcement learning models in the presence of noisy rewards.” 2019. Masters Thesis, University of Texas – Austin. Accessed October 22, 2019. http://dx.doi.org/10.26153/tsw/2860.

Note: this citation may be lacking information needed for this citation format:
Author name may be incomplete

MLA Handbook (7th Edition):

-5370-2810. “Enhancing performance of reinforcement learning models in the presence of noisy rewards.” 2019. Web. 22 Oct 2019.

Note: this citation may be lacking information needed for this citation format:
Author name may be incomplete

Vancouver:

-5370-2810. Enhancing performance of reinforcement learning models in the presence of noisy rewards. [Internet] [Masters thesis]. University of Texas – Austin; 2019. [cited 2019 Oct 22]. Available from: http://dx.doi.org/10.26153/tsw/2860.

Note: this citation may be lacking information needed for this citation format:
Author name may be incomplete

Council of Science Editors:

-5370-2810. Enhancing performance of reinforcement learning models in the presence of noisy rewards. [Masters Thesis]. University of Texas – Austin; 2019. Available from: http://dx.doi.org/10.26153/tsw/2860

Note: this citation may be lacking information needed for this citation format:
Author name may be incomplete

17. Manoel Leandro de Lima Júnior. Uma contribuição à solução do problema dos k-servos usando aprendizagem por reforço.

Degree: 2005, Universidade Federal do Rio Grande do Norte

 Neste trabalho é proposto um novo algoritmo online para o resolver o Problema dos k-Servos (PKS). O desempenho desta solução é comparado com o de… (more)

Subjects/Keywords: K-Servos; Aprendizado por Reforço; Q-Learning; ENGENHARIA ELETRICA; K-Servos; Reinforcement Learning; Q-Learning

Record DetailsSimilar RecordsGoogle PlusoneFacebookTwitterCiteULikeMendeleyreddit

APA · Chicago · MLA · Vancouver · CSE | Export to Zotero / EndNote / Reference Manager

APA (6th Edition):

Júnior, M. L. d. L. (2005). Uma contribuição à solução do problema dos k-servos usando aprendizagem por reforço. (Thesis). Universidade Federal do Rio Grande do Norte. Retrieved from http://bdtd.bczm.ufrn.br/tedesimplificado//tde_busca/arquivo.php?codArquivo=629

Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation

Chicago Manual of Style (16th Edition):

Júnior, Manoel Leandro de Lima. “Uma contribuição à solução do problema dos k-servos usando aprendizagem por reforço.” 2005. Thesis, Universidade Federal do Rio Grande do Norte. Accessed October 22, 2019. http://bdtd.bczm.ufrn.br/tedesimplificado//tde_busca/arquivo.php?codArquivo=629.

Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation

MLA Handbook (7th Edition):

Júnior, Manoel Leandro de Lima. “Uma contribuição à solução do problema dos k-servos usando aprendizagem por reforço.” 2005. Web. 22 Oct 2019.

Vancouver:

Júnior MLdL. Uma contribuição à solução do problema dos k-servos usando aprendizagem por reforço. [Internet] [Thesis]. Universidade Federal do Rio Grande do Norte; 2005. [cited 2019 Oct 22]. Available from: http://bdtd.bczm.ufrn.br/tedesimplificado//tde_busca/arquivo.php?codArquivo=629.

Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation

Council of Science Editors:

Júnior MLdL. Uma contribuição à solução do problema dos k-servos usando aprendizagem por reforço. [Thesis]. Universidade Federal do Rio Grande do Norte; 2005. Available from: http://bdtd.bczm.ufrn.br/tedesimplificado//tde_busca/arquivo.php?codArquivo=629

Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation


University of Missouri – Columbia

18. Lv, Xutao. Local learning algorithms with application to action recognition and video analysis.

Degree: 2012, University of Missouri – Columbia

 [ACCESS RESTRICTED TO THE UNIVERSITY OF MISSOURI AT AUTHOR'S REQUEST.] Activity analysis has been an active research area in recent years, due to its difficulties… (more)

Subjects/Keywords: activity analysis; superpixelization; learning algorithm

Record DetailsSimilar RecordsGoogle PlusoneFacebookTwitterCiteULikeMendeleyreddit

APA · Chicago · MLA · Vancouver · CSE | Export to Zotero / EndNote / Reference Manager

APA (6th Edition):

Lv, X. (2012). Local learning algorithms with application to action recognition and video analysis. (Thesis). University of Missouri – Columbia. Retrieved from http://hdl.handle.net/10355/33234

Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation

Chicago Manual of Style (16th Edition):

Lv, Xutao. “Local learning algorithms with application to action recognition and video analysis.” 2012. Thesis, University of Missouri – Columbia. Accessed October 22, 2019. http://hdl.handle.net/10355/33234.

Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation

MLA Handbook (7th Edition):

Lv, Xutao. “Local learning algorithms with application to action recognition and video analysis.” 2012. Web. 22 Oct 2019.

Vancouver:

Lv X. Local learning algorithms with application to action recognition and video analysis. [Internet] [Thesis]. University of Missouri – Columbia; 2012. [cited 2019 Oct 22]. Available from: http://hdl.handle.net/10355/33234.

Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation

Council of Science Editors:

Lv X. Local learning algorithms with application to action recognition and video analysis. [Thesis]. University of Missouri – Columbia; 2012. Available from: http://hdl.handle.net/10355/33234

Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation


University of Missouri – Columbia

19. Lv, Xutao. Local learning algorithms with application to action recognition and video analysis.

Degree: 2012, University of Missouri – Columbia

 [ACCESS RESTRICTED TO THE UNIVERSITY OF MISSOURI AT AUTHOR'S REQUEST.] Activity analysis has been an active research area in recent years, due to its difficulties… (more)

Subjects/Keywords: activity analysis; superpixelization; learning algorithm

Record DetailsSimilar RecordsGoogle PlusoneFacebookTwitterCiteULikeMendeleyreddit

APA · Chicago · MLA · Vancouver · CSE | Export to Zotero / EndNote / Reference Manager

APA (6th Edition):

Lv, X. (2012). Local learning algorithms with application to action recognition and video analysis. (Thesis). University of Missouri – Columbia. Retrieved from https://doi.org/10.32469/10355/33234

Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation

Chicago Manual of Style (16th Edition):

Lv, Xutao. “Local learning algorithms with application to action recognition and video analysis.” 2012. Thesis, University of Missouri – Columbia. Accessed October 22, 2019. https://doi.org/10.32469/10355/33234.

Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation

MLA Handbook (7th Edition):

Lv, Xutao. “Local learning algorithms with application to action recognition and video analysis.” 2012. Web. 22 Oct 2019.

Vancouver:

Lv X. Local learning algorithms with application to action recognition and video analysis. [Internet] [Thesis]. University of Missouri – Columbia; 2012. [cited 2019 Oct 22]. Available from: https://doi.org/10.32469/10355/33234.

Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation

Council of Science Editors:

Lv X. Local learning algorithms with application to action recognition and video analysis. [Thesis]. University of Missouri – Columbia; 2012. Available from: https://doi.org/10.32469/10355/33234

Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation


Indian Institute of Science

20. Prashanth, L A. Resource Allocation for Sequential Decision Making Under Uncertainaty : Studies in Vehicular Traffic Control, Service Systems, Sensor Networks and Mechanism Design.

Degree: 2013, Indian Institute of Science

 A fundamental question in a sequential decision making setting under uncertainty is “how to allocate resources amongst competing entities so as to maximize the rewards… (more)

Subjects/Keywords: Vehicular Traffic Control; Service Systems; Sensor Networks; Mechanism Design; Traffic Signal Control - Q-Learning; Traffic Signal Control; Signal Control - Threshold Tuning; Traffic Light Control Algorithm; Adaptive Labor Staffing; Sleep-Wake Scheduling Algorithms; Reinforcement Learning; Vehicular Control; Graded Signal Control; Adaptive Sleep–wake Control; Computer Science

Record DetailsSimilar RecordsGoogle PlusoneFacebookTwitterCiteULikeMendeleyreddit

APA · Chicago · MLA · Vancouver · CSE | Export to Zotero / EndNote / Reference Manager

APA (6th Edition):

Prashanth, L. A. (2013). Resource Allocation for Sequential Decision Making Under Uncertainaty : Studies in Vehicular Traffic Control, Service Systems, Sensor Networks and Mechanism Design. (Thesis). Indian Institute of Science. Retrieved from http://hdl.handle.net/2005/2810

Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation

Chicago Manual of Style (16th Edition):

Prashanth, L A. “Resource Allocation for Sequential Decision Making Under Uncertainaty : Studies in Vehicular Traffic Control, Service Systems, Sensor Networks and Mechanism Design.” 2013. Thesis, Indian Institute of Science. Accessed October 22, 2019. http://hdl.handle.net/2005/2810.

Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation

MLA Handbook (7th Edition):

Prashanth, L A. “Resource Allocation for Sequential Decision Making Under Uncertainaty : Studies in Vehicular Traffic Control, Service Systems, Sensor Networks and Mechanism Design.” 2013. Web. 22 Oct 2019.

Vancouver:

Prashanth LA. Resource Allocation for Sequential Decision Making Under Uncertainaty : Studies in Vehicular Traffic Control, Service Systems, Sensor Networks and Mechanism Design. [Internet] [Thesis]. Indian Institute of Science; 2013. [cited 2019 Oct 22]. Available from: http://hdl.handle.net/2005/2810.

Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation

Council of Science Editors:

Prashanth LA. Resource Allocation for Sequential Decision Making Under Uncertainaty : Studies in Vehicular Traffic Control, Service Systems, Sensor Networks and Mechanism Design. [Thesis]. Indian Institute of Science; 2013. Available from: http://hdl.handle.net/2005/2810

Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation

21. Daugherty, Stephen Greyson. A Q-Learning Approach to Minefield Characterization from Unmanned Aerial Vehicles .

Degree: 2012, Duke University

  The treasure hunt problem to determine how a computational agent can maximize its ability to detect and/or classify multiple targets located in a region… (more)

Subjects/Keywords: Mechanical engineering; Computer science; Aircraft; Algorithm; Machine Learning; Q-Learning; Sensor

…optimal control policy will be. One problem with the basic Q-Learning algorithm, however, is… …A Q-Learning Approach to Minefield Characterization from Unmanned Aerial Vehicles by… …University 2012 Abstract (Intelligent Control Systems) A Q-Learning Approach to… …Bayesian Networks, Neural Networks, and Q-Learning 7 2.1 Bayesian Network Interpretation… …Network Training . . . . . . . . . . . . . . . . . . . . . . . . . 17 2.6 Q-Learning… 

Page 1 Page 2 Page 3 Page 4 Page 5 Sample image

Record DetailsSimilar RecordsGoogle PlusoneFacebookTwitterCiteULikeMendeleyreddit

APA · Chicago · MLA · Vancouver · CSE | Export to Zotero / EndNote / Reference Manager

APA (6th Edition):

Daugherty, S. G. (2012). A Q-Learning Approach to Minefield Characterization from Unmanned Aerial Vehicles . (Thesis). Duke University. Retrieved from http://hdl.handle.net/10161/5757

Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation

Chicago Manual of Style (16th Edition):

Daugherty, Stephen Greyson. “A Q-Learning Approach to Minefield Characterization from Unmanned Aerial Vehicles .” 2012. Thesis, Duke University. Accessed October 22, 2019. http://hdl.handle.net/10161/5757.

Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation

MLA Handbook (7th Edition):

Daugherty, Stephen Greyson. “A Q-Learning Approach to Minefield Characterization from Unmanned Aerial Vehicles .” 2012. Web. 22 Oct 2019.

Vancouver:

Daugherty SG. A Q-Learning Approach to Minefield Characterization from Unmanned Aerial Vehicles . [Internet] [Thesis]. Duke University; 2012. [cited 2019 Oct 22]. Available from: http://hdl.handle.net/10161/5757.

Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation

Council of Science Editors:

Daugherty SG. A Q-Learning Approach to Minefield Characterization from Unmanned Aerial Vehicles . [Thesis]. Duke University; 2012. Available from: http://hdl.handle.net/10161/5757

Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation


University of New South Wales

22. Kwok, Hing-Wah. Hierarchical reinforcement learning in adversarial environments.

Degree: Computer Science & Engineering, 2009, University of New South Wales

 It is known that one of the downfalls of reinforcement learning is the amount of time required to learn an optimal policy. This especially holds… (more)

Subjects/Keywords: Bowling's Win or Learn Fast (WoLF) algorithm; Q-Learning; Reinforcement learning; Thesis Digitisation Program

Record DetailsSimilar RecordsGoogle PlusoneFacebookTwitterCiteULikeMendeleyreddit

APA · Chicago · MLA · Vancouver · CSE | Export to Zotero / EndNote / Reference Manager

APA (6th Edition):

Kwok, H. (2009). Hierarchical reinforcement learning in adversarial environments. (Masters Thesis). University of New South Wales. Retrieved from http://handle.unsw.edu.au/1959.4/43424 ; https://unsworks.unsw.edu.au/fapi/datastream/unsworks:7097/SOURCE02?view=true

Chicago Manual of Style (16th Edition):

Kwok, Hing-Wah. “Hierarchical reinforcement learning in adversarial environments.” 2009. Masters Thesis, University of New South Wales. Accessed October 22, 2019. http://handle.unsw.edu.au/1959.4/43424 ; https://unsworks.unsw.edu.au/fapi/datastream/unsworks:7097/SOURCE02?view=true.

MLA Handbook (7th Edition):

Kwok, Hing-Wah. “Hierarchical reinforcement learning in adversarial environments.” 2009. Web. 22 Oct 2019.

Vancouver:

Kwok H. Hierarchical reinforcement learning in adversarial environments. [Internet] [Masters thesis]. University of New South Wales; 2009. [cited 2019 Oct 22]. Available from: http://handle.unsw.edu.au/1959.4/43424 ; https://unsworks.unsw.edu.au/fapi/datastream/unsworks:7097/SOURCE02?view=true.

Council of Science Editors:

Kwok H. Hierarchical reinforcement learning in adversarial environments. [Masters Thesis]. University of New South Wales; 2009. Available from: http://handle.unsw.edu.au/1959.4/43424 ; https://unsworks.unsw.edu.au/fapi/datastream/unsworks:7097/SOURCE02?view=true


NSYSU

23. Chung, Chi-Hsiu. Achieving Imitation-Based Learning for a Humanoid Robot by Evolutionary Computation.

Degree: Master, Information Management, 2009, NSYSU

 This thesis presents an imitation-based methodology, also a simple and easy way, for a service robot to learn the behaviors demonstrated by the user. With… (more)

Subjects/Keywords: Genetic Algorithm; Imitation Learning; Robot Learning

Record DetailsSimilar RecordsGoogle PlusoneFacebookTwitterCiteULikeMendeleyreddit

APA · Chicago · MLA · Vancouver · CSE | Export to Zotero / EndNote / Reference Manager

APA (6th Edition):

Chung, C. (2009). Achieving Imitation-Based Learning for a Humanoid Robot by Evolutionary Computation. (Thesis). NSYSU. Retrieved from http://etd.lib.nsysu.edu.tw/ETD-db/ETD-search/view_etd?URN=etd-0729109-002840

Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation

Chicago Manual of Style (16th Edition):

Chung, Chi-Hsiu. “Achieving Imitation-Based Learning for a Humanoid Robot by Evolutionary Computation.” 2009. Thesis, NSYSU. Accessed October 22, 2019. http://etd.lib.nsysu.edu.tw/ETD-db/ETD-search/view_etd?URN=etd-0729109-002840.

Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation

MLA Handbook (7th Edition):

Chung, Chi-Hsiu. “Achieving Imitation-Based Learning for a Humanoid Robot by Evolutionary Computation.” 2009. Web. 22 Oct 2019.

Vancouver:

Chung C. Achieving Imitation-Based Learning for a Humanoid Robot by Evolutionary Computation. [Internet] [Thesis]. NSYSU; 2009. [cited 2019 Oct 22]. Available from: http://etd.lib.nsysu.edu.tw/ETD-db/ETD-search/view_etd?URN=etd-0729109-002840.

Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation

Council of Science Editors:

Chung C. Achieving Imitation-Based Learning for a Humanoid Robot by Evolutionary Computation. [Thesis]. NSYSU; 2009. Available from: http://etd.lib.nsysu.edu.tw/ETD-db/ETD-search/view_etd?URN=etd-0729109-002840

Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation


Texas A&M University

24. Redwine, Tobin Dean. A Model of Global Learning: How Students Change Through International High-Impact Experiences.

Degree: 2014, Texas A&M University

 Study abroad programs are important, increasing, impactful, and influential in agricultural disciplines. Research suggests that students who participate are changed. However, we do not yet… (more)

Subjects/Keywords: Global Learning; Study Abroad; Q Methodology

Record DetailsSimilar RecordsGoogle PlusoneFacebookTwitterCiteULikeMendeleyreddit

APA · Chicago · MLA · Vancouver · CSE | Export to Zotero / EndNote / Reference Manager

APA (6th Edition):

Redwine, T. D. (2014). A Model of Global Learning: How Students Change Through International High-Impact Experiences. (Thesis). Texas A&M University. Retrieved from http://hdl.handle.net/1969.1/154044

Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation

Chicago Manual of Style (16th Edition):

Redwine, Tobin Dean. “A Model of Global Learning: How Students Change Through International High-Impact Experiences.” 2014. Thesis, Texas A&M University. Accessed October 22, 2019. http://hdl.handle.net/1969.1/154044.

Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation

MLA Handbook (7th Edition):

Redwine, Tobin Dean. “A Model of Global Learning: How Students Change Through International High-Impact Experiences.” 2014. Web. 22 Oct 2019.

Vancouver:

Redwine TD. A Model of Global Learning: How Students Change Through International High-Impact Experiences. [Internet] [Thesis]. Texas A&M University; 2014. [cited 2019 Oct 22]. Available from: http://hdl.handle.net/1969.1/154044.

Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation

Council of Science Editors:

Redwine TD. A Model of Global Learning: How Students Change Through International High-Impact Experiences. [Thesis]. Texas A&M University; 2014. Available from: http://hdl.handle.net/1969.1/154044

Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation


Universidade do Rio Grande do Norte

25. Silva, Lucileide Medeiros Dantas da. Proposta de arquitetura em Hardware para FPGA da técnica Qlearning de aprendizagem por reforço .

Degree: 2016, Universidade do Rio Grande do Norte

Q-learning is a off-policy reinforcement learning technique which has as main advantage the possibility of obtaining an optimal policy interacting with an unknown model environment.… (more)

Subjects/Keywords: FPGA; Q-learning; Aprendizagem por reforço; Hardware

Record DetailsSimilar RecordsGoogle PlusoneFacebookTwitterCiteULikeMendeleyreddit

APA · Chicago · MLA · Vancouver · CSE | Export to Zotero / EndNote / Reference Manager

APA (6th Edition):

Silva, L. M. D. d. (2016). Proposta de arquitetura em Hardware para FPGA da técnica Qlearning de aprendizagem por reforço . (Masters Thesis). Universidade do Rio Grande do Norte. Retrieved from http://repositorio.ufrn.br/handle/123456789/22395

Chicago Manual of Style (16th Edition):

Silva, Lucileide Medeiros Dantas da. “Proposta de arquitetura em Hardware para FPGA da técnica Qlearning de aprendizagem por reforço .” 2016. Masters Thesis, Universidade do Rio Grande do Norte. Accessed October 22, 2019. http://repositorio.ufrn.br/handle/123456789/22395.

MLA Handbook (7th Edition):

Silva, Lucileide Medeiros Dantas da. “Proposta de arquitetura em Hardware para FPGA da técnica Qlearning de aprendizagem por reforço .” 2016. Web. 22 Oct 2019.

Vancouver:

Silva LMDd. Proposta de arquitetura em Hardware para FPGA da técnica Qlearning de aprendizagem por reforço . [Internet] [Masters thesis]. Universidade do Rio Grande do Norte; 2016. [cited 2019 Oct 22]. Available from: http://repositorio.ufrn.br/handle/123456789/22395.

Council of Science Editors:

Silva LMDd. Proposta de arquitetura em Hardware para FPGA da técnica Qlearning de aprendizagem por reforço . [Masters Thesis]. Universidade do Rio Grande do Norte; 2016. Available from: http://repositorio.ufrn.br/handle/123456789/22395


Oklahoma State University

26. Banala, Lakshmi Prasad. Activation Mechanism in Robots.

Degree: Computer Science Department, 2009, Oklahoma State University

 In robotics, implementing strategies based on human body would help a robot to handle situations in a hostile environment where human intervention is not possible.… (more)

Subjects/Keywords: activation; information retrieval; q-learning; robot

Record DetailsSimilar RecordsGoogle PlusoneFacebookTwitterCiteULikeMendeleyreddit

APA · Chicago · MLA · Vancouver · CSE | Export to Zotero / EndNote / Reference Manager

APA (6th Edition):

Banala, L. P. (2009). Activation Mechanism in Robots. (Thesis). Oklahoma State University. Retrieved from http://hdl.handle.net/11244/8121

Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation

Chicago Manual of Style (16th Edition):

Banala, Lakshmi Prasad. “Activation Mechanism in Robots.” 2009. Thesis, Oklahoma State University. Accessed October 22, 2019. http://hdl.handle.net/11244/8121.

Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation

MLA Handbook (7th Edition):

Banala, Lakshmi Prasad. “Activation Mechanism in Robots.” 2009. Web. 22 Oct 2019.

Vancouver:

Banala LP. Activation Mechanism in Robots. [Internet] [Thesis]. Oklahoma State University; 2009. [cited 2019 Oct 22]. Available from: http://hdl.handle.net/11244/8121.

Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation

Council of Science Editors:

Banala LP. Activation Mechanism in Robots. [Thesis]. Oklahoma State University; 2009. Available from: http://hdl.handle.net/11244/8121

Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation


Brno University of Technology

27. Buchal, Petr. Hraní her pomocí neuronových sítí .

Degree: 2018, Brno University of Technology

 Cílem této práce je naučit neuronovou síť pohybu v prostředích s klasickou kontrolou řízení, hraní tahové hry 2048 a několika Atari her. Jedná se o… (more)

Subjects/Keywords: Strojové učení; Zpětnovazební učení; Neuronové sítě; Q-učení; Hluboké Q-učení; Machine learning; Reinforcement learning; Neural networks; Q-learning; Deep Q-learning

Record DetailsSimilar RecordsGoogle PlusoneFacebookTwitterCiteULikeMendeleyreddit

APA · Chicago · MLA · Vancouver · CSE | Export to Zotero / EndNote / Reference Manager

APA (6th Edition):

Buchal, P. (2018). Hraní her pomocí neuronových sítí . (Thesis). Brno University of Technology. Retrieved from http://hdl.handle.net/11012/85237

Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation

Chicago Manual of Style (16th Edition):

Buchal, Petr. “Hraní her pomocí neuronových sítí .” 2018. Thesis, Brno University of Technology. Accessed October 22, 2019. http://hdl.handle.net/11012/85237.

Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation

MLA Handbook (7th Edition):

Buchal, Petr. “Hraní her pomocí neuronových sítí .” 2018. Web. 22 Oct 2019.

Vancouver:

Buchal P. Hraní her pomocí neuronových sítí . [Internet] [Thesis]. Brno University of Technology; 2018. [cited 2019 Oct 22]. Available from: http://hdl.handle.net/11012/85237.

Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation

Council of Science Editors:

Buchal P. Hraní her pomocí neuronových sítí . [Thesis]. Brno University of Technology; 2018. Available from: http://hdl.handle.net/11012/85237

Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation


Brno University of Technology

28. Chábek, Lukáš. Posilované učení pro hraní hry Starcraft .

Degree: 2018, Brno University of Technology

 Práce se zabývá metodami strojového učení aplikovanými pro hraní strategických her v realném čase. V práci se zabývám metodou strojového učení Q-learning založenou na zpětnovazebním… (more)

Subjects/Keywords: Strojové učení; Umělá Intelience; Starcraft II; RTS; Q-Learning; Deep-Q-Learning; PySC2; Machine Learnig; Artificial Intelligence; Stracraft II; RTS; Q-Learning; Deep-Q-Learning; PySC2

Record DetailsSimilar RecordsGoogle PlusoneFacebookTwitterCiteULikeMendeleyreddit

APA · Chicago · MLA · Vancouver · CSE | Export to Zotero / EndNote / Reference Manager

APA (6th Edition):

Chábek, L. (2018). Posilované učení pro hraní hry Starcraft . (Thesis). Brno University of Technology. Retrieved from http://hdl.handle.net/11012/85144

Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation

Chicago Manual of Style (16th Edition):

Chábek, Lukáš. “Posilované učení pro hraní hry Starcraft .” 2018. Thesis, Brno University of Technology. Accessed October 22, 2019. http://hdl.handle.net/11012/85144.

Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation

MLA Handbook (7th Edition):

Chábek, Lukáš. “Posilované učení pro hraní hry Starcraft .” 2018. Web. 22 Oct 2019.

Vancouver:

Chábek L. Posilované učení pro hraní hry Starcraft . [Internet] [Thesis]. Brno University of Technology; 2018. [cited 2019 Oct 22]. Available from: http://hdl.handle.net/11012/85144.

Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation

Council of Science Editors:

Chábek L. Posilované učení pro hraní hry Starcraft . [Thesis]. Brno University of Technology; 2018. Available from: http://hdl.handle.net/11012/85144

Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation


University of Minnesota

29. Tseng, Kuo-Shih. Learning in Human and Robot Search: Subgoal, Submodularity, and Sparsity.

Degree: PhD, Computer Science, 2016, University of Minnesota

 Search is an essential technology for various robotic applications and it is also central to human daily activities. Searching for targets efficiently consists of NP-hard… (more)

Subjects/Keywords: Compressed sensing; Probabilistic search; Q-learning; Sparse learning; Structure learning; Submodularity

Record DetailsSimilar RecordsGoogle PlusoneFacebookTwitterCiteULikeMendeleyreddit

APA · Chicago · MLA · Vancouver · CSE | Export to Zotero / EndNote / Reference Manager

APA (6th Edition):

Tseng, K. (2016). Learning in Human and Robot Search: Subgoal, Submodularity, and Sparsity. (Doctoral Dissertation). University of Minnesota. Retrieved from http://hdl.handle.net/11299/183325

Chicago Manual of Style (16th Edition):

Tseng, Kuo-Shih. “Learning in Human and Robot Search: Subgoal, Submodularity, and Sparsity.” 2016. Doctoral Dissertation, University of Minnesota. Accessed October 22, 2019. http://hdl.handle.net/11299/183325.

MLA Handbook (7th Edition):

Tseng, Kuo-Shih. “Learning in Human and Robot Search: Subgoal, Submodularity, and Sparsity.” 2016. Web. 22 Oct 2019.

Vancouver:

Tseng K. Learning in Human and Robot Search: Subgoal, Submodularity, and Sparsity. [Internet] [Doctoral dissertation]. University of Minnesota; 2016. [cited 2019 Oct 22]. Available from: http://hdl.handle.net/11299/183325.

Council of Science Editors:

Tseng K. Learning in Human and Robot Search: Subgoal, Submodularity, and Sparsity. [Doctoral Dissertation]. University of Minnesota; 2016. Available from: http://hdl.handle.net/11299/183325


Colorado State University

30. Elliott, Daniel L. Wisdom of the crowd : reliable deep reinforcement learning through ensembles of Q-functions, The.

Degree: PhD, Computer Science, 2018, Colorado State University

 Reinforcement learning agents learn by exploring the environment and then exploiting what they have learned. This frees the human trainers from having to know the… (more)

Subjects/Keywords: machine learning; Q-learning; ensemble; reinforcement learning; neural networks

Record DetailsSimilar RecordsGoogle PlusoneFacebookTwitterCiteULikeMendeleyreddit

APA · Chicago · MLA · Vancouver · CSE | Export to Zotero / EndNote / Reference Manager

APA (6th Edition):

Elliott, D. L. (2018). Wisdom of the crowd : reliable deep reinforcement learning through ensembles of Q-functions, The. (Doctoral Dissertation). Colorado State University. Retrieved from http://hdl.handle.net/10217/191477

Chicago Manual of Style (16th Edition):

Elliott, Daniel L. “Wisdom of the crowd : reliable deep reinforcement learning through ensembles of Q-functions, The.” 2018. Doctoral Dissertation, Colorado State University. Accessed October 22, 2019. http://hdl.handle.net/10217/191477.

MLA Handbook (7th Edition):

Elliott, Daniel L. “Wisdom of the crowd : reliable deep reinforcement learning through ensembles of Q-functions, The.” 2018. Web. 22 Oct 2019.

Vancouver:

Elliott DL. Wisdom of the crowd : reliable deep reinforcement learning through ensembles of Q-functions, The. [Internet] [Doctoral dissertation]. Colorado State University; 2018. [cited 2019 Oct 22]. Available from: http://hdl.handle.net/10217/191477.

Council of Science Editors:

Elliott DL. Wisdom of the crowd : reliable deep reinforcement learning through ensembles of Q-functions, The. [Doctoral Dissertation]. Colorado State University; 2018. Available from: http://hdl.handle.net/10217/191477

[1] [2] [3] [4] [5] … [1888]

.