Advanced search options
You searched for subject:(Reinforcement Learning)
.
Showing records 1 – 30 of
1472 total matches.
◁ [1] [2] [3] [4] [5] … [50] ▶
Search Limiters
Dates
Universities
Department
Degrees
Country
▼ Search Limiters
Oregon State University
1. Proper, Scott. Scaling multiagent reinforcement learning.
Degree: PhD, Computer Science, 2009, Oregon State University
URL: http://hdl.handle.net/1957/13662
Subjects/Keywords: Reinforcement learning; Reinforcement learning
Record Details
Similar Records
❌
APA · Chicago · MLA · Vancouver · CSE | Export to Zotero / EndNote / Reference Manager
APA (6th Edition):
Proper, S. (2009). Scaling multiagent reinforcement learning. (Doctoral Dissertation). Oregon State University. Retrieved from http://hdl.handle.net/1957/13662
Chicago Manual of Style (16th Edition):
Proper, Scott. “Scaling multiagent reinforcement learning.” 2009. Doctoral Dissertation, Oregon State University. Accessed March 04, 2021. http://hdl.handle.net/1957/13662.
MLA Handbook (7th Edition):
Proper, Scott. “Scaling multiagent reinforcement learning.” 2009. Web. 04 Mar 2021.
Vancouver:
Proper S. Scaling multiagent reinforcement learning. [Internet] [Doctoral dissertation]. Oregon State University; 2009. [cited 2021 Mar 04]. Available from: http://hdl.handle.net/1957/13662.
Council of Science Editors:
Proper S. Scaling multiagent reinforcement learning. [Doctoral Dissertation]. Oregon State University; 2009. Available from: http://hdl.handle.net/1957/13662
Oregon State University
2. Mehta, Neville. Hierarchical structure discovery and transfer in sequential decision problems.
Degree: PhD, Computer Science, 2011, Oregon State University
URL: http://hdl.handle.net/1957/25199
Subjects/Keywords: hierarchical reinforcement learning; Reinforcement learning
Record Details
Similar Records
❌
APA · Chicago · MLA · Vancouver · CSE | Export to Zotero / EndNote / Reference Manager
APA (6th Edition):
Mehta, N. (2011). Hierarchical structure discovery and transfer in sequential decision problems. (Doctoral Dissertation). Oregon State University. Retrieved from http://hdl.handle.net/1957/25199
Chicago Manual of Style (16th Edition):
Mehta, Neville. “Hierarchical structure discovery and transfer in sequential decision problems.” 2011. Doctoral Dissertation, Oregon State University. Accessed March 04, 2021. http://hdl.handle.net/1957/25199.
MLA Handbook (7th Edition):
Mehta, Neville. “Hierarchical structure discovery and transfer in sequential decision problems.” 2011. Web. 04 Mar 2021.
Vancouver:
Mehta N. Hierarchical structure discovery and transfer in sequential decision problems. [Internet] [Doctoral dissertation]. Oregon State University; 2011. [cited 2021 Mar 04]. Available from: http://hdl.handle.net/1957/25199.
Council of Science Editors:
Mehta N. Hierarchical structure discovery and transfer in sequential decision problems. [Doctoral Dissertation]. Oregon State University; 2011. Available from: http://hdl.handle.net/1957/25199
Oregon State University
3.
Lauer, Christopher Joseph.
Determining optimal timber harvest and fuel treatment on a fire-threatened landscape using approximate dynamic programming.
Degree: PhD, 2017, Oregon State University
URL: http://hdl.handle.net/1957/61678
Subjects/Keywords: reinforcement learning
Record Details
Similar Records
❌
APA · Chicago · MLA · Vancouver · CSE | Export to Zotero / EndNote / Reference Manager
APA (6th Edition):
Lauer, C. J. (2017). Determining optimal timber harvest and fuel treatment on a fire-threatened landscape using approximate dynamic programming. (Doctoral Dissertation). Oregon State University. Retrieved from http://hdl.handle.net/1957/61678
Chicago Manual of Style (16th Edition):
Lauer, Christopher Joseph. “Determining optimal timber harvest and fuel treatment on a fire-threatened landscape using approximate dynamic programming.” 2017. Doctoral Dissertation, Oregon State University. Accessed March 04, 2021. http://hdl.handle.net/1957/61678.
MLA Handbook (7th Edition):
Lauer, Christopher Joseph. “Determining optimal timber harvest and fuel treatment on a fire-threatened landscape using approximate dynamic programming.” 2017. Web. 04 Mar 2021.
Vancouver:
Lauer CJ. Determining optimal timber harvest and fuel treatment on a fire-threatened landscape using approximate dynamic programming. [Internet] [Doctoral dissertation]. Oregon State University; 2017. [cited 2021 Mar 04]. Available from: http://hdl.handle.net/1957/61678.
Council of Science Editors:
Lauer CJ. Determining optimal timber harvest and fuel treatment on a fire-threatened landscape using approximate dynamic programming. [Doctoral Dissertation]. Oregon State University; 2017. Available from: http://hdl.handle.net/1957/61678
4. Frank, Mikhail Alexander. Learning to reach and reaching to learn: a unified approach to path planning and reactive control through reinforcement learning.
Degree: 2014, Università della Svizzera italiana
URL: http://doc.rero.ch/record/234387
Subjects/Keywords: Reinforcement learning
Record Details
Similar Records
❌
APA · Chicago · MLA · Vancouver · CSE | Export to Zotero / EndNote / Reference Manager
APA (6th Edition):
Frank, M. A. (2014). Learning to reach and reaching to learn: a unified approach to path planning and reactive control through reinforcement learning. (Thesis). Università della Svizzera italiana. Retrieved from http://doc.rero.ch/record/234387
Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation
Chicago Manual of Style (16th Edition):
Frank, Mikhail Alexander. “Learning to reach and reaching to learn: a unified approach to path planning and reactive control through reinforcement learning.” 2014. Thesis, Università della Svizzera italiana. Accessed March 04, 2021. http://doc.rero.ch/record/234387.
Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation
MLA Handbook (7th Edition):
Frank, Mikhail Alexander. “Learning to reach and reaching to learn: a unified approach to path planning and reactive control through reinforcement learning.” 2014. Web. 04 Mar 2021.
Vancouver:
Frank MA. Learning to reach and reaching to learn: a unified approach to path planning and reactive control through reinforcement learning. [Internet] [Thesis]. Università della Svizzera italiana; 2014. [cited 2021 Mar 04]. Available from: http://doc.rero.ch/record/234387.
Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation
Council of Science Editors:
Frank MA. Learning to reach and reaching to learn: a unified approach to path planning and reactive control through reinforcement learning. [Thesis]. Università della Svizzera italiana; 2014. Available from: http://doc.rero.ch/record/234387
Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation
Delft University of Technology
5. Tian, Yuan (author). Model Free Reinforcement Learning with Stability Guarantee.
Degree: 2019, Delft University of Technology
URL: http://resolver.tudelft.nl/uuid:dde4e58f-e109-4e7f-8ecb-ed1734294e5c
Subjects/Keywords: Reinforcement Learning
Record Details
Similar Records
❌
APA · Chicago · MLA · Vancouver · CSE | Export to Zotero / EndNote / Reference Manager
APA (6th Edition):
Tian, Y. (. (2019). Model Free Reinforcement Learning with Stability Guarantee. (Masters Thesis). Delft University of Technology. Retrieved from http://resolver.tudelft.nl/uuid:dde4e58f-e109-4e7f-8ecb-ed1734294e5c
Chicago Manual of Style (16th Edition):
Tian, Yuan (author). “Model Free Reinforcement Learning with Stability Guarantee.” 2019. Masters Thesis, Delft University of Technology. Accessed March 04, 2021. http://resolver.tudelft.nl/uuid:dde4e58f-e109-4e7f-8ecb-ed1734294e5c.
MLA Handbook (7th Edition):
Tian, Yuan (author). “Model Free Reinforcement Learning with Stability Guarantee.” 2019. Web. 04 Mar 2021.
Vancouver:
Tian Y(. Model Free Reinforcement Learning with Stability Guarantee. [Internet] [Masters thesis]. Delft University of Technology; 2019. [cited 2021 Mar 04]. Available from: http://resolver.tudelft.nl/uuid:dde4e58f-e109-4e7f-8ecb-ed1734294e5c.
Council of Science Editors:
Tian Y(. Model Free Reinforcement Learning with Stability Guarantee. [Masters Thesis]. Delft University of Technology; 2019. Available from: http://resolver.tudelft.nl/uuid:dde4e58f-e109-4e7f-8ecb-ed1734294e5c
Delft University of Technology
6. Van Rooijen, J.C. (author). Learning Parameter Selection in Continuous Reinforcement Learning: Attempting to Reduce Tuning Effords.
Degree: 2012, Delft University of Technology
URL: http://resolver.tudelft.nl/uuid:94b81bc2-aff6-457f-9b54-be5e005def38
Subjects/Keywords: reinforcement learning
Record Details
Similar Records
❌
APA · Chicago · MLA · Vancouver · CSE | Export to Zotero / EndNote / Reference Manager
APA (6th Edition):
Van Rooijen, J. C. (. (2012). Learning Parameter Selection in Continuous Reinforcement Learning: Attempting to Reduce Tuning Effords. (Masters Thesis). Delft University of Technology. Retrieved from http://resolver.tudelft.nl/uuid:94b81bc2-aff6-457f-9b54-be5e005def38
Chicago Manual of Style (16th Edition):
Van Rooijen, J C (author). “Learning Parameter Selection in Continuous Reinforcement Learning: Attempting to Reduce Tuning Effords.” 2012. Masters Thesis, Delft University of Technology. Accessed March 04, 2021. http://resolver.tudelft.nl/uuid:94b81bc2-aff6-457f-9b54-be5e005def38.
MLA Handbook (7th Edition):
Van Rooijen, J C (author). “Learning Parameter Selection in Continuous Reinforcement Learning: Attempting to Reduce Tuning Effords.” 2012. Web. 04 Mar 2021.
Vancouver:
Van Rooijen JC(. Learning Parameter Selection in Continuous Reinforcement Learning: Attempting to Reduce Tuning Effords. [Internet] [Masters thesis]. Delft University of Technology; 2012. [cited 2021 Mar 04]. Available from: http://resolver.tudelft.nl/uuid:94b81bc2-aff6-457f-9b54-be5e005def38.
Council of Science Editors:
Van Rooijen JC(. Learning Parameter Selection in Continuous Reinforcement Learning: Attempting to Reduce Tuning Effords. [Masters Thesis]. Delft University of Technology; 2012. Available from: http://resolver.tudelft.nl/uuid:94b81bc2-aff6-457f-9b54-be5e005def38
Delft University of Technology
7. Van Diepen, M.D.M. (author). Avoiding failure states during reinforcement learning.
Degree: 2011, Delft University of Technology
URL: http://resolver.tudelft.nl/uuid:1f03c580-9fd5-4807-87b5-d70890e05ff6
Subjects/Keywords: reinforcement learning
Record Details
Similar Records
❌
APA · Chicago · MLA · Vancouver · CSE | Export to Zotero / EndNote / Reference Manager
APA (6th Edition):
Van Diepen, M. D. M. (. (2011). Avoiding failure states during reinforcement learning. (Masters Thesis). Delft University of Technology. Retrieved from http://resolver.tudelft.nl/uuid:1f03c580-9fd5-4807-87b5-d70890e05ff6
Chicago Manual of Style (16th Edition):
Van Diepen, M D M (author). “Avoiding failure states during reinforcement learning.” 2011. Masters Thesis, Delft University of Technology. Accessed March 04, 2021. http://resolver.tudelft.nl/uuid:1f03c580-9fd5-4807-87b5-d70890e05ff6.
MLA Handbook (7th Edition):
Van Diepen, M D M (author). “Avoiding failure states during reinforcement learning.” 2011. Web. 04 Mar 2021.
Vancouver:
Van Diepen MDM(. Avoiding failure states during reinforcement learning. [Internet] [Masters thesis]. Delft University of Technology; 2011. [cited 2021 Mar 04]. Available from: http://resolver.tudelft.nl/uuid:1f03c580-9fd5-4807-87b5-d70890e05ff6.
Council of Science Editors:
Van Diepen MDM(. Avoiding failure states during reinforcement learning. [Masters Thesis]. Delft University of Technology; 2011. Available from: http://resolver.tudelft.nl/uuid:1f03c580-9fd5-4807-87b5-d70890e05ff6
University of Illinois – Urbana-Champaign
8. Potok, Matthew. Safe reinforcement learning: An overview, a hybrid systems perspective, and a case study.
Degree: MS, Electrical & Computer Engr, 2018, University of Illinois – Urbana-Champaign
URL: http://hdl.handle.net/2142/102518
Subjects/Keywords: Reinforcement Learning
Record Details
Similar Records
❌
APA · Chicago · MLA · Vancouver · CSE | Export to Zotero / EndNote / Reference Manager
APA (6th Edition):
Potok, M. (2018). Safe reinforcement learning: An overview, a hybrid systems perspective, and a case study. (Thesis). University of Illinois – Urbana-Champaign. Retrieved from http://hdl.handle.net/2142/102518
Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation
Chicago Manual of Style (16th Edition):
Potok, Matthew. “Safe reinforcement learning: An overview, a hybrid systems perspective, and a case study.” 2018. Thesis, University of Illinois – Urbana-Champaign. Accessed March 04, 2021. http://hdl.handle.net/2142/102518.
Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation
MLA Handbook (7th Edition):
Potok, Matthew. “Safe reinforcement learning: An overview, a hybrid systems perspective, and a case study.” 2018. Web. 04 Mar 2021.
Vancouver:
Potok M. Safe reinforcement learning: An overview, a hybrid systems perspective, and a case study. [Internet] [Thesis]. University of Illinois – Urbana-Champaign; 2018. [cited 2021 Mar 04]. Available from: http://hdl.handle.net/2142/102518.
Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation
Council of Science Editors:
Potok M. Safe reinforcement learning: An overview, a hybrid systems perspective, and a case study. [Thesis]. University of Illinois – Urbana-Champaign; 2018. Available from: http://hdl.handle.net/2142/102518
Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation
Delft University of Technology
9. Keulen, Bart (author). Smart Start: A Directed and Persistent Exploration Framework for Reinforcement Learning.
Degree: 2018, Delft University of Technology
URL: http://resolver.tudelft.nl/uuid:eca20454-7905-42a1-9fb0-f72776fd5422
Subjects/Keywords: Reinforcement Learning; Exploration
Record Details
Similar Records
❌
APA · Chicago · MLA · Vancouver · CSE | Export to Zotero / EndNote / Reference Manager
APA (6th Edition):
Keulen, B. (. (2018). Smart Start: A Directed and Persistent Exploration Framework for Reinforcement Learning. (Masters Thesis). Delft University of Technology. Retrieved from http://resolver.tudelft.nl/uuid:eca20454-7905-42a1-9fb0-f72776fd5422
Chicago Manual of Style (16th Edition):
Keulen, Bart (author). “Smart Start: A Directed and Persistent Exploration Framework for Reinforcement Learning.” 2018. Masters Thesis, Delft University of Technology. Accessed March 04, 2021. http://resolver.tudelft.nl/uuid:eca20454-7905-42a1-9fb0-f72776fd5422.
MLA Handbook (7th Edition):
Keulen, Bart (author). “Smart Start: A Directed and Persistent Exploration Framework for Reinforcement Learning.” 2018. Web. 04 Mar 2021.
Vancouver:
Keulen B(. Smart Start: A Directed and Persistent Exploration Framework for Reinforcement Learning. [Internet] [Masters thesis]. Delft University of Technology; 2018. [cited 2021 Mar 04]. Available from: http://resolver.tudelft.nl/uuid:eca20454-7905-42a1-9fb0-f72776fd5422.
Council of Science Editors:
Keulen B(. Smart Start: A Directed and Persistent Exploration Framework for Reinforcement Learning. [Masters Thesis]. Delft University of Technology; 2018. Available from: http://resolver.tudelft.nl/uuid:eca20454-7905-42a1-9fb0-f72776fd5422
University of Illinois – Urbana-Champaign
10. Van Stralen, Neale A. Hierarchical reinforcement learning for adaptive and autonomous decision-making in robotics.
Degree: MS, Aerospace Engineering, 2020, University of Illinois – Urbana-Champaign
URL: http://hdl.handle.net/2142/108536
Subjects/Keywords: Reinforcement Learning; Adaptation
Record Details
Similar Records
❌
APA · Chicago · MLA · Vancouver · CSE | Export to Zotero / EndNote / Reference Manager
APA (6th Edition):
Van Stralen, N. A. (2020). Hierarchical reinforcement learning for adaptive and autonomous decision-making in robotics. (Thesis). University of Illinois – Urbana-Champaign. Retrieved from http://hdl.handle.net/2142/108536
Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation
Chicago Manual of Style (16th Edition):
Van Stralen, Neale A. “Hierarchical reinforcement learning for adaptive and autonomous decision-making in robotics.” 2020. Thesis, University of Illinois – Urbana-Champaign. Accessed March 04, 2021. http://hdl.handle.net/2142/108536.
Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation
MLA Handbook (7th Edition):
Van Stralen, Neale A. “Hierarchical reinforcement learning for adaptive and autonomous decision-making in robotics.” 2020. Web. 04 Mar 2021.
Vancouver:
Van Stralen NA. Hierarchical reinforcement learning for adaptive and autonomous decision-making in robotics. [Internet] [Thesis]. University of Illinois – Urbana-Champaign; 2020. [cited 2021 Mar 04]. Available from: http://hdl.handle.net/2142/108536.
Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation
Council of Science Editors:
Van Stralen NA. Hierarchical reinforcement learning for adaptive and autonomous decision-making in robotics. [Thesis]. University of Illinois – Urbana-Champaign; 2020. Available from: http://hdl.handle.net/2142/108536
Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation
University of New South Wales
11.
Ismail, Hafsa.
A neural network framework for combining different task types and motivations in motivated reinforcement learning.
Degree: Engineering & Information Technology, 2014, University of New South Wales
URL: http://handle.unsw.edu.au/1959.4/53975
;
https://unsworks.unsw.edu.au/fapi/datastream/unsworks:12686/SOURCE02?view=true
Subjects/Keywords: Motivated Reinforcement Learning
Record Details
Similar Records
❌
APA · Chicago · MLA · Vancouver · CSE | Export to Zotero / EndNote / Reference Manager
APA (6th Edition):
Ismail, H. (2014). A neural network framework for combining different task types and motivations in motivated reinforcement learning. (Masters Thesis). University of New South Wales. Retrieved from http://handle.unsw.edu.au/1959.4/53975 ; https://unsworks.unsw.edu.au/fapi/datastream/unsworks:12686/SOURCE02?view=true
Chicago Manual of Style (16th Edition):
Ismail, Hafsa. “A neural network framework for combining different task types and motivations in motivated reinforcement learning.” 2014. Masters Thesis, University of New South Wales. Accessed March 04, 2021. http://handle.unsw.edu.au/1959.4/53975 ; https://unsworks.unsw.edu.au/fapi/datastream/unsworks:12686/SOURCE02?view=true.
MLA Handbook (7th Edition):
Ismail, Hafsa. “A neural network framework for combining different task types and motivations in motivated reinforcement learning.” 2014. Web. 04 Mar 2021.
Vancouver:
Ismail H. A neural network framework for combining different task types and motivations in motivated reinforcement learning. [Internet] [Masters thesis]. University of New South Wales; 2014. [cited 2021 Mar 04]. Available from: http://handle.unsw.edu.au/1959.4/53975 ; https://unsworks.unsw.edu.au/fapi/datastream/unsworks:12686/SOURCE02?view=true.
Council of Science Editors:
Ismail H. A neural network framework for combining different task types and motivations in motivated reinforcement learning. [Masters Thesis]. University of New South Wales; 2014. Available from: http://handle.unsw.edu.au/1959.4/53975 ; https://unsworks.unsw.edu.au/fapi/datastream/unsworks:12686/SOURCE02?view=true
Oregon State University
12. Wilson, Aaron (Aaron Creighton). Bayesian methods for knowledge transfer and policy search in reinforcement learning.
Degree: PhD, Computer Science, 2012, Oregon State University
URL: http://hdl.handle.net/1957/34550
Subjects/Keywords: Machine Learning; Reinforcement learning
Record Details
Similar Records
❌
APA · Chicago · MLA · Vancouver · CSE | Export to Zotero / EndNote / Reference Manager
APA (6th Edition):
Wilson, A. (. C. (2012). Bayesian methods for knowledge transfer and policy search in reinforcement learning. (Doctoral Dissertation). Oregon State University. Retrieved from http://hdl.handle.net/1957/34550
Chicago Manual of Style (16th Edition):
Wilson, Aaron (Aaron Creighton). “Bayesian methods for knowledge transfer and policy search in reinforcement learning.” 2012. Doctoral Dissertation, Oregon State University. Accessed March 04, 2021. http://hdl.handle.net/1957/34550.
MLA Handbook (7th Edition):
Wilson, Aaron (Aaron Creighton). “Bayesian methods for knowledge transfer and policy search in reinforcement learning.” 2012. Web. 04 Mar 2021.
Vancouver:
Wilson A(C. Bayesian methods for knowledge transfer and policy search in reinforcement learning. [Internet] [Doctoral dissertation]. Oregon State University; 2012. [cited 2021 Mar 04]. Available from: http://hdl.handle.net/1957/34550.
Council of Science Editors:
Wilson A(C. Bayesian methods for knowledge transfer and policy search in reinforcement learning. [Doctoral Dissertation]. Oregon State University; 2012. Available from: http://hdl.handle.net/1957/34550
Rutgers University
13. Marivate, Vukosi N. Improved empirical methods in reinforcement-learning evaluation.
Degree: PhD, Computer Science, 2015, Rutgers University
URL: https://rucore.libraries.rutgers.edu/rutgers-lib/46389/
Subjects/Keywords: Reinforcement learning; Machine learning; Algorithms
Record Details
Similar Records
❌
APA · Chicago · MLA · Vancouver · CSE | Export to Zotero / EndNote / Reference Manager
APA (6th Edition):
Marivate, V. N. (2015). Improved empirical methods in reinforcement-learning evaluation. (Doctoral Dissertation). Rutgers University. Retrieved from https://rucore.libraries.rutgers.edu/rutgers-lib/46389/
Chicago Manual of Style (16th Edition):
Marivate, Vukosi N. “Improved empirical methods in reinforcement-learning evaluation.” 2015. Doctoral Dissertation, Rutgers University. Accessed March 04, 2021. https://rucore.libraries.rutgers.edu/rutgers-lib/46389/.
MLA Handbook (7th Edition):
Marivate, Vukosi N. “Improved empirical methods in reinforcement-learning evaluation.” 2015. Web. 04 Mar 2021.
Vancouver:
Marivate VN. Improved empirical methods in reinforcement-learning evaluation. [Internet] [Doctoral dissertation]. Rutgers University; 2015. [cited 2021 Mar 04]. Available from: https://rucore.libraries.rutgers.edu/rutgers-lib/46389/.
Council of Science Editors:
Marivate VN. Improved empirical methods in reinforcement-learning evaluation. [Doctoral Dissertation]. Rutgers University; 2015. Available from: https://rucore.libraries.rutgers.edu/rutgers-lib/46389/
University of Aberdeen
14. Alexander, John W. Transfer in reinforcement learning.
Degree: PhD, 2015, University of Aberdeen
URL: https://abdn.alma.exlibrisgroup.com/view/delivery/44ABE_INST/12152715040005941
;
https://ethos.bl.uk/OrderDetails.do?uin=uk.bl.ethos.675561
Subjects/Keywords: 004; Reinforcement learning; Learning
Record Details
Similar Records
❌
APA · Chicago · MLA · Vancouver · CSE | Export to Zotero / EndNote / Reference Manager
APA (6th Edition):
Alexander, J. W. (2015). Transfer in reinforcement learning. (Doctoral Dissertation). University of Aberdeen. Retrieved from https://abdn.alma.exlibrisgroup.com/view/delivery/44ABE_INST/12152715040005941 ; https://ethos.bl.uk/OrderDetails.do?uin=uk.bl.ethos.675561
Chicago Manual of Style (16th Edition):
Alexander, John W. “Transfer in reinforcement learning.” 2015. Doctoral Dissertation, University of Aberdeen. Accessed March 04, 2021. https://abdn.alma.exlibrisgroup.com/view/delivery/44ABE_INST/12152715040005941 ; https://ethos.bl.uk/OrderDetails.do?uin=uk.bl.ethos.675561.
MLA Handbook (7th Edition):
Alexander, John W. “Transfer in reinforcement learning.” 2015. Web. 04 Mar 2021.
Vancouver:
Alexander JW. Transfer in reinforcement learning. [Internet] [Doctoral dissertation]. University of Aberdeen; 2015. [cited 2021 Mar 04]. Available from: https://abdn.alma.exlibrisgroup.com/view/delivery/44ABE_INST/12152715040005941 ; https://ethos.bl.uk/OrderDetails.do?uin=uk.bl.ethos.675561.
Council of Science Editors:
Alexander JW. Transfer in reinforcement learning. [Doctoral Dissertation]. University of Aberdeen; 2015. Available from: https://abdn.alma.exlibrisgroup.com/view/delivery/44ABE_INST/12152715040005941 ; https://ethos.bl.uk/OrderDetails.do?uin=uk.bl.ethos.675561
University of Waterloo
15. Gaurav, Ashish. Safety-Oriented Stability Biases for Continual Learning.
Degree: 2020, University of Waterloo
URL: http://hdl.handle.net/10012/15579
Subjects/Keywords: deep learning; continual learning; classification; reinforcement learning
Record Details
Similar Records
❌
APA · Chicago · MLA · Vancouver · CSE | Export to Zotero / EndNote / Reference Manager
APA (6th Edition):
Gaurav, A. (2020). Safety-Oriented Stability Biases for Continual Learning. (Thesis). University of Waterloo. Retrieved from http://hdl.handle.net/10012/15579
Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation
Chicago Manual of Style (16th Edition):
Gaurav, Ashish. “Safety-Oriented Stability Biases for Continual Learning.” 2020. Thesis, University of Waterloo. Accessed March 04, 2021. http://hdl.handle.net/10012/15579.
Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation
MLA Handbook (7th Edition):
Gaurav, Ashish. “Safety-Oriented Stability Biases for Continual Learning.” 2020. Web. 04 Mar 2021.
Vancouver:
Gaurav A. Safety-Oriented Stability Biases for Continual Learning. [Internet] [Thesis]. University of Waterloo; 2020. [cited 2021 Mar 04]. Available from: http://hdl.handle.net/10012/15579.
Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation
Council of Science Editors:
Gaurav A. Safety-Oriented Stability Biases for Continual Learning. [Thesis]. University of Waterloo; 2020. Available from: http://hdl.handle.net/10012/15579
Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation
Hong Kong University of Science and Technology
16. Wellmer, Zachary William CSE. Building and leveraging implicit models for policy gradient methods.
Degree: 2019, Hong Kong University of Science and Technology
URL: http://repository.ust.hk/ir/Record/1783.1-102375
;
https://doi.org/10.14711/thesis-991012757568703412
;
http://repository.ust.hk/ir/bitstream/1783.1-102375/1/th_redirect.html
Subjects/Keywords: Reinforcement learning ; Machine learning ; Implicit learning
Record Details
Similar Records
❌
APA · Chicago · MLA · Vancouver · CSE | Export to Zotero / EndNote / Reference Manager
APA (6th Edition):
Wellmer, Z. W. C. (2019). Building and leveraging implicit models for policy gradient methods. (Thesis). Hong Kong University of Science and Technology. Retrieved from http://repository.ust.hk/ir/Record/1783.1-102375 ; https://doi.org/10.14711/thesis-991012757568703412 ; http://repository.ust.hk/ir/bitstream/1783.1-102375/1/th_redirect.html
Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation
Chicago Manual of Style (16th Edition):
Wellmer, Zachary William CSE. “Building and leveraging implicit models for policy gradient methods.” 2019. Thesis, Hong Kong University of Science and Technology. Accessed March 04, 2021. http://repository.ust.hk/ir/Record/1783.1-102375 ; https://doi.org/10.14711/thesis-991012757568703412 ; http://repository.ust.hk/ir/bitstream/1783.1-102375/1/th_redirect.html.
Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation
MLA Handbook (7th Edition):
Wellmer, Zachary William CSE. “Building and leveraging implicit models for policy gradient methods.” 2019. Web. 04 Mar 2021.
Vancouver:
Wellmer ZWC. Building and leveraging implicit models for policy gradient methods. [Internet] [Thesis]. Hong Kong University of Science and Technology; 2019. [cited 2021 Mar 04]. Available from: http://repository.ust.hk/ir/Record/1783.1-102375 ; https://doi.org/10.14711/thesis-991012757568703412 ; http://repository.ust.hk/ir/bitstream/1783.1-102375/1/th_redirect.html.
Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation
Council of Science Editors:
Wellmer ZWC. Building and leveraging implicit models for policy gradient methods. [Thesis]. Hong Kong University of Science and Technology; 2019. Available from: http://repository.ust.hk/ir/Record/1783.1-102375 ; https://doi.org/10.14711/thesis-991012757568703412 ; http://repository.ust.hk/ir/bitstream/1783.1-102375/1/th_redirect.html
Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation
University of Texas – Austin
17. Shah, Rishi Alpesh. Deep R learning for continual area sweeping.
Degree: MSin Computational Science, Engineering, and Mathematics, Computational Science, Engineering, and Mathematics, 2019, University of Texas – Austin
URL: http://dx.doi.org/10.26153/tsw/8201
Subjects/Keywords: Machine learning; Reinforcement learning; Robotics; Robot learning
Record Details
Similar Records
❌
APA · Chicago · MLA · Vancouver · CSE | Export to Zotero / EndNote / Reference Manager
APA (6th Edition):
Shah, R. A. (2019). Deep R learning for continual area sweeping. (Masters Thesis). University of Texas – Austin. Retrieved from http://dx.doi.org/10.26153/tsw/8201
Chicago Manual of Style (16th Edition):
Shah, Rishi Alpesh. “Deep R learning for continual area sweeping.” 2019. Masters Thesis, University of Texas – Austin. Accessed March 04, 2021. http://dx.doi.org/10.26153/tsw/8201.
MLA Handbook (7th Edition):
Shah, Rishi Alpesh. “Deep R learning for continual area sweeping.” 2019. Web. 04 Mar 2021.
Vancouver:
Shah RA. Deep R learning for continual area sweeping. [Internet] [Masters thesis]. University of Texas – Austin; 2019. [cited 2021 Mar 04]. Available from: http://dx.doi.org/10.26153/tsw/8201.
Council of Science Editors:
Shah RA. Deep R learning for continual area sweeping. [Masters Thesis]. University of Texas – Austin; 2019. Available from: http://dx.doi.org/10.26153/tsw/8201
University of Illinois – Urbana-Champaign
18. Benson, Christopher Edward. Improving cache replacement policy using deep reinforcement learning.
Degree: MS, Computer Science, 2018, University of Illinois – Urbana-Champaign
URL: http://hdl.handle.net/2142/102858
Subjects/Keywords: Reinforcement Learning; Machine Learning; Deep Learning
Record Details
Similar Records
❌
APA · Chicago · MLA · Vancouver · CSE | Export to Zotero / EndNote / Reference Manager
APA (6th Edition):
Benson, C. E. (2018). Improving cache replacement policy using deep reinforcement learning. (Thesis). University of Illinois – Urbana-Champaign. Retrieved from http://hdl.handle.net/2142/102858
Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation
Chicago Manual of Style (16th Edition):
Benson, Christopher Edward. “Improving cache replacement policy using deep reinforcement learning.” 2018. Thesis, University of Illinois – Urbana-Champaign. Accessed March 04, 2021. http://hdl.handle.net/2142/102858.
Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation
MLA Handbook (7th Edition):
Benson, Christopher Edward. “Improving cache replacement policy using deep reinforcement learning.” 2018. Web. 04 Mar 2021.
Vancouver:
Benson CE. Improving cache replacement policy using deep reinforcement learning. [Internet] [Thesis]. University of Illinois – Urbana-Champaign; 2018. [cited 2021 Mar 04]. Available from: http://hdl.handle.net/2142/102858.
Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation
Council of Science Editors:
Benson CE. Improving cache replacement policy using deep reinforcement learning. [Thesis]. University of Illinois – Urbana-Champaign; 2018. Available from: http://hdl.handle.net/2142/102858
Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation
Delft University of Technology
19. Kovács, B. (author). Safe reinforcement learning in long-horizon partially observable environments.
Degree: 2020, Delft University of Technology
URL: http://resolver.tudelft.nl/uuid:83556b37-1484-41f4-b168-3bd316def4a0
Subjects/Keywords: reinforcement learning; self-attention; partially observable; deep reinforcement learning; safe reinforcement learning
Record Details
Similar Records
❌
APA · Chicago · MLA · Vancouver · CSE | Export to Zotero / EndNote / Reference Manager
APA (6th Edition):
Kovács, B. (. (2020). Safe reinforcement learning in long-horizon partially observable environments. (Masters Thesis). Delft University of Technology. Retrieved from http://resolver.tudelft.nl/uuid:83556b37-1484-41f4-b168-3bd316def4a0
Chicago Manual of Style (16th Edition):
Kovács, B (author). “Safe reinforcement learning in long-horizon partially observable environments.” 2020. Masters Thesis, Delft University of Technology. Accessed March 04, 2021. http://resolver.tudelft.nl/uuid:83556b37-1484-41f4-b168-3bd316def4a0.
MLA Handbook (7th Edition):
Kovács, B (author). “Safe reinforcement learning in long-horizon partially observable environments.” 2020. Web. 04 Mar 2021.
Vancouver:
Kovács B(. Safe reinforcement learning in long-horizon partially observable environments. [Internet] [Masters thesis]. Delft University of Technology; 2020. [cited 2021 Mar 04]. Available from: http://resolver.tudelft.nl/uuid:83556b37-1484-41f4-b168-3bd316def4a0.
Council of Science Editors:
Kovács B(. Safe reinforcement learning in long-horizon partially observable environments. [Masters Thesis]. Delft University of Technology; 2020. Available from: http://resolver.tudelft.nl/uuid:83556b37-1484-41f4-b168-3bd316def4a0
NSYSU
20. Lin, Kun-da. Deep Reinforcement Learning with a Gating Network.
Degree: Master, Electrical Engineering, 2017, NSYSU
URL: http://etd.lib.nsysu.edu.tw/ETD-db/ETD-search/view_etd?URN=etd-0223117-131536
Subjects/Keywords: Reinforcement Learning; Deep Reinforcement Learning; Deep Learning; Gating network; Neural network
Record Details
Similar Records
❌
APA · Chicago · MLA · Vancouver · CSE | Export to Zotero / EndNote / Reference Manager
APA (6th Edition):
Lin, K. (2017). Deep Reinforcement Learning with a Gating Network. (Thesis). NSYSU. Retrieved from http://etd.lib.nsysu.edu.tw/ETD-db/ETD-search/view_etd?URN=etd-0223117-131536
Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation
Chicago Manual of Style (16th Edition):
Lin, Kun-da. “Deep Reinforcement Learning with a Gating Network.” 2017. Thesis, NSYSU. Accessed March 04, 2021. http://etd.lib.nsysu.edu.tw/ETD-db/ETD-search/view_etd?URN=etd-0223117-131536.
Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation
MLA Handbook (7th Edition):
Lin, Kun-da. “Deep Reinforcement Learning with a Gating Network.” 2017. Web. 04 Mar 2021.
Vancouver:
Lin K. Deep Reinforcement Learning with a Gating Network. [Internet] [Thesis]. NSYSU; 2017. [cited 2021 Mar 04]. Available from: http://etd.lib.nsysu.edu.tw/ETD-db/ETD-search/view_etd?URN=etd-0223117-131536.
Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation
Council of Science Editors:
Lin K. Deep Reinforcement Learning with a Gating Network. [Thesis]. NSYSU; 2017. Available from: http://etd.lib.nsysu.edu.tw/ETD-db/ETD-search/view_etd?URN=etd-0223117-131536
Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation
NSYSU
21. Tseng, Yi-Chia. An Unified Approach to Inverse Reinforcement Learning by Oppositive Demonstrations.
Degree: Master, Electrical Engineering, 2015, NSYSU
URL: http://etd.lib.nsysu.edu.tw/ETD-db/ETD-search/view_etd?URN=etd-0727115-130716
Subjects/Keywords: Apprenticeship Learning; Feature weight; Inverse Reinforcement learning; Reward function; Reinforcement learning
Record Details
Similar Records
❌
APA · Chicago · MLA · Vancouver · CSE | Export to Zotero / EndNote / Reference Manager
APA (6th Edition):
Tseng, Y. (2015). An Unified Approach to Inverse Reinforcement Learning by Oppositive Demonstrations. (Thesis). NSYSU. Retrieved from http://etd.lib.nsysu.edu.tw/ETD-db/ETD-search/view_etd?URN=etd-0727115-130716
Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation
Chicago Manual of Style (16th Edition):
Tseng, Yi-Chia. “An Unified Approach to Inverse Reinforcement Learning by Oppositive Demonstrations.” 2015. Thesis, NSYSU. Accessed March 04, 2021. http://etd.lib.nsysu.edu.tw/ETD-db/ETD-search/view_etd?URN=etd-0727115-130716.
Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation
MLA Handbook (7th Edition):
Tseng, Yi-Chia. “An Unified Approach to Inverse Reinforcement Learning by Oppositive Demonstrations.” 2015. Web. 04 Mar 2021.
Vancouver:
Tseng Y. An Unified Approach to Inverse Reinforcement Learning by Oppositive Demonstrations. [Internet] [Thesis]. NSYSU; 2015. [cited 2021 Mar 04]. Available from: http://etd.lib.nsysu.edu.tw/ETD-db/ETD-search/view_etd?URN=etd-0727115-130716.
Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation
Council of Science Editors:
Tseng Y. An Unified Approach to Inverse Reinforcement Learning by Oppositive Demonstrations. [Thesis]. NSYSU; 2015. Available from: http://etd.lib.nsysu.edu.tw/ETD-db/ETD-search/view_etd?URN=etd-0727115-130716
Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation
NSYSU
22. Lin, Hung-shyuan. Applying The Concept of Fuzzy Logic to Inverse Reinforcement Learning.
Degree: Master, Electrical Engineering, 2015, NSYSU
URL: http://etd.lib.nsysu.edu.tw/ETD-db/ETD-search/view_etd?URN=etd-1025115-185021
Subjects/Keywords: Inverse reinforcement learning; Reward function; Fuzzy; Reinforcement learning; AdaBoost; Apprenticeship learning
Record Details
Similar Records
❌
APA · Chicago · MLA · Vancouver · CSE | Export to Zotero / EndNote / Reference Manager
APA (6th Edition):
Lin, H. (2015). Applying The Concept of Fuzzy Logic to Inverse Reinforcement Learning. (Thesis). NSYSU. Retrieved from http://etd.lib.nsysu.edu.tw/ETD-db/ETD-search/view_etd?URN=etd-1025115-185021
Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation
Chicago Manual of Style (16th Edition):
Lin, Hung-shyuan. “Applying The Concept of Fuzzy Logic to Inverse Reinforcement Learning.” 2015. Thesis, NSYSU. Accessed March 04, 2021. http://etd.lib.nsysu.edu.tw/ETD-db/ETD-search/view_etd?URN=etd-1025115-185021.
Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation
MLA Handbook (7th Edition):
Lin, Hung-shyuan. “Applying The Concept of Fuzzy Logic to Inverse Reinforcement Learning.” 2015. Web. 04 Mar 2021.
Vancouver:
Lin H. Applying The Concept of Fuzzy Logic to Inverse Reinforcement Learning. [Internet] [Thesis]. NSYSU; 2015. [cited 2021 Mar 04]. Available from: http://etd.lib.nsysu.edu.tw/ETD-db/ETD-search/view_etd?URN=etd-1025115-185021.
Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation
Council of Science Editors:
Lin H. Applying The Concept of Fuzzy Logic to Inverse Reinforcement Learning. [Thesis]. NSYSU; 2015. Available from: http://etd.lib.nsysu.edu.tw/ETD-db/ETD-search/view_etd?URN=etd-1025115-185021
Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation
University of Waterloo
23. Bhalla, Sushrut. Deep Multi Agent Reinforcement Learning for Autonomous Driving.
Degree: 2020, University of Waterloo
URL: http://hdl.handle.net/10012/15799
Subjects/Keywords: Machine Learning; Reinforcement Learning; Multi-Agent Reinforcement Learning
Record Details
Similar Records
❌
APA · Chicago · MLA · Vancouver · CSE | Export to Zotero / EndNote / Reference Manager
APA (6th Edition):
Bhalla, S. (2020). Deep Multi Agent Reinforcement Learning for Autonomous Driving. (Thesis). University of Waterloo. Retrieved from http://hdl.handle.net/10012/15799
Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation
Chicago Manual of Style (16th Edition):
Bhalla, Sushrut. “Deep Multi Agent Reinforcement Learning for Autonomous Driving.” 2020. Thesis, University of Waterloo. Accessed March 04, 2021. http://hdl.handle.net/10012/15799.
Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation
MLA Handbook (7th Edition):
Bhalla, Sushrut. “Deep Multi Agent Reinforcement Learning for Autonomous Driving.” 2020. Web. 04 Mar 2021.
Vancouver:
Bhalla S. Deep Multi Agent Reinforcement Learning for Autonomous Driving. [Internet] [Thesis]. University of Waterloo; 2020. [cited 2021 Mar 04]. Available from: http://hdl.handle.net/10012/15799.
Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation
Council of Science Editors:
Bhalla S. Deep Multi Agent Reinforcement Learning for Autonomous Driving. [Thesis]. University of Waterloo; 2020. Available from: http://hdl.handle.net/10012/15799
Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation
Delft University of Technology
24. van der Wijden, R. (author). Preference-driven demonstrations ranking for inverse reinforcement learning.
Degree: 2016, Delft University of Technology
URL: http://resolver.tudelft.nl/uuid:4a85d32d-79da-4983-97d7-530c7bb1da98
Subjects/Keywords: robotics; reinforcement learning; preference learning; inverse reinforcement learning
Record Details
Similar Records
❌
APA · Chicago · MLA · Vancouver · CSE | Export to Zotero / EndNote / Reference Manager
APA (6th Edition):
van der Wijden, R. (. (2016). Preference-driven demonstrations ranking for inverse reinforcement learning. (Masters Thesis). Delft University of Technology. Retrieved from http://resolver.tudelft.nl/uuid:4a85d32d-79da-4983-97d7-530c7bb1da98
Chicago Manual of Style (16th Edition):
van der Wijden, R (author). “Preference-driven demonstrations ranking for inverse reinforcement learning.” 2016. Masters Thesis, Delft University of Technology. Accessed March 04, 2021. http://resolver.tudelft.nl/uuid:4a85d32d-79da-4983-97d7-530c7bb1da98.
MLA Handbook (7th Edition):
van der Wijden, R (author). “Preference-driven demonstrations ranking for inverse reinforcement learning.” 2016. Web. 04 Mar 2021.
Vancouver:
van der Wijden R(. Preference-driven demonstrations ranking for inverse reinforcement learning. [Internet] [Masters thesis]. Delft University of Technology; 2016. [cited 2021 Mar 04]. Available from: http://resolver.tudelft.nl/uuid:4a85d32d-79da-4983-97d7-530c7bb1da98.
Council of Science Editors:
van der Wijden R(. Preference-driven demonstrations ranking for inverse reinforcement learning. [Masters Thesis]. Delft University of Technology; 2016. Available from: http://resolver.tudelft.nl/uuid:4a85d32d-79da-4983-97d7-530c7bb1da98
Oregon State University
25. Zhang, Wei, 1960-. Reinforcement learning for job-shop scheduling.
Degree: PhD, Computer Science, 1996, Oregon State University
URL: http://hdl.handle.net/1957/11721
Subjects/Keywords: Reinforcement learning
Record Details
Similar Records
❌
APA · Chicago · MLA · Vancouver · CSE | Export to Zotero / EndNote / Reference Manager
APA (6th Edition):
Zhang, Wei, 1. (1996). Reinforcement learning for job-shop scheduling. (Doctoral Dissertation). Oregon State University. Retrieved from http://hdl.handle.net/1957/11721
Chicago Manual of Style (16th Edition):
Zhang, Wei, 1960-. “Reinforcement learning for job-shop scheduling.” 1996. Doctoral Dissertation, Oregon State University. Accessed March 04, 2021. http://hdl.handle.net/1957/11721.
MLA Handbook (7th Edition):
Zhang, Wei, 1960-. “Reinforcement learning for job-shop scheduling.” 1996. Web. 04 Mar 2021.
Vancouver:
Zhang, Wei 1. Reinforcement learning for job-shop scheduling. [Internet] [Doctoral dissertation]. Oregon State University; 1996. [cited 2021 Mar 04]. Available from: http://hdl.handle.net/1957/11721.
Council of Science Editors:
Zhang, Wei 1. Reinforcement learning for job-shop scheduling. [Doctoral Dissertation]. Oregon State University; 1996. Available from: http://hdl.handle.net/1957/11721
26. Clark, Kendrick Cheng Go. A Reinforcement Learning Model of the Shepherding Task : 羊飼い課題の強化学習モデル; ヒツジ カイ カダイ ノ キョウカ ガクシュウ モデル.
Degree: Nara Institute of Science and Technology / 奈良先端科学技術大学院大学
URL: http://hdl.handle.net/10061/10997
Subjects/Keywords: reinforcement learning
Record Details
Similar Records
❌
APA · Chicago · MLA · Vancouver · CSE | Export to Zotero / EndNote / Reference Manager
APA (6th Edition):
Clark, K. C. G. (n.d.). A Reinforcement Learning Model of the Shepherding Task : 羊飼い課題の強化学習モデル; ヒツジ カイ カダイ ノ キョウカ ガクシュウ モデル. (Thesis). Nara Institute of Science and Technology / 奈良先端科学技術大学院大学. Retrieved from http://hdl.handle.net/10061/10997
Note: this citation may be lacking information needed for this citation format:
No year of publication.
Not specified: Masters Thesis or Doctoral Dissertation
Chicago Manual of Style (16th Edition):
Clark, Kendrick Cheng Go. “A Reinforcement Learning Model of the Shepherding Task : 羊飼い課題の強化学習モデル; ヒツジ カイ カダイ ノ キョウカ ガクシュウ モデル.” Thesis, Nara Institute of Science and Technology / 奈良先端科学技術大学院大学. Accessed March 04, 2021. http://hdl.handle.net/10061/10997.
Note: this citation may be lacking information needed for this citation format:
No year of publication.
Not specified: Masters Thesis or Doctoral Dissertation
MLA Handbook (7th Edition):
Clark, Kendrick Cheng Go. “A Reinforcement Learning Model of the Shepherding Task : 羊飼い課題の強化学習モデル; ヒツジ カイ カダイ ノ キョウカ ガクシュウ モデル.” Web. 04 Mar 2021.
Note: this citation may be lacking information needed for this citation format:
No year of publication.
Vancouver:
Clark KCG. A Reinforcement Learning Model of the Shepherding Task : 羊飼い課題の強化学習モデル; ヒツジ カイ カダイ ノ キョウカ ガクシュウ モデル. [Internet] [Thesis]. Nara Institute of Science and Technology / 奈良先端科学技術大学院大学; [cited 2021 Mar 04]. Available from: http://hdl.handle.net/10061/10997.
Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation
No year of publication.
Council of Science Editors:
Clark KCG. A Reinforcement Learning Model of the Shepherding Task : 羊飼い課題の強化学習モデル; ヒツジ カイ カダイ ノ キョウカ ガクシュウ モデル. [Thesis]. Nara Institute of Science and Technology / 奈良先端科学技術大学院大学; Available from: http://hdl.handle.net/10061/10997
Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation
No year of publication.
27. Mauricio Alexandre Parente Burdelis. Temporal Difference Approach in Linearly Solvable Markov Decision Processes : 線形可解マルコフ決定過程における受動的ダイナミクスのモデリングと推定; センケイ カカイ マルコフ ケッテイ カテイ ニ オケル ジュドウテキ ダイナミクス ノ モデリング ト スイテイ.
Degree: 博士(工学), Nara Institute of Science and Technology / 奈良先端科学技術大学院大学
URL: http://hdl.handle.net/10061/9189
Subjects/Keywords: Reinforcement learning
Record Details
Similar Records
❌
APA · Chicago · MLA · Vancouver · CSE | Export to Zotero / EndNote / Reference Manager
APA (6th Edition):
Burdelis, M. A. P. (n.d.). Temporal Difference Approach in Linearly Solvable Markov Decision Processes : 線形可解マルコフ決定過程における受動的ダイナミクスのモデリングと推定; センケイ カカイ マルコフ ケッテイ カテイ ニ オケル ジュドウテキ ダイナミクス ノ モデリング ト スイテイ. (Thesis). Nara Institute of Science and Technology / 奈良先端科学技術大学院大学. Retrieved from http://hdl.handle.net/10061/9189
Note: this citation may be lacking information needed for this citation format:
No year of publication.
Not specified: Masters Thesis or Doctoral Dissertation
Chicago Manual of Style (16th Edition):
Burdelis, Mauricio Alexandre Parente. “Temporal Difference Approach in Linearly Solvable Markov Decision Processes : 線形可解マルコフ決定過程における受動的ダイナミクスのモデリングと推定; センケイ カカイ マルコフ ケッテイ カテイ ニ オケル ジュドウテキ ダイナミクス ノ モデリング ト スイテイ.” Thesis, Nara Institute of Science and Technology / 奈良先端科学技術大学院大学. Accessed March 04, 2021. http://hdl.handle.net/10061/9189.
Note: this citation may be lacking information needed for this citation format:
No year of publication.
Not specified: Masters Thesis or Doctoral Dissertation
MLA Handbook (7th Edition):
Burdelis, Mauricio Alexandre Parente. “Temporal Difference Approach in Linearly Solvable Markov Decision Processes : 線形可解マルコフ決定過程における受動的ダイナミクスのモデリングと推定; センケイ カカイ マルコフ ケッテイ カテイ ニ オケル ジュドウテキ ダイナミクス ノ モデリング ト スイテイ.” Web. 04 Mar 2021.
Note: this citation may be lacking information needed for this citation format:
No year of publication.
Vancouver:
Burdelis MAP. Temporal Difference Approach in Linearly Solvable Markov Decision Processes : 線形可解マルコフ決定過程における受動的ダイナミクスのモデリングと推定; センケイ カカイ マルコフ ケッテイ カテイ ニ オケル ジュドウテキ ダイナミクス ノ モデリング ト スイテイ. [Internet] [Thesis]. Nara Institute of Science and Technology / 奈良先端科学技術大学院大学; [cited 2021 Mar 04]. Available from: http://hdl.handle.net/10061/9189.
Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation
No year of publication.
Council of Science Editors:
Burdelis MAP. Temporal Difference Approach in Linearly Solvable Markov Decision Processes : 線形可解マルコフ決定過程における受動的ダイナミクスのモデリングと推定; センケイ カカイ マルコフ ケッテイ カテイ ニ オケル ジュドウテキ ダイナミクス ノ モデリング ト スイテイ. [Thesis]. Nara Institute of Science and Technology / 奈良先端科学技術大学院大学; Available from: http://hdl.handle.net/10061/9189
Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation
No year of publication.
28. 森本, 淳. Hierarchical Decomposition and Min-max Strategy for Fast and Robust Reinforcement Learning in the Real Environment : 階層分割とMin-max戦略による実環境での高速かつロバストな強化学習; カイソウ ブンカツ ト Min-max センリャク ニヨル ジツカンキョウ デノ コウソク カツ ロバストナ キョウカ ガクシュウ.
Degree: Nara Institute of Science and Technology / 奈良先端科学技術大学院大学
URL: http://hdl.handle.net/10061/2966
Subjects/Keywords: reinforcement learning
Record Details
Similar Records
❌
APA · Chicago · MLA · Vancouver · CSE | Export to Zotero / EndNote / Reference Manager
APA (6th Edition):
森本, . (n.d.). Hierarchical Decomposition and Min-max Strategy for Fast and Robust Reinforcement Learning in the Real Environment : 階層分割とMin-max戦略による実環境での高速かつロバストな強化学習; カイソウ ブンカツ ト Min-max センリャク ニヨル ジツカンキョウ デノ コウソク カツ ロバストナ キョウカ ガクシュウ. (Thesis). Nara Institute of Science and Technology / 奈良先端科学技術大学院大学. Retrieved from http://hdl.handle.net/10061/2966
Note: this citation may be lacking information needed for this citation format:
No year of publication.
Not specified: Masters Thesis or Doctoral Dissertation
Chicago Manual of Style (16th Edition):
森本, 淳. “Hierarchical Decomposition and Min-max Strategy for Fast and Robust Reinforcement Learning in the Real Environment : 階層分割とMin-max戦略による実環境での高速かつロバストな強化学習; カイソウ ブンカツ ト Min-max センリャク ニヨル ジツカンキョウ デノ コウソク カツ ロバストナ キョウカ ガクシュウ.” Thesis, Nara Institute of Science and Technology / 奈良先端科学技術大学院大学. Accessed March 04, 2021. http://hdl.handle.net/10061/2966.
Note: this citation may be lacking information needed for this citation format:
No year of publication.
Not specified: Masters Thesis or Doctoral Dissertation
MLA Handbook (7th Edition):
森本, 淳. “Hierarchical Decomposition and Min-max Strategy for Fast and Robust Reinforcement Learning in the Real Environment : 階層分割とMin-max戦略による実環境での高速かつロバストな強化学習; カイソウ ブンカツ ト Min-max センリャク ニヨル ジツカンキョウ デノ コウソク カツ ロバストナ キョウカ ガクシュウ.” Web. 04 Mar 2021.
Note: this citation may be lacking information needed for this citation format:
No year of publication.
Vancouver:
森本 . Hierarchical Decomposition and Min-max Strategy for Fast and Robust Reinforcement Learning in the Real Environment : 階層分割とMin-max戦略による実環境での高速かつロバストな強化学習; カイソウ ブンカツ ト Min-max センリャク ニヨル ジツカンキョウ デノ コウソク カツ ロバストナ キョウカ ガクシュウ. [Internet] [Thesis]. Nara Institute of Science and Technology / 奈良先端科学技術大学院大学; [cited 2021 Mar 04]. Available from: http://hdl.handle.net/10061/2966.
Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation
No year of publication.
Council of Science Editors:
森本 . Hierarchical Decomposition and Min-max Strategy for Fast and Robust Reinforcement Learning in the Real Environment : 階層分割とMin-max戦略による実環境での高速かつロバストな強化学習; カイソウ ブンカツ ト Min-max センリャク ニヨル ジツカンキョウ デノ コウソク カツ ロバストナ キョウカ ガクシュウ. [Thesis]. Nara Institute of Science and Technology / 奈良先端科学技術大学院大学; Available from: http://hdl.handle.net/10061/2966
Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation
No year of publication.
29. 南條, 信人. 動的な部分空間生成による価値の逐次推定を行う強化学習法 : An effective reinforcement learning with automatic construction of basis functions and sequential approximation; ドウテキナ ブブン クウカン セイセイ ニヨル カチ ノ チクジ スイテイ オ オコナウ キョウカ ガクシュウホウ.
Degree: Nara Institute of Science and Technology / 奈良先端科学技術大学院大学
URL: http://hdl.handle.net/10061/4584
Subjects/Keywords: Reinforcement learning
Record Details
Similar Records
❌
APA · Chicago · MLA · Vancouver · CSE | Export to Zotero / EndNote / Reference Manager
APA (6th Edition):
南條, . (n.d.). 動的な部分空間生成による価値の逐次推定を行う強化学習法 : An effective reinforcement learning with automatic construction of basis functions and sequential approximation; ドウテキナ ブブン クウカン セイセイ ニヨル カチ ノ チクジ スイテイ オ オコナウ キョウカ ガクシュウホウ. (Thesis). Nara Institute of Science and Technology / 奈良先端科学技術大学院大学. Retrieved from http://hdl.handle.net/10061/4584
Note: this citation may be lacking information needed for this citation format:
No year of publication.
Not specified: Masters Thesis or Doctoral Dissertation
Chicago Manual of Style (16th Edition):
南條, 信人. “動的な部分空間生成による価値の逐次推定を行う強化学習法 : An effective reinforcement learning with automatic construction of basis functions and sequential approximation; ドウテキナ ブブン クウカン セイセイ ニヨル カチ ノ チクジ スイテイ オ オコナウ キョウカ ガクシュウホウ.” Thesis, Nara Institute of Science and Technology / 奈良先端科学技術大学院大学. Accessed March 04, 2021. http://hdl.handle.net/10061/4584.
Note: this citation may be lacking information needed for this citation format:
No year of publication.
Not specified: Masters Thesis or Doctoral Dissertation
MLA Handbook (7th Edition):
南條, 信人. “動的な部分空間生成による価値の逐次推定を行う強化学習法 : An effective reinforcement learning with automatic construction of basis functions and sequential approximation; ドウテキナ ブブン クウカン セイセイ ニヨル カチ ノ チクジ スイテイ オ オコナウ キョウカ ガクシュウホウ.” Web. 04 Mar 2021.
Note: this citation may be lacking information needed for this citation format:
No year of publication.
Vancouver:
南條 . 動的な部分空間生成による価値の逐次推定を行う強化学習法 : An effective reinforcement learning with automatic construction of basis functions and sequential approximation; ドウテキナ ブブン クウカン セイセイ ニヨル カチ ノ チクジ スイテイ オ オコナウ キョウカ ガクシュウホウ. [Internet] [Thesis]. Nara Institute of Science and Technology / 奈良先端科学技術大学院大学; [cited 2021 Mar 04]. Available from: http://hdl.handle.net/10061/4584.
Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation
No year of publication.
Council of Science Editors:
南條 . 動的な部分空間生成による価値の逐次推定を行う強化学習法 : An effective reinforcement learning with automatic construction of basis functions and sequential approximation; ドウテキナ ブブン クウカン セイセイ ニヨル カチ ノ チクジ スイテイ オ オコナウ キョウカ ガクシュウホウ. [Thesis]. Nara Institute of Science and Technology / 奈良先端科学技術大学院大学; Available from: http://hdl.handle.net/10061/4584
Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation
No year of publication.
30. Rodrigues, Alan de Souza. Model-Free and Model-Based Reinforcement Learning Strategies in the Acquisition of Sequential Behaviors : 系列運動の獲得におけるモデルフリーとモデルベース強化学習戦略; ケイレツ ウンドウ ノ カクトク ニオケル モデル フリー ト モデル ベース キョウカ ガクシュウ センリャク.
Degree: Nara Institute of Science and Technology / 奈良先端科学技術大学院大学
URL: http://hdl.handle.net/10061/4681
Subjects/Keywords: Reinforcement Learning
Record Details
Similar Records
❌
APA · Chicago · MLA · Vancouver · CSE | Export to Zotero / EndNote / Reference Manager
APA (6th Edition):
Rodrigues, A. d. S. (n.d.). Model-Free and Model-Based Reinforcement Learning Strategies in the Acquisition of Sequential Behaviors : 系列運動の獲得におけるモデルフリーとモデルベース強化学習戦略; ケイレツ ウンドウ ノ カクトク ニオケル モデル フリー ト モデル ベース キョウカ ガクシュウ センリャク. (Thesis). Nara Institute of Science and Technology / 奈良先端科学技術大学院大学. Retrieved from http://hdl.handle.net/10061/4681
Note: this citation may be lacking information needed for this citation format:
No year of publication.
Not specified: Masters Thesis or Doctoral Dissertation
Chicago Manual of Style (16th Edition):
Rodrigues, Alan de Souza. “Model-Free and Model-Based Reinforcement Learning Strategies in the Acquisition of Sequential Behaviors : 系列運動の獲得におけるモデルフリーとモデルベース強化学習戦略; ケイレツ ウンドウ ノ カクトク ニオケル モデル フリー ト モデル ベース キョウカ ガクシュウ センリャク.” Thesis, Nara Institute of Science and Technology / 奈良先端科学技術大学院大学. Accessed March 04, 2021. http://hdl.handle.net/10061/4681.
Note: this citation may be lacking information needed for this citation format:
No year of publication.
Not specified: Masters Thesis or Doctoral Dissertation
MLA Handbook (7th Edition):
Rodrigues, Alan de Souza. “Model-Free and Model-Based Reinforcement Learning Strategies in the Acquisition of Sequential Behaviors : 系列運動の獲得におけるモデルフリーとモデルベース強化学習戦略; ケイレツ ウンドウ ノ カクトク ニオケル モデル フリー ト モデル ベース キョウカ ガクシュウ センリャク.” Web. 04 Mar 2021.
Note: this citation may be lacking information needed for this citation format:
No year of publication.
Vancouver:
Rodrigues AdS. Model-Free and Model-Based Reinforcement Learning Strategies in the Acquisition of Sequential Behaviors : 系列運動の獲得におけるモデルフリーとモデルベース強化学習戦略; ケイレツ ウンドウ ノ カクトク ニオケル モデル フリー ト モデル ベース キョウカ ガクシュウ センリャク. [Internet] [Thesis]. Nara Institute of Science and Technology / 奈良先端科学技術大学院大学; [cited 2021 Mar 04]. Available from: http://hdl.handle.net/10061/4681.
Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation
No year of publication.
Council of Science Editors:
Rodrigues AdS. Model-Free and Model-Based Reinforcement Learning Strategies in the Acquisition of Sequential Behaviors : 系列運動の獲得におけるモデルフリーとモデルベース強化学習戦略; ケイレツ ウンドウ ノ カクトク ニオケル モデル フリー ト モデル ベース キョウカ ガクシュウ センリャク. [Thesis]. Nara Institute of Science and Technology / 奈良先端科学技術大学院大学; Available from: http://hdl.handle.net/10061/4681
Note: this citation may be lacking information needed for this citation format:
Not specified: Masters Thesis or Doctoral Dissertation
No year of publication.