-
Notifications
You must be signed in to change notification settings - Fork 0
/
index.xml
122 lines (101 loc) · 8.04 KB
/
index.xml
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
<?xml version="1.0" encoding="utf-8" standalone="yes"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom">
<channel>
<title>Scott Jeen</title>
<link>https://enjeeneer.io/</link>
<description>Recent content on Scott Jeen</description>
<generator>Hugo -- gohugo.io</generator>
<language>en-us</language>
<copyright><a href="https://creativecommons.org/licenses/by-nc/4.0/" target="_blank" rel="noopener">CC BY-NC 4.0</a></copyright>
<lastBuildDate>Tue, 12 Nov 2024 14:18:07 +0000</lastBuildDate><atom:link href="https://enjeeneer.io/index.xml" rel="self" type="application/rss+xml" />
<item>
<title>4 Years in 4(ish) Minutes</title>
<link>https://enjeeneer.io/talks/2024-11-12-phd-reflection/</link>
<pubDate>Tue, 12 Nov 2024 14:18:07 +0000</pubDate>
<guid>https://enjeeneer.io/talks/2024-11-12-phd-reflection/</guid>
<description>[Click here for full-screen slideshow]</description>
</item>
<item>
<title>Searching for useful problems</title>
<link>https://enjeeneer.io/posts/2024/08/searching-for-useful-problems/</link>
<pubDate>Mon, 26 Aug 2024 19:36:27 +0200</pubDate>
<guid>https://enjeeneer.io/posts/2024/08/searching-for-useful-problems/</guid>
<description>This post is based on a talk I gave to my research group at Cambridge University in June 2024. You can find the notes and slides for the talk here.
Solutions to most problems aren’t particularly useful. Solutions to a small number of problems are extremely useful. If you’re interested in doing good, you’ll want to search for problems that look like the latter. It’s a hard search, but the ROI is likely greater than any other use of your time, and I have some ways of running it that I think make it a little more tractable.</description>
</item>
<item>
<title>Dynamics Generalisation with Behaviour Foundation Models</title>
<link>https://enjeeneer.io/talks/2024-08-09-rlc-tafm/</link>
<pubDate>Wed, 07 Aug 2024 17:01:15 -0400</pubDate>
<guid>https://enjeeneer.io/talks/2024-08-09-rlc-tafm/</guid>
<description>[Click here for full-screen slideshow]</description>
</item>
<item>
<title>The problem problem: choosing impactful research problems</title>
<link>https://enjeeneer.io/talks/2024-06-14-reffciency/</link>
<pubDate>Fri, 14 Jun 2024 09:34:36 +0100</pubDate>
<guid>https://enjeeneer.io/talks/2024-06-14-reffciency/</guid>
<description>[Click here for full-screen slideshow]
What are problems? “A problem is a situation in which we experience conflicting ideas.”
David Deutsch
Let’s start with a definition, here’s the best I’ve found for what constitutes a problem. David Deutsch says a problem is a situation in which we experience conflicting ideas. In other words we have two or more explanations about how we might proceed, and it&rsquo;s not initially clear how to choose between them.</description>
</item>
<item>
<title>Zero-Shot Reinforcement Learning from Low Quality Data</title>
<link>https://enjeeneer.io/projects/zero-shot-rl/</link>
<pubDate>Tue, 26 Sep 2023 17:27:21 +0100</pubDate>
<guid>https://enjeeneer.io/projects/zero-shot-rl/</guid>
<description>NeurIPS 2024 Scott Jeen\(^{1}\), Tom Bewley\(^{2}\), &amp; Jonathan M. Cullen\(^{1}\) \(^{1}\) University of Cambridge
\(^{2}\) University of Bristol
[Paper] [Code] [Poster] [Slides]
Summary Zero-shot reinforcement learning (RL) methods learn general policies that can, in principle, solve any unseen task in an environment. Recently, methods leveraging successor features and successor measures have emerged as viable zero-shot RL candidates, returning near-optimal policies for many unseen tasks. However, to enable this, they have assumed access to unrealistically large and heterogeneous datasets of transitions for pre-training.</description>
</item>
<item>
<title>NeurIPS 2022</title>
<link>https://enjeeneer.io/posts/2023/01/neurips-2022/</link>
<pubDate>Thu, 26 Jan 2023 21:08:05 +0000</pubDate>
<guid>https://enjeeneer.io/posts/2023/01/neurips-2022/</guid>
<description>I was fortunate to attend NeurIPS in New Orleans in November. Here, I publish my takeaways to give you a feel for the zeitgeist. I’ll discuss, firstly, the papers, then the workshops, and finally, and briefly, the keynotes.
Papers Here’s a ranked list of my top 8 papers. Most are on Offline RL, which is representative of the conference writ large.
1. Does Zero-Shot Reinforcement Learning Exist (Touati et. al, 2022)</description>
</item>
<item>
<title>Low Emission Building Control with Zero-Shot Reinforcement Learning</title>
<link>https://enjeeneer.io/projects/pearl/</link>
<pubDate>Fri, 12 Aug 2022 09:34:36 +0100</pubDate>
<guid>https://enjeeneer.io/projects/pearl/</guid>
<description>In Proceedings of the AAAI Conference on Artificial Intelligence (Vol. 37, No. 12, pp. 14259-14267)
Scott R. Jeen\(^{1,3}\), Alessandro Abate\(^{2,3}\), Jonathan M. Cullen\(^{1}\) \(^{1}\) University of Cambridge
\(^{2}\) University of Oxford
\(^{3}\) Alan Turing Institute
[Paper] [Code] [Talk] [Poster]
PEARL: Probabilistic Emission-Abating Reinforcement Learning Presenting PEARL: Probabilistic Emission-Abating Reinforcement Learning, a deep RL algorithm that can find performant building control policies online, without pre-training&ndash;the first time this has been shown to be possible.</description>
</item>
<item>
<title>Zero Shot Building Control</title>
<link>https://enjeeneer.io/talks/2022-06-23oxcav/</link>
<pubDate>Thu, 23 Jun 2022 09:34:36 +0100</pubDate>
<guid>https://enjeeneer.io/talks/2022-06-23oxcav/</guid>
<description>Abstract
Heating and cooling systems in buildings account for 31% of global energy use, much of which are regulated by Rule Based Controllers (RBCs) that neither maximise energy efficiency nor minimise emissions by interacting optimally with the grid. Control via Reinforcement Learning (RL) has been shown to significantly improve building energy efficiency, but existing solutions require pre-training in simulators that are prohibitively expensive to obtain for every building in the world.</description>
</item>
<item>
<title>One Hour RL</title>
<link>https://enjeeneer.io/posts/2022/02/one-hour-rl/</link>
<pubDate>Fri, 25 Feb 2022 15:23:20 +0000</pubDate>
<guid>https://enjeeneer.io/posts/2022/02/one-hour-rl/</guid>
<description>An Introduction to Reinforcement Learning Tom Bewley &amp; Scott Jeen Alan Turing Institute 24/02/2022 The best way to walk through this tutorial is using the accompanying Jupyter Notebook:
[Jupyter Notebook]
1 | Markov Decision Processes: A Model of Sequential Decision Making 1.1. MDP (semi-)Formalism In reinforcement learning (RL), an agent takes actions in an environment to change its state over discrete timesteps $t$, with the goal of maximising the future sum of a scalar quantity known as reward.</description>
</item>
<item>
<title>Presenting with Jupyter Notebooks</title>
<link>https://enjeeneer.io/posts/2021/11/presenting-with-jupyter-notebooks/</link>
<pubDate>Wed, 17 Nov 2021 09:03:11 -0500</pubDate>
<guid>https://enjeeneer.io/posts/2021/11/presenting-with-jupyter-notebooks/</guid>
<description>The best way to walk through this tutorial is using the accompanying Jupyter Notebook:
[Jupyter Notebook]
- In the last year I&rsquo;ve started presenting work using Jupyter Notebooks, rebelling against the Bill Gates'-driven status-quo. Here I&rsquo;ll explain how to do it. It&rsquo;s not difficult, but in my opinion makes presentations look slicker, whilst allowing you to run code live in a presentation if you like. First, we need to download the plug-in that gives us the presentation functionality, it&rsquo;s called RISE.</description>
</item>
</channel>
</rss>