Skip to content

Commit

Permalink
Add BigAI speakers to the schedule
Browse files Browse the repository at this point in the history
  • Loading branch information
YzyLmc committed Feb 27, 2024
1 parent e834af3 commit 5ccb404
Show file tree
Hide file tree
Showing 6 changed files with 59 additions and 0 deletions.
24 changes: 24 additions & 0 deletions _speakers/5_aldopacchiano.md
Original file line number Diff line number Diff line change
@@ -0,0 +1,24 @@
---
# Name of the speaker
name: Aldo Pacchiano

# Link to the speaker's webpage
webpage: https://www.aldopacchiano.ai/

# Primary affiliation of the speaker
affil: Broad Institute
# Link to the speaker's primary affiliation
affil_link: https://www.broadinstitute.org/

# An image of the speaker (square aspect ratio works the best) (place in the `assets/img/speakers` directory)
img: aldopacchiano.jpeg

# # (Optional) Secondary affiliation of the speaker
# affil2: BuzzFizz Corp
# # Link to the speaker's secondary affiliation
# affil2_link: https://buzzfizz.corp
---

<!-- Whatever you write below will show up as the speaker's bio -->

In settings where there is a significant overhead to deploying adaptive algorithms -- for example, when the execution of the data collection policies is required to be distributed, or a human in the loop is needed to implement these policies -- producing in advance a set of policies for data collection is paramount. We study the setting where a large dataset of contexts but not rewards is available and may be used by the learner to design an effective data collection strategy. Although when rewards are linear this problem has been well studied, results are still missing for more complex reward models. In this work we propose two experiment planning strategies compatible with function approximation. The first is an eluder planning and sampling procedure that can recover optimality guarantees depending on the eluder dimension of the reward function class. For the second, we show that a uniform sampler achieves competitive optimality rates in the setting where the number of actions is small. We finalize our results introducing a statistical gap fleshing out the fundamental differences between planning and adaptive learning and provide results for planning with model selection.
24 changes: 24 additions & 0 deletions _speakers/6_katedarling.md
Original file line number Diff line number Diff line change
@@ -0,0 +1,24 @@
---
# Name of the speaker
name: Kate Darling

# Link to the speaker's webpage
webpage: https://www.katedarling.org/

# Primary affiliation of the speaker
affil: MIT Media Lab
# Link to the speaker's primary affiliation
affil_link: https://www.media.mit.edu/

# An image of the speaker (square aspect ratio works the best) (place in the `assets/img/speakers` directory)
img: katedarling.jpeg

# # (Optional) Secondary affiliation of the speaker
# affil2: BuzzFizz Corp
# # Link to the speaker's secondary affiliation
# affil2_link: https://buzzfizz.corp
---

<!-- Whatever you write below will show up as the speaker's bio -->

__TBD__
1 change: 1 addition & 0 deletions assets/abstracts/aldopacchiano.txt
Original file line number Diff line number Diff line change
@@ -0,0 +1 @@
In settings where there is a significant overhead to deploying adaptive algorithms -- for example, when the execution of the data collection policies is required to be distributed, or a human in the loop is needed to implement these policies -- producing in advance a set of policies for data collection is paramount. We study the setting where a large dataset of contexts but not rewards is available and may be used by the learner to design an effective data collection strategy. Although when rewards are linear this problem has been well studied, results are still missing for more complex reward models. In this work we propose two experiment planning strategies compatible with function approximation. The first is an eluder planning and sampling procedure that can recover optimality guarantees depending on the eluder dimension of the reward function class. For the second, we show that a uniform sampler achieves competitive optimality rates in the setting where the number of actions is small. We finalize our results introducing a statistical gap fleshing out the fundamental differences between planning and adaptive learning and provide results for planning with model selection.
Binary file added assets/img/speakers/aldopacchiano.jpeg
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
Binary file added assets/img/speakers/katedarling.jpeg
Loading
Sorry, something went wrong. Reload?
Sorry, we cannot display this file.
Sorry, this file is invalid so it cannot be displayed.
10 changes: 10 additions & 0 deletions index.md
Original file line number Diff line number Diff line change
Expand Up @@ -46,5 +46,15 @@ Brown Robotics Talks consists of BigAI talks and lab talks ([CIT](https://www.go
<td><b>TBD</b></td>
<td>Jiayuan Mao</td>
</tr>
<tr>
<td>04/26</td>
<td><b>Experiment Planning with Function Approximation</b> [<a href='assets/abstracts/aldopacchiano.txt' target="_blank">abstract</a>]</td>
<td>Aldo Pacchiano</td>
</tr>
<tr>
<td>05/10</td>
<td><b>TBD</b></td>
<td>Kate Darling</td>
</tr>
</tbody>
</table>

0 comments on commit 5ccb404

Please sign in to comment.