@@ -100,72 +100,72 @@ nav_order: 4
100100 </thead >
101101 <tbody >
102102
103- <tr>
104- <td>August 29th</td>
105- <td>
106- <strong>Governance: LLM + AI Agents</strong><br>
107- Sabina (Yudi) Nong, Stanford<br><br>
108-
109- <strong>Key Points:</strong><br>
110- <ul>
111- <li>AI Governance & its Stakes</li>
112- <li>Why AI Agents Require a Distinct Governance Lens (Procedural vs. Predictive)</li>
113- <li>Developer Governance</li>
114- <li>Regulatory Governance</li>
115- <li>What does a good governance structure look like?</li>
116- </ul>
117-
118- <a href="https://docs.google.com/presentation/d/145F-UnXUBswnz03WvZkliK7eLiLyhGY0bTrDJo8VUAk/mobilepresent?slide=id.g376427fc988_0_361" target="_blank">Slides (Website)</a> ·
119- <a href="https://www.youtube.com/watch?v=MpbzMipStXk&pp=0gcJCU0KAYcqIYzv" target="_blank">Recording</a>
120- </td>
121- <td>
122- <em>To be updated</em>
123- </td>
124- </tr>
125-
126- <tr>
127- <td>September 3rd</td>
128- <td>
129- <strong>Interactive symbolic regression with co-design mechanism</strong><br>
130- Yuan Tian, ETH Zurich<br><br>
131-
132- <strong>Key Topics:</strong><br>
133- <ul>
134- <li>Introduction to symbolic regression and its applications in scientific discovery</li>
135- <li>Challenges in traditional symbolic regression approaches</li>
136- <li>Co-design mechanism: integrating human expertise with machine learning</li>
137- <li>Interactive frameworks for collaborative model discovery</li>
138- <li>Case studies and practical applications in scientific computing</li>
139- </ul>
140-
141- <a href="https://www.youtube.com/watch?v=Q-l6pVR3z-Y" target="_blank">Recording</a>
142- </td>
143- <td>
144- <em>To be updated</em>
145- </td>
146- </tr>
147-
148- <tr>
149- <td>October 3rd</td>
150- <td>
151- <strong>Rewiring the reward pathways</strong><br>
152- Kumar Shridhar, ETH Zürich<br><br>
153-
154- <strong>Key Topics:</strong><br>
155- <ul>
156- <li>Reward models in ML pipeline</li>
157- <li>Brittleness of current reward models</li>
158- <li>What to do when there is no clear reward (non verifiable rewards)</li>
159- <li>How to rewire reward pathways</li>
160- <li>A common reward model for both verifiable and non verifiable rewards</li>
161- </ul>
162-
163- <a href="https://www.youtube.com/watch?v=x7wvhD28EMU" target="_blank">Recording</a>
164- </td>
165- <td>
166- <em>To be updated</em>
167- </td>
168- </tr>
103+ <tr >
104+ <td >August 29th</td >
105+ <td >
106+ <strong>Governance: LLM + AI Agents</strong><br>
107+ Sabina (Yudi) Nong, Stanford<br><br>
108+
109+ <strong>Key Points:</strong><br>
110+ <ul>
111+ <li>AI Governance & its Stakes</li>
112+ <li>Why AI Agents Require a Distinct Governance Lens (Procedural vs. Predictive)</li>
113+ <li>Developer Governance</li>
114+ <li>Regulatory Governance</li>
115+ <li>What does a good governance structure look like?</li>
116+ </ul>
117+
118+ <a href="https://docs.google.com/presentation/d/145F-UnXUBswnz03WvZkliK7eLiLyhGY0bTrDJo8VUAk/mobilepresent?slide=id.g376427fc988_0_361" target="_blank">Slides (Website)</a> ·
119+ <a href="https://www.youtube.com/watch?v=MpbzMipStXk&pp=0gcJCU0KAYcqIYzv" target="_blank">Recording</a>
120+ </td >
121+ <td >
122+ <em>To be updated</em>
123+ </td >
124+ </tr >
125+
126+ <tr >
127+ <td >September 3rd</td >
128+ <td >
129+ <strong>Interactive symbolic regression with co-design mechanism</strong><br>
130+ Yuan Tian, ETH Zurich<br><br>
131+
132+ <strong>Key Topics:</strong><br>
133+ <ul>
134+ <li>Introduction to symbolic regression and its applications in scientific discovery</li>
135+ <li>Challenges in traditional symbolic regression approaches</li>
136+ <li>Co-design mechanism: integrating human expertise with machine learning</li>
137+ <li>Interactive frameworks for collaborative model discovery</li>
138+ <li>Case studies and practical applications in scientific computing</li>
139+ </ul>
140+
141+ <a href="https://www.youtube.com/watch?v=Q-l6pVR3z-Y" target="_blank">Recording</a>
142+ </td >
143+ <td >
144+ <em>To be updated</em>
145+ </td >
146+ </tr >
147+
148+ <tr >
149+ <td >October 3rd</td >
150+ <td >
151+ <strong>Rewiring the reward pathways</strong><br>
152+ Kumar Shridhar, ETH Zürich<br><br>
153+
154+ <strong>Key Topics:</strong><br>
155+ <ul>
156+ <li>Reward models in ML pipeline</li>
157+ <li>Brittleness of current reward models</li>
158+ <li>What to do when there is no clear reward (non verifiable rewards)</li>
159+ <li>How to rewire reward pathways</li>
160+ <li>A common reward model for both verifiable and non verifiable rewards</li>
161+ </ul>
162+
163+ <a href="https://www.youtube.com/watch?v=x7wvhD28EMU" target="_blank">Recording</a>
164+ </td >
165+ <td >
166+ <em>To be updated</em>
167+ </td >
168+ </tr >
169169
170170 </tbody >
171171</table >
0 commit comments