<?xml version="1.0" encoding="utf-8" standalone="yes" ?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom">
  <channel>
    <title>Aldo Pacchiano</title>
    <link>/authors/admin/</link>
      <atom:link href="/authors/admin/index.xml" rel="self" type="application/rss+xml" />
    <description>Aldo Pacchiano</description>
    <generator>Source Themes Academic (https://sourcethemes.com/academic/)</generator><language>en-us</language><lastBuildDate>Wed, 21 Jan 2026 00:00:00 +0000</lastBuildDate>
    <image>
      <url>/img/icon.png</url>
      <title>Aldo Pacchiano</title>
      <link>/authors/admin/</link>
    </image>
    
    <item>
      <title>Autonomous Discovery from Data</title>
      <link>/post/autonomousdiscovery/</link>
      <pubDate>Wed, 21 Jan 2026 00:00:00 +0000</pubDate>
      <guid>/post/autonomousdiscovery/</guid>
      <description>&lt;h2 id=&#34;overview&#34;&gt;Overview&lt;/h2&gt;
&lt;p&gt;Sequential decision-making algorithms are used to interact with environments where the objective is to learn a policy that improves with data. Among others these are scenarios such as bandit problems, where one wants to compete against the best arm in hindsight, reinforcement learning domains, where finding an optimal policy is paramount, or active learning settings where the goal is to produce an accurate model of the world in as few online interactions as possible. Applications where these models are used include scenarios such as robotics and experiment design in scientific domains. Designing sequential-decision making algorithms typically involves a careful and painstaking modeling process that requires deep domain knowledge as well as mathematical proficiency in the techniques of sequential decision-making. In these works we explore an alternative way of producing algorithms for sequential decision-making domains. Instead, we will use the in-context learning abilities of transformer models to encode history dependent policies (i.e. algorithms) that will be learned from data by minimizing an appropriate loss. In our work &lt;a href=&#34;https://arxiv.org/abs/2306.14892&#34;&gt;&lt;strong&gt;Supervised Pretraining Can Learn In-Context Reinforcement Learning&lt;/strong&gt;&lt;/a&gt; we design the &lt;strong&gt;Decision Pretrained Transformer (DPT)&lt;/strong&gt; a supervised learning strategy over offline data that can be used to recover the Thompson Sampling algorithm. We expand on this line of work in our paper &lt;a href=&#34;https://arxiv.org/abs/2506.01876&#34;&gt;&lt;strong&gt;Learning to Explore: An In-Context Learning Approach for Pure Exploration&lt;/strong&gt;&lt;/a&gt; where we introduce the &lt;strong&gt;In Context Pure Exploration (ICPE)&lt;/strong&gt; algorithm. In this work learning is done via an online reinforcement learning approach. ICPE is able to recover a bayes optimal discovery strategy for sequential hypothesis testing adapted to specific problem families. Learning how to autonomously discover is an exciting area of research that our research group is proudly pioneering.&lt;/p&gt;
&lt;h2 id=&#34;papers&#34;&gt;Papers:&lt;/h2&gt;
&lt;ul&gt;
&lt;li&gt;&lt;a href=&#34;https://arxiv.org/abs/2306.14892&#34;&gt;&lt;strong&gt;Supervised Pretraining Can Learn In-Context Reinforcement Learning&lt;/strong&gt;&lt;/a&gt;&lt;/li&gt;
&lt;li&gt;&lt;a href=&#34;https://arxiv.org/abs/2506.01876&#34;&gt;&lt;strong&gt;Learning to Explore: An In-Context Learning Approach for Pure Exploration&lt;/strong&gt;&lt;/a&gt;&lt;/li&gt;
&lt;/ul&gt;
</description>
    </item>
    
    <item>
      <title>The Dissimilarity Dimension</title>
      <link>/post/dissimilarity/</link>
      <pubDate>Fri, 27 Sep 2024 00:00:00 +0000</pubDate>
      <guid>/post/dissimilarity/</guid>
      <description>&lt;h2 id=&#34;overview&#34;&gt;Overview&lt;/h2&gt;
&lt;p&gt;For a time the Eluder dimension has been used to provide bounds for optimistic algorithms in function approximation regimes. We introduce the dissimilarity dimension that provides us with tighter bounds. The dissimilarity dimension is a combinatorial statistical dimension that can be used to define the statistical complexity of optimistic algorithms in structured bandit problems. This statistical dimension&amp;rsquo;s definition is based on defining maximal sequences that achieve good historical fit while also satisfying what we call a large self evalution property that is connected with optimism. Using a graph theoretic argument, we prove the regret of optimistic least squares achieves a regret bound depending on this statistical dimension that is sharper than the typical eluder dimension bound.&lt;/p&gt;
&lt;h2 id=&#34;paper&#34;&gt;Paper:&lt;/h2&gt;
&lt;ul&gt;
&lt;li&gt;&lt;a href=&#34;https://arxiv.org/abs/2306.06184&#34;&gt;&lt;strong&gt;A Unified Model and Dimension for Interactive Estimation&lt;/strong&gt;&lt;/a&gt;&lt;/li&gt;
&lt;/ul&gt;
</description>
    </item>
    
    <item>
      <title>Neural Optimism for Genetic Perturbation Experiments</title>
      <link>/post/neuralperturbation/</link>
      <pubDate>Wed, 10 Aug 2022 00:00:00 +0000</pubDate>
      <guid>/post/neuralperturbation/</guid>
      <description>&lt;h2 id=&#34;overview&#34;&gt;Overview&lt;/h2&gt;
&lt;p&gt;The problem of how to genetically modify cells in order to maximize a certain cellular phenotype has taken center stage in drug development over the last few years (with, for example, genetically edited CAR-T, CAR-NK, and CAR-NKT cells entering cancer clinical trials). Exhausting the search space for all possible genetic edits (perturbations) or combinations thereof is infeasible due to cost and experimental limitations. This work provides a theoretically sound framework for iteratively exploring the space of perturbations in pooled batches in order to maximize a target phenotype under an experimental budget. Inspired by this application domain, we study the problem of batch query bandit optimization and introduce the Optimistic Arm Elimination (OAE) principle designed to find an almost optimal arm under different functional relationships between the queries (arms) and the outputs (rewards). In this work we analyze several tractable mechanisms for optimistic gene perturbation discovery that use neural network function approximation. We analyze the convergence properties of OAE by relating it to the Eluder dimension of the algorithm&amp;rsquo;s function class and validate that OAE outperforms other strategies in finding optimal actions in experiments on simulated problems, public datasets well-studied in bandit contexts, and in genetic perturbation datasets when the regression model is a deep neural network. OAE also outperforms the benchmark algorithms in 3 of 4 datasets in the GeneDisco experimental planning challenge. Our algorithms can be used in problems beyond the setting of genetic perturbation experiments.&lt;/p&gt;
&lt;h2 id=&#34;paper&#34;&gt;Paper&lt;/h2&gt;
&lt;ul&gt;
&lt;li&gt;&lt;a href=&#34;https://arxiv.org/abs/2207.12805&#34;&gt;&lt;strong&gt;Neural Optimism for Genetic Perturbation Experiments&lt;/strong&gt;&lt;/a&gt;.&lt;/li&gt;
&lt;/ul&gt;
</description>
    </item>
    
    <item>
      <title>Model Selection for Contextual Bandits and Reinforcement Learning</title>
      <link>/post/modelselection/</link>
      <pubDate>Mon, 27 Jun 2022 00:00:00 +0000</pubDate>
      <guid>/post/modelselection/</guid>
      <description>&lt;h2 id=&#34;overview&#34;&gt;Overview&lt;/h2&gt;
&lt;p&gt;In many domains ranging from internet commerce, to robotics and computational biology, many algorithms have been developed that make decisions with the objective of maximizing a reward, while learning how to make better decisions in the future. In hopes of realizing this objective a vast literature focused on the study of Bandits and Reinforcement Learning algorithms has arisen. Although in most practical applications, precise knowledge of the nature of the problem faced by the learner may not be known in advance most of this work has chiefly focused on designing algorithms with provable regret guarantees that work under specific modeling assumptions. Less work has been spent on the problem of model selection where the objective is to design algorithms that can select in an online fashion the best suitable algorithm among a set of candidates to deal with a specific problem instance.&lt;/p&gt;
&lt;h2 id=&#34;papers&#34;&gt;Papers&lt;/h2&gt;
&lt;ul&gt;
&lt;li&gt;&lt;a href=&#34;https://arxiv.org/abs/2003.01704&#34;&gt;&lt;strong&gt;Model Selection in Contextual Stochastic Bandit Problems&lt;/strong&gt;&lt;/a&gt; - In this work we introduce the Stochastic CORRAL algorithm for model selection in stochastic contextual bandit problems and RL. The Stochastic CORRAL algorithm makes use of a master algorithm based on the &lt;a href=&#34;https://arxiv.org/abs/1612.06246&#34;&gt;CORRAL&lt;/a&gt; master. We also show some initial lower bounds showing the imposibility of deriving model selection results in a variety of settings.&lt;/li&gt;
&lt;li&gt;&lt;a href=&#34;https://arxiv.org/abs/2011.09750&#34;&gt;&lt;strong&gt;Online Model Selection for Reinforcement Learning with Function Approximation&lt;/strong&gt;&lt;/a&gt; - We introduce the Explore-Commit-Eliminate (&lt;strong&gt;ECE&lt;/strong&gt;) algorithm for model selection. This algorithm is suitable for model selection between stochastic bandit and reinforcement learning algorithms. &lt;strong&gt;ECE&lt;/strong&gt; is based on a simple misspecification test coupled with a play schedule reminiscent of epsilon greedy algorithms.&lt;/li&gt;
&lt;li&gt;&lt;a href=&#34;https://arxiv.org/abs/2011.09750&#34;&gt;&lt;strong&gt;Regret Bound Balancing and Elimination for Model Selection in Bandits and RL&lt;/strong&gt;&lt;/a&gt; - This work proposes the Regret Bound Balancing and Elimination Algorithm (&lt;strong&gt;RBBE&lt;/strong&gt;). This algorithm can be used for model selection among stochastic bandit and Reinforcement Learning algorithms. Similar to the &lt;strong&gt;ECE&lt;/strong&gt; algorithm, &lt;strong&gt;RBBE&lt;/strong&gt; is based on a simple misspecification test, in this case coupled with an algorithm play schedule based on &lt;strong&gt;balancing&lt;/strong&gt; the base learner&amp;rsquo;s regret bounds. Additionally, we show this approach can be extended to the setting of adversarially generated contexts in stochastic contextual linear bandits.&lt;/li&gt;
&lt;li&gt;&lt;a href=&#34;http://proceedings.mlr.press/v139/cutkosky21a/cutkosky21a.pdf&#34;&gt;&lt;strong&gt;Dynamic Balancing for Model Selection in Bandits and RL&lt;/strong&gt;&lt;/a&gt; - This work extends and refines the &lt;strong&gt;RBBE&lt;/strong&gt; approach, shaving polynomial factors in the number of models from the resulting model selection regret upper bounds.&lt;/li&gt;
&lt;li&gt;&lt;a href=&#34;https://www2.eecs.berkeley.edu/Pubs/TechRpts/2021/EECS-2021-164.pdf&#34;&gt;&lt;strong&gt;Model Selection for Contextual Bandits and Reinforcement Learning&lt;/strong&gt;&lt;/a&gt; - PhD Thesis. A compendium of the results introduced above.
-&lt;a href=&#34;https://arxiv.org/abs/2206.14912&#34;&gt;&lt;strong&gt;Best of Both Worlds Model Selection&lt;/strong&gt;&lt;/a&gt; - This works extends the regret balancing technology to work for adversarial bandits. It also explores the question of wether it is possible to obtain model selection and best of both worlds guarantees between adversarial and stochastic bandit scenarios.&lt;/li&gt;
&lt;/ul&gt;
</description>
    </item>
    
    <item>
      <title>Beyond the Standard Assumptions in Reinforcement Learning</title>
      <link>/post/beyondstandardrl/</link>
      <pubDate>Sun, 27 Mar 2022 00:00:00 +0000</pubDate>
      <guid>/post/beyondstandardrl/</guid>
      <description>&lt;h2 id=&#34;overview&#34;&gt;Overview&lt;/h2&gt;
&lt;p&gt;We study a theory of reinforcement learning (RL) in which the learner receives binary feedback only once at the end of an episode. While this is an extreme test case for theory, it is also arguably more representative of real-world applications than the traditional requirement in RL practice that the learner receive feedback at every time step. Indeed, in many real-world applications of reinforcement learning, such as self-driving cars and robotics, it is easier to evaluate whether a learner&amp;rsquo;s complete trajectory was either &amp;ldquo;good&amp;rdquo; or &amp;ldquo;bad,&amp;rdquo; but harder to provide a reward signal at each step. To show that learning is possible in this more challenging setting, we study the case where trajectory labels are generated by an unknown parametric model, and provide a statistically and computationally efficient algorithm that achieves sub-linear regret. We also study the corresponding Dueling Reinforcement Learning setting where the learner&amp;rsquo;s feedback comes in the form of noisy binary comparisons between trajectories.&lt;/p&gt;
&lt;h2 id=&#34;papers&#34;&gt;Papers&lt;/h2&gt;
&lt;ul&gt;
&lt;li&gt;&lt;a href=&#34;https://arxiv.org/abs/2105.14363&#34;&gt;&lt;strong&gt;On the Theory of Reinforcement Learning with Once-per-Episode Feedback&lt;/strong&gt;&lt;/a&gt;.&lt;/li&gt;
&lt;li&gt;&lt;a href=&#34;https://arxiv.org/abs/2111.04850&#34;&gt;&lt;strong&gt;Dueling RL: Reinforcement Learning with Trajectory Preferences&lt;/strong&gt;&lt;/a&gt;&lt;/li&gt;
&lt;/ul&gt;
</description>
    </item>
    
  </channel>
</rss>
