<?xml version="1.0" encoding="utf-8"?><feed xmlns="http://www.w3.org/2005/Atom" ><generator uri="https://jekyllrb.com/" version="3.10.0">Jekyll</generator><link href="https://yifeichen2024.github.io/feed.xml" rel="self" type="application/atom+xml" /><link href="https://yifeichen2024.github.io/" rel="alternate" type="text/html" /><updated>2025-10-29T21:01:31+00:00</updated><id>https://yifeichen2024.github.io/feed.xml</id><title type="html">Yifei Chen’s Website</title><subtitle>personal description</subtitle><author><name>Yifei Chen(陈逸飞)</name><email>yifeichen2026@@u.northwestern.edu</email><uri>https://yifeichen2024.github.io/</uri></author><entry><title type="html">Thoughts on Dexterous Manipulation Tasks</title><link href="https://yifeichen2024.github.io/posts/2025/05/Dexterous%20Tasks/" rel="alternate" type="text/html" title="Thoughts on Dexterous Manipulation Tasks" /><published>2025-05-04T00:00:00+00:00</published><updated>2025-05-04T00:00:00+00:00</updated><id>https://yifeichen2024.github.io/posts/2025/05/blog-post-1</id><content type="html" xml:base="https://yifeichen2024.github.io/posts/2025/05/Dexterous%20Tasks/"><![CDATA[<h2 id="️-two-types-of-dexterous-manipulation-tasks-functional-grasping-vs-in-hand-manipulation">🖐️ Two Types of Dexterous Manipulation Tasks: Functional Grasping vs. In-Hand Manipulation</h2>

<p>In my recent research and experimentation on dexterous manipulation, I’ve found that tasks in this domain can generally be divided into two major categories. These differ in objectives, strategy requirements, and especially in physical modeling. In this post, I attempt a preliminary classification and analysis of the two, along with some personal reflections and questions.</p>

<h3 id="-1-task-categorization">📌 1. Task Categorization</h3>

<p>Based on existing literature and practical trials, dexterous manipulation tasks can be roughly divided into:</p>

<h4 id="1-functional-grasping">1. <strong>Functional Grasping</strong></h4>

<p>The goal is for the robot to <strong>grasp an object stably for use in downstream interaction tasks</strong>, such as using tools, transporting items, or lifting containers. Typical examples include:</p>

<ul>
  <li>Grasping a cup and moving it to a target location</li>
  <li>Holding a screwdriver in preparation to tighten a screw</li>
  <li>Firmly holding a pen to write</li>
</ul>

<p>The core of functional grasping is <strong>stability</strong>, often evaluated via force-closure constraints or other grasp quality metrics. The main concern is <strong>whether the grasp can resist external perturbations</strong>—for instance, whether the object remains stable when external forces are applied after grasping.</p>

<p>An interesting example is using a pen to write. This task inherently involves both grasping and manipulation, but few systems today can perform this kind of stable interaction.</p>

<h4 id="2-in-hand-manipulation">2. <strong>In-Hand Manipulation</strong></h4>

<p>This task type involves <strong>dynamic adjustment</strong>. The goal is to perform fine pose tuning of an object within the hand through coordinated control of high-DOF fingers, often as a prerequisite for later tasks. Examples include:</p>

<ul>
  <li>Rotating a screw cap in-hand to align with a threaded hole</li>
  <li>Adjusting the orientation of a pen after picking it up to a writing-ready pose</li>
  <li>Rotating a cube in-palm to a target orientation</li>
</ul>

<p>Compared to functional grasping, this task is <strong>more dynamic and complex</strong>, often involving finger gaiting, sliding contacts, or even control of inertial responses within the hand. A simplified version often used in RL research is rotating a cube to a target orientation — a popular benchmark today.</p>

<h3 id="️-2-reflections-on-physical-modeling">⚙️ 2. Reflections on Physical Modeling</h3>

<p>These two task types differ significantly in terms of <strong>physical constraints</strong> and <strong>policy design</strong>:</p>

<h4 id="for-functional-grasping">For Functional Grasping:</h4>

<ul>
  <li>Introducing force-closure or wrench-based metrics into optimization or reward design is natural and effective;</li>
  <li>The main focus is grasp stability, especially under external contact;</li>
  <li>Physical modeling is largely quasi-static;</li>
  <li>Some interesting downstream tasks (e.g., writing) are highly sensitive to grasp stability, and incorporating physical modeling can significantly improve policy performance.</li>
</ul>

<h4 id="for-in-hand-manipulation">For In-Hand Manipulation:</h4>

<ul>
  <li>This is a <strong>more dynamic</strong> process;</li>
  <li>A major challenge is modeling and controlling object responses such as inertia and contact forces (which is also a tough problem for model-based control);</li>
  <li>While physical constraints are necessary, they’re not as <strong>directly effective</strong> as in grasping. Instead, we need integration of contact dynamics modeling and complex control strategies;</li>
  <li>There’s still no effective method for <strong>real-time pose adjustment based on tactile feedback</strong>.</li>
</ul>

<hr />

<h3 id="-summary">🧩 Summary</h3>

<table>
  <thead>
    <tr>
      <th><strong>Category</strong></th>
      <th><strong>Functional Grasping</strong></th>
      <th><strong>In-Hand Manipulation</strong></th>
    </tr>
  </thead>
  <tbody>
    <tr>
      <td><strong>Goal</strong></td>
      <td>Stable grasp for downstream tasks</td>
      <td>Fine in-hand pose adjustment</td>
    </tr>
    <tr>
      <td><strong>Characteristics</strong></td>
      <td>Quasi-static, force closure, stability</td>
      <td>Dynamic, finger gaiting, dynamic modeling</td>
    </tr>
    <tr>
      <td><strong>Challenge</strong></td>
      <td>Evaluating and optimizing grasp quality</td>
      <td>Modeling contact dynamics and regrasping</td>
    </tr>
    <tr>
      <td><strong>Examples</strong></td>
      <td>Writing, transporting, tool usage</td>
      <td>Rotating cap, pen adjustment, cube reorientation</td>
    </tr>
  </tbody>
</table>

<p>The second type of task demands richer <strong>dynamic contact information</strong>. If the robotic hand is equipped with quality <strong>tactile sensors</strong> that can provide contact-rich feedback, then the policy could leverage this to more effectively control object dynamics.</p>

<hr />

<h2 id="️-关于灵巧操作的两类任务思考functional-grasping-vs-in-hand-manipulation">🖐️ 关于灵巧操作的两类任务思考：Functional Grasping vs. In-Hand Manipulation</h2>

<p><strong>在我最近对 dexterous manipulation 的调研与实践中，我发现这类任务能够分成了两大类，它们在目标、策略需求上都展现出一定的差异，尤其是物理建模上。这里我尝试对它们做一个初步的分类与分析，也记录一些我目前的理解与疑问。</strong></p>

<h3 id="-一任务类型划分">📌 一、任务类型划分</h3>

<p>从已有文献和实际尝试来看，dexterous manipulation 的任务大致可以分为以下两类：</p>

<h4 id="1-functional-grasping功能性持握--稳定抓取">1. <strong>Functional Grasping</strong>（功能性持握 / 稳定抓取）</h4>

<p>目标是让机器人<strong>稳定地抓起物体，并用于后续的交互任务</strong>，如使用工具、搬运物体、举起容器等。典型任务如：</p>

<ul>
  <li>抓住一只杯子将其移动至指定位置</li>
  <li>拿起一把螺丝刀准备拧紧螺丝</li>
  <li>稳定持握一支笔进行写字</li>
</ul>

<p>Functional Grasping 的核心在于<strong>稳定性（stability）</strong>，可以借助 force-closure constraints 或其他 grasp evaluation metrics 来评估一个抓取的“质量”。这一任务关注的是<strong>能否抓住并抵抗外部扰动</strong>，比如机器人持握物体后，能否在遇到外部力施加时仍保持稳定。</p>

<p>比较有意思的例子是机器人用笔写字——这本质上既是 grasp 也是 manipulation。目前很少有稳定的交互。</p>

<h4 id="2-in-hand-manipulation手中操作--灵巧操作">2. <strong>In-Hand Manipulation</strong>（手中操作 / 灵巧操作）</h4>

<p>这类任务更偏向于一个<strong>动态调整的过程</strong>。目标是通过高自由度手指之间的协调，对物体进行位姿上的微调，以便于后续任务执行。典型示例包括：</p>

<ul>
  <li>手中旋转一个螺丝帽以对齐螺纹孔</li>
  <li>拿起一支笔后在指间调整至便于书写的姿态</li>
  <li>将一个小方块在掌中旋转至目标姿态</li>
</ul>

<p>相比 Functional Grasping，这类任务<strong>更复杂、更动态</strong>，通常涉及 finger gaiting、滑动接触、甚至控制物体在手内的惯性响应。因此，policy 本身需要对 object dynamics 有更深的建模和理解。一个常用的简化版本任务是旋转一个 cube 至目标姿态 —— 这是当前 RL manipulation 研究中较为流行的 benchmark。</p>

<h3 id="️-二关于物理建模的思考">⚙️ 二、关于物理建模的思考</h3>

<p>这两类任务在 <strong>物理建模（physical constraints）</strong> 和 <strong>策略设计（policy learning）</strong> 上有着显著不同：</p>

<h4 id="对于-functional-grasping">对于 Functional Grasping：</h4>

<ul>
  <li>引入 force-closure 或 wrench-based metrics 作为 optimization/reward，是较自然、直接的方式；</li>
  <li>重点在于提升 grasp 稳定性，尤其是在面对 external contact 时能否维持持握；</li>
  <li>物理建模更偏向于 quasi-static；</li>
  <li>一些 cool 的 downstream task（如书写）需要对 grasp 稳定性极为敏感，而这可以通过物理建模来提升策略性能。</li>
</ul>

<h4 id="对于-in-hand-manipulation">对于 In-Hand Manipulation：</h4>

<ul>
  <li>是一个更加动态（dynamic）的过程；</li>
  <li>难点在于对物体的动态响应（如转动惯量、接触力滑动）进行建模和控制（这一点也是model-based control比较困难的点）；Learning的policy也有这样一个</li>
  <li>引入物理约束虽然必要，但效果不如在 grasping 中那样“直接有效”，更需要融合 contact dynamics 模型与更复杂的 control strategy；</li>
  <li>如何基于 tactile feedback 做在线姿态调整目前还没有很好的方法。</li>
</ul>

<hr />

<h3 id="-总结">🧩 总结</h3>

<table>
  <thead>
    <tr>
      <th><strong>类别</strong></th>
      <th><strong>Functional Grasping</strong></th>
      <th><strong>In-Hand Manipulation</strong></th>
    </tr>
  </thead>
  <tbody>
    <tr>
      <td>目标</td>
      <td>稳定持握用于后续交互</td>
      <td>手中物体精细姿态调整</td>
    </tr>
    <tr>
      <td>特征</td>
      <td>Quasi-static, 力闭合, 稳定性为主</td>
      <td>Dynamic, finger gaiting, 动态建模</td>
    </tr>
    <tr>
      <td>难点</td>
      <td>如何评估 grasp 质量并优化持握稳定性</td>
      <td>如何建模 contact dynamics 和控制再定位策略</td>
    </tr>
    <tr>
      <td>示例</td>
      <td>写字、搬运、使用工具</td>
      <td>转动螺丝帽、笔姿态调整、cube reorientation</td>
    </tr>
  </tbody>
</table>

<p>第二类task对于动态的contact信息获取要求更高一些，如果hand有好的sensor来提供contact-rich information, 那policy应该能够利用这些信息来控制object的dynamics.</p>]]></content><author><name>Yifei Chen(陈逸飞)</name><email>yifeichen2026@@u.northwestern.edu</email><uri>https://yifeichen2024.github.io/</uri></author><category term="Dexterous Manipulation" /><category term="Reinforcement Learning" /><category term="Grasp Stability" /><category term="Contact-Rich Tasks" /><summary type="html"><![CDATA[🖐️ Two Types of Dexterous Manipulation Tasks: Functional Grasping vs. In-Hand Manipulation]]></summary></entry><entry><title type="html">Learning Stability: Reflections on Physical Constraints in Dexterous Manipulation</title><link href="https://yifeichen2024.github.io/posts/2025/05/learning-stability/" rel="alternate" type="text/html" title="Learning Stability: Reflections on Physical Constraints in Dexterous Manipulation" /><published>2025-05-02T00:00:00+00:00</published><updated>2025-05-02T00:00:00+00:00</updated><id>https://yifeichen2024.github.io/posts/2025/05/blog-post-1</id><content type="html" xml:base="https://yifeichen2024.github.io/posts/2025/05/learning-stability/"><![CDATA[<p>Over the past few months, I’ve been exploring how physical constraints—like force closure and contact stability—can be introduced directly into the learning process for in-hand manipulation. While most recent progress in dexterous robotic skills has focused on end-to-end reinforcement learning using vision or tactile images, I found myself wondering: what if we teach policies to <em>care</em> about physical grasp stability?</p>

<h3 id="stability-isnt-just-a-side-effect">Stability isn’t just a side effect</h3>

<p>Most in-hand policies today can spin a cube, but they don’t <em>really</em> hold it. Often, the object is precariously balanced on a fingertip or nudged by the palm, barely resisting even small external disturbances. In contrast, when humans manipulate objects, we instinctively adjust our grasp in response to micro-contact cues—especially force and torque at our fingertips.</p>

<p>That idea became the starting point of this workshop paper. Instead of treating grasp stability as a side effect of task success, we made it part of the objective: using force-torque sensing and force-closure-based rewards, we trained a policy that explicitly seeks to maintain a physically stable grasp during in-hand manipulation.</p>

<h3 id="what-we-learned">What we learned</h3>

<p>It turns out, adding physical constraints had some interesting effects:</p>

<ul>
  <li><strong>Faster convergence</strong>: The policy trained with grasp stability rewards reached functional behavior in far fewer steps.</li>
  <li><strong>Smoother behavior</strong>: The hand didn’t just spin the cube—it cradled it, adjusted to shifting forces, and moved more like a hand, not a claw.</li>
  <li><strong>Trade-offs emerged</strong>: Stability and dexterity aren’t always aligned. By rewarding force closure, the policy occasionally sacrificed rotational speed to keep its grip secure.</li>
</ul>

<h3 id="looking-ahead">Looking ahead</h3>

<p>This is still early-stage work—just a workshop paper—but the idea of integrating physically meaningful objectives into RL for manipulation feels promising. In future work, I’m curious to see how these strategies scale with more complex objects, external perturbations, and real-world sim-to-real deployment.</p>

<p>Maybe one day, we’ll build not just policies that solve tasks, but hands that <em>understand</em> what it means to hold on.</p>]]></content><author><name>Yifei Chen(陈逸飞)</name><email>yifeichen2026@@u.northwestern.edu</email><uri>https://yifeichen2024.github.io/</uri></author><category term="Dexterous Manipulation" /><category term="Reinforcement Learning" /><category term="Grasp Stability" /><category term="Contact-Rich Tasks" /><summary type="html"><![CDATA[Over the past few months, I’ve been exploring how physical constraints—like force closure and contact stability—can be introduced directly into the learning process for in-hand manipulation. While most recent progress in dexterous robotic skills has focused on end-to-end reinforcement learning using vision or tactile images, I found myself wondering: what if we teach policies to care about physical grasp stability?]]></summary></entry></feed>