Learning models of physical systems can sometimes be difficult. Vanilla neural networks—like residual networks—particularly struggle to learn invariant properties like the conservation of energy which is fundamental to physical systems. To counteract this, a number of recent works such as *Hamiltonian Neural Networks* and *Variational Integrator Networks* introduce *inductive biases*, also referred to as *physics priors*, which improve reliability of predictions and speed up learning. These network classes exhibit good approximation behavior for continuous physical systems but they are fundamentally limited to smooth dynamics, and not designed to handle non-smooth physical behavior, such as resolving collision events between different objects. Such behavior is of key interest in robotics, and other areas of engineering. In this work, we explore neural network architectures designed for accurately modeling contact dynamics, which incorporate the structure necessary to reliably resolve non-smooth collision events.

# Contact Dynamics

*Contact dynamics* are a class of equations which describe the motion of physical systems consisting of multiple solid objects which interact with each other and their environment. One of these equations’ defining features is that when two objects collide, their velocities change direction *instantaneously* in a non-smooth manner—this describes, for instance, how a bouncing ball immediately changes direction upon hitting the ground, resulting from a transfer of momentum and other physical considerations.

Because of the resulting non-smoothness and non-linearity, contacts dynamics are considered notoriously difficult to compute. For example, a numerical regime must decide whether to enforce non-interpenetration constraints by precisely calculating contact times using an optimization procedure, or instead allow physically incorrect interpenetration. Below, we illustrate sample numerical trajectory of a bouncing ball.

These difficulties are further compounded when the equations of motion of the system under study are unknown, which occurs in robotics when learning to interact with unknown objects. The aim of this work is to explore neural network architectures which are capable of accurately modeling such systems.

# Central Difference Lagrange Networks

We begin with the perspective of *neural ordinary differential equations*, which view deep networks as discretizations of continuous-time dynamical systems. Our system’s state is defined as velocity pairs `$(\mathbf{Q}, \mathbf{\dot Q})$`

. In-between contact events, the trajectories follow the *Euler-Lagrange equations*

```
$$
\frac{\partial L}{\partial \mathbf{Q}} - \frac{\mathrm{d}}{\mathrm{d}t} \frac{\partial L}{\partial \mathbf{\dot Q}} = 0
$$
```

where `$L=T-V$`

is the Lagrangian with the potential `$V$`

and kinetic energy `$T$`

. At contact times, the above equations do not hold, and a set of instantaneous *transfer of momentum* equations apply instead.

To model these dynamics, we adopt an approach similar in spirit to *Variational Integrator Networks*^{1} by modeling `$V$`

using a fully connected neural network and discretizing the resulting equations of motion to construct a recurrent network architecture. The choice of differential equation class and discretization scheme determines the inductive bias that is introduced. These biases can include physical properties such as conservation of momentum and of energy, as well as other fundamental mechanical characteristics.

To design a scheme for contact dynamics, we employ the Central Difference–Lagrange (CDL) scheme,^{2} whose equations form the basis of our network architecture. Between contact times, the dynamics evolve smoothly, which we denoted by `$(\cdot)^S$`

, in a manner that mirrors variational integrator networks. During contact events, these equations are augmented by a contact term, denoted by `$(\cdot)^{C}$`

, that handles the transfer of momentum and makes sure that (1) Newton’s restitution law, as well as (2) the law of conservation of momentum both hold. Below, we illustrate how the different states in the CDL-Network are calculated.

**Illustration:**a CD-Lagrange network. Here, we begin from initial states

`$(\mathbf{Q}_0,\mathbf{\dot Q}_{\frac{1}{2}})$`

. We calculate the next position `$\mathbf{Q}_1$`

, and proceed to calculate the next velocity `$\mathbf{\dot Q}_{1 + \frac{1}{2}} = \mathbf{\dot Q}_{1 + \frac{1}{2}}^S + \mathbf{\dot Q}_{1 + \frac{1}{2}}^C$`

as a sum of smooth and contact terms. These terms are in turn calculated using the conservative forces `$\mathbf{F}$`

and the impulse `$\mathbf{I}$`

, which are calculated from the parameterized Lagrangian, whose potential energy is given by a fully connected network.# Touch Feedback

In an unknown system, one must determine when a system’s trajectory evolves according to smooth dynamics, and when it evolves according to contact events. Our results suggest that external touch feedback—such as that available from an idealized touch sensor—is necessary to make the problem tractable. This is handled by introducing a *contact network* `$\hat{c}$`

which learns to predict contact events, using training data obtained via said touch sensor. Without this, the network struggles to differentiate noise from real contact events, as small contact events and noise both generate similar data. We illustrate this behavior on the following examples.

**Bouncing ball**. Here, we see that without touch feedback, the CD–Lagrange network struggles to learn the contact events properly and instead approximates the discontinuous behavior using the potential network. With training data that includes touch feedback, we see that performance, shown below, is significantly better.**(a)**Bouncing ball: CD–Lagrange**Newton’s cradle**. As an additional baseline, we employ a vanilla residual network (ResNet) as well as a residual network with additional contact inputs (ResNet contact). The CD-Lagrange network, shown below, exhibits the best approximation behavior and learn both the potential and contact events more accurately than the residual networks.**(a)**Newton’s cradle: CD–Lagrange**(b)**Newton’s cradle: ResNet

# Summary

State-of-the-art physics-inspired neural networks generally struggle to learn contact dynamics. Central-Difference-Lagrange networks are a class of networks that not only exhibit strong conservation properties, comparable to other physically structured neural networks, but also allow accurate learning of contact dynamics from observed data. In this regime, the information available to the network when making predictions has a significant effect on performance: the addition of touch feedback sensor data ensures that noise and contact events are correctly differentiated. We hope these contributions enable neural network models to be used in wider settings.

# References

^{1}

S. Sæmundsson, A. Terenin, K. Hofmann, M. P. Deisenroth. Variational Integrator Networks for Physically Structured Embeddings. AISTATS, 2020.

^{2}

F.-E. Fekak, M. Brun, A. Gravouil, and B. Depale. A new heterogeneous asynchronous explicit–implicit time integrator for nonsmooth dynamics. Computational Mechanics, 60(1):1–21, 2017.