Traditionally, IL/IRL methods assume access to expert demonstrations and minimize some divergence between policy and expert's trajectory distribution. However, in many cases, it may be easier to directly specify the state distribution explicitly or via samples of the desired behavior rather than to provide fully-specified demonstrations (with actions) of the desired behavior. For example, in a safety-critical application, it may be easier to specify that the expert never visits some unsafe states rather than tweaking reward to penalize safety violations. Similarly, we can specify a uniform density over the whole state space for exploration tasks, or a Gaussian centered at the goal for goal-reaching tasks.