I'm running with VSync at a constant 60fps which is a calculated dt of 16.667ms (blank project with just a cube). When I call Time.deltaTime it varies wildly. Time.deltaTime can be as low as 2ms and as high as 31ms. The deltaTime "averages" out to be ~16.67ms. Why does Time.deltaTime vary wildly? Shouldn't it be a constant 16.667ms assuming the Update() loop has minimal computations and lots of idle time? My understanding is the Update loop runs as fast as possible (and will have variable Time.deltaTime) but if VSync is enabled then it should sync the Update rate to the display (60fps). Assuming the Update loop isn't overloaded and it can complete in < 16.667ms then shouldn't Time.deltaTime always be a constant 16.667ms? The question relates to camera jitter (camera following the player). Because the Time.deltaTime varies so much it is making the camera jitter. Thanks in advance.