Below are examples of simple scenes showing the main features of the Varjo headset. You can freely copy code from the examples to your own project. The example projects are supplied together with the Varjo Plugin for Unity.
Note: Make sure to copy the example to your project before modifying it. These examples are constantly updated and any changes may be overwritten.
This is a basic example for testing whether the headset is working and showing an image correctly. It can be used as a starting point to understand how to develop for the Varjo headset. The scripts demonstrated in this example are common for all the following examples explained in this section.
- Tracked devices
This is a script for enabling/disabling objects based on connectivity and assigned roles. It ensures that the Varjo headset and controllers are assigned correct roles in VR and behave as expected.
This is a script for handling the rendering of all SteamVR_Cameras.
This is a script for controlling in-game objects with tracked devices.
This example shows how to reset tracking position and record different events in the application.
This is a script for resetting the tracking origin position based on current user location by pressing a keyboard button. It can be useful for seated experiences where the user’s position is stationary. Settings for the example can be found in the inspector:
This is a script for logging events. It shows how to log info about the Application and System buttons on the Varjo Headset when pressed.
Note: You can only use the Application button on your Varjo headset for applications as the System button is reserved. However, you still can detect when the System button is pressed.
This example shows how to perform input and move a player around with basic teleportation and pick up scripts.
This is a basic teleport script for holding the touchpad of the controller. Teleportation is used for moving the user/player around the scene without causing motion sickness.
This is a throwing script for holding the controller trigger and releasing when throwing. It is used for interacting with objects, picking up objects using controllers, and throwing them with a throwing motion of the hand.
EYE TRACKING EXAMPLE
If you are not familiar with the eye tracking capabilities of the Varjo headset, it is recommended that you familiarize yourself with the information found in Eye Tracking with Varjo headset before proceeding. In order for eye tracking to work, please make sure you have enabled eye tracking in VarjoBase as described in Developing with 20/20 Eye Tracker section.
This example shows how to request eye tracking calibration and use the 20/20 Eye Tracker. Additionally, it shows how to poll button events from the headset as calibration is requested by pressing the Application button on the headset. The Application button on the Varjo headset is meant for interacting with VR apps. You can freely choose what this button does when using your application.
To access the Eye tracking code, check VarjoGazeRay.
This example also shows how eye tracking information can be logged into a file. Logged information includes information for frames, headset position and rotation, as well ass separate data for each eye’s pupil size, direction, and focus distance. All of the information is stored in the .csv file in /Assets/logs/ folder.
To access the Gaze Logging code, check VarjoGazeLog.
An example for accessing the Varjo headset’s Application button is demonstrated in the VarjoGazeCalibrationRequest file.
This shoots rays towards where the user is looking and sends hit events to VarjoGazeTargets. You may use it for projecting to the point at which the user is looking.
Note: This requires eye tracking calibration before using.
When the user presses the Application button or Space, an eye tracking calibration request is sent. Calibration needs to be repeated every time the headset is taken off and put back on in order to ensure the best eye tracking performance.
This handles sphere color change (target) when the user looks at it. You may use it as a starting point for changing behavior based on gaze interactions.
This logs gaze data after pressing a button and requires eye tracking calibration to be performed before using. Logged data is stored in the project’s folder.
Settings for logging can be found in the Unity editor and you can modify them according to your needs.
This example shows how to create a Unity Canvas UI for the headset. The UI works by moving your head or using hand controllers. Selections can be made with a keyboard, the headset’s Application button or the hand controller trigger. Buttons, toggles and dropdowns are supported.
This is a replacement for the Standalone Input Module. It is used with world space canvases in VR. It handles the keyboard, headset and hand controller for the canvas. It is recommended for use with VarjoPointer (see below) and requires a dummy camera for pointing. Rendering of the camera can be disabled.
This is a visual representation for UI pointing and works with both canvas and geometry. It changes input automatically to the most recently used or connected method.
This example shows how to display a video pass-through image from both cameras of XR-1 Developer Edition in the headset and all mixed reality related functionalities.
This controls for mixed reality related functionalities.
MR STREAM EXAMPLE
This example shows how to process a raw video pass-through image from the cameras of XR-1 Developer Edition and trigger an event every frame.
This example demonstrates the controls for mixed reality stream-related functionalities.