virtualx-engine/modules/openxr/openxr_api.cpp

Ignoring revisions in .git-blame-ignore-revs. Click here to bypass and see the normal blame view.

3546 lines
117 KiB
C++
Raw Normal View History

2021-12-14 02:44:12 +01:00
/**************************************************************************/
/* openxr_api.cpp */
/**************************************************************************/
/* This file is part of: */
/* GODOT ENGINE */
/* https://godotengine.org */
/**************************************************************************/
/* Copyright (c) 2014-present Godot Engine contributors (see AUTHORS.md). */
/* Copyright (c) 2007-2014 Juan Linietsky, Ariel Manzur. */
/* */
/* Permission is hereby granted, free of charge, to any person obtaining */
/* a copy of this software and associated documentation files (the */
/* "Software"), to deal in the Software without restriction, including */
/* without limitation the rights to use, copy, modify, merge, publish, */
/* distribute, sublicense, and/or sell copies of the Software, and to */
/* permit persons to whom the Software is furnished to do so, subject to */
/* the following conditions: */
/* */
/* The above copyright notice and this permission notice shall be */
/* included in all copies or substantial portions of the Software. */
/* */
/* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, */
/* EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF */
/* MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. */
/* IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY */
/* CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, */
/* TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE */
/* SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE. */
/**************************************************************************/
#include "openxr_api.h"
#include "extensions/openxr_extension_wrapper_extension.h"
#include "openxr_interface.h"
2021-12-14 02:44:12 +01:00
#include "openxr_util.h"
#include "core/config/engine.h"
#include "core/config/project_settings.h"
#include "core/os/memory.h"
#include "core/version.h"
#ifdef TOOLS_ENABLED
#include "editor/editor_settings.h"
#endif
#include "openxr_platform_inc.h"
2022-10-23 03:29:15 +02:00
2021-12-14 02:44:12 +01:00
#ifdef VULKAN_ENABLED
2023-12-27 04:49:25 +01:00
#include "extensions/platform/openxr_vulkan_extension.h"
2021-12-14 02:44:12 +01:00
#endif
2023-07-18 16:02:47 +02:00
#if defined(GLES3_ENABLED) && !defined(MACOS_ENABLED)
2023-12-27 04:49:25 +01:00
#include "extensions/platform/openxr_opengl_extension.h"
2022-10-23 03:29:15 +02:00
#endif
#include "extensions/openxr_composition_layer_depth_extension.h"
#include "extensions/openxr_debug_utils_extension.h"
2024-03-26 02:57:26 +01:00
#include "extensions/openxr_eye_gaze_interaction.h"
#include "extensions/openxr_fb_display_refresh_rate_extension.h"
#include "extensions/openxr_fb_foveation_extension.h"
#include "extensions/openxr_fb_update_swapchain_extension.h"
2024-02-19 03:06:36 +01:00
#include "extensions/openxr_hand_tracking_extension.h"
2022-03-09 09:09:25 +01:00
#ifdef ANDROID_ENABLED
#define OPENXR_LOADER_NAME "libopenxr_loader.so"
#endif
////////////////////////////////////
// OpenXRAPI::OpenXRSwapChainInfo
Vector<OpenXRAPI::OpenXRSwapChainInfo> OpenXRAPI::OpenXRSwapChainInfo::free_queue;
bool OpenXRAPI::OpenXRSwapChainInfo::create(XrSwapchainCreateFlags p_create_flags, XrSwapchainUsageFlags p_usage_flags, int64_t p_swapchain_format, uint32_t p_width, uint32_t p_height, uint32_t p_sample_count, uint32_t p_array_size) {
OpenXRAPI *openxr_api = OpenXRAPI::get_singleton();
ERR_FAIL_NULL_V(openxr_api, false);
XrSession xr_session = openxr_api->get_session();
ERR_FAIL_COND_V(xr_session == XR_NULL_HANDLE, false);
OpenXRGraphicsExtensionWrapper *xr_graphics_extension = openxr_api->get_graphics_extension();
ERR_FAIL_NULL_V(xr_graphics_extension, false);
// We already have a swapchain?
ERR_FAIL_COND_V(swapchain != XR_NULL_HANDLE, false);
XrResult result;
void *next_pointer = nullptr;
for (OpenXRExtensionWrapper *wrapper : openxr_api->get_registered_extension_wrappers()) {
void *np = wrapper->set_swapchain_create_info_and_get_next_pointer(next_pointer);
if (np != nullptr) {
next_pointer = np;
}
}
XrSwapchainCreateInfo swapchain_create_info = {
XR_TYPE_SWAPCHAIN_CREATE_INFO, // type
next_pointer, // next
p_create_flags, // createFlags
p_usage_flags, // usageFlags
p_swapchain_format, // format
p_sample_count, // sampleCount
p_width, // width
p_height, // height
1, // faceCount
p_array_size, // arraySize
1 // mipCount
};
XrSwapchain new_swapchain;
result = openxr_api->xrCreateSwapchain(xr_session, &swapchain_create_info, &new_swapchain);
if (XR_FAILED(result)) {
print_line("OpenXR: Failed to get swapchain [", openxr_api->get_error_string(result), "]");
return false;
}
if (!xr_graphics_extension->get_swapchain_image_data(new_swapchain, p_swapchain_format, p_width, p_height, p_sample_count, p_array_size, &swapchain_graphics_data)) {
openxr_api->xrDestroySwapchain(new_swapchain);
return false;
}
swapchain = new_swapchain;
return true;
}
void OpenXRAPI::OpenXRSwapChainInfo::queue_free() {
if (image_acquired) {
release();
}
if (swapchain != XR_NULL_HANDLE) {
free_queue.push_back(*this);
swapchain_graphics_data = nullptr;
swapchain = XR_NULL_HANDLE;
}
}
void OpenXRAPI::OpenXRSwapChainInfo::free_queued() {
for (OpenXRAPI::OpenXRSwapChainInfo &swapchain_info : free_queue) {
swapchain_info.free();
}
free_queue.clear();
}
void OpenXRAPI::OpenXRSwapChainInfo::free() {
OpenXRAPI *openxr_api = OpenXRAPI::get_singleton();
ERR_FAIL_NULL(openxr_api);
if (image_acquired) {
release();
}
if (openxr_api->get_graphics_extension() && swapchain_graphics_data != nullptr) {
openxr_api->get_graphics_extension()->cleanup_swapchain_graphics_data(&swapchain_graphics_data);
}
if (swapchain != XR_NULL_HANDLE) {
openxr_api->xrDestroySwapchain(swapchain);
swapchain = XR_NULL_HANDLE;
}
}
bool OpenXRAPI::OpenXRSwapChainInfo::acquire(bool &p_should_render) {
ERR_FAIL_COND_V(image_acquired, true); // This was not released when it should be, error out and reuse...
OpenXRAPI *openxr_api = OpenXRAPI::get_singleton();
ERR_FAIL_NULL_V(openxr_api, false);
XrResult result;
if (!skip_acquire_swapchain) {
XrSwapchainImageAcquireInfo swapchain_image_acquire_info = {
XR_TYPE_SWAPCHAIN_IMAGE_ACQUIRE_INFO, // type
nullptr // next
};
result = openxr_api->xrAcquireSwapchainImage(swapchain, &swapchain_image_acquire_info, &image_index);
if (!XR_UNQUALIFIED_SUCCESS(result)) {
// Make sure end_frame knows we need to submit an empty frame
p_should_render = false;
if (XR_FAILED(result)) {
// Unexpected failure, log this!
print_line("OpenXR: failed to acquire swapchain image [", openxr_api->get_error_string(result), "]");
return false;
} else {
// In this scenario we silently fail, the XR runtime is simply not ready yet to acquire the swapchain.
return false;
}
}
}
XrSwapchainImageWaitInfo swapchain_image_wait_info = {
XR_TYPE_SWAPCHAIN_IMAGE_WAIT_INFO, // type
nullptr, // next
1000000000 // 1s timeout in nanoseconds
};
// Wait for a maximum of 10 seconds before calling it a critical failure...
for (int retry = 0; retry < 10; retry++) {
result = openxr_api->xrWaitSwapchainImage(swapchain, &swapchain_image_wait_info);
if (result != XR_TIMEOUT_EXPIRED) {
break;
}
WARN_PRINT("OpenXR: timed out waiting for swapchain image.");
}
if (!XR_UNQUALIFIED_SUCCESS(result)) {
// Make sure end_frame knows we need to submit an empty frame
p_should_render = false;
if (XR_FAILED(result)) {
// Unexpected failure, log this!
print_line("OpenXR: failed to wait for swapchain image [", openxr_api->get_error_string(result), "]");
return false;
} else {
WARN_PRINT("OpenXR: couldn't to wait for swapchain but not a complete error [" + openxr_api->get_error_string(result) + "]");
// Make sure to skip trying to acquire the swapchain image in the next frame
skip_acquire_swapchain = true;
return false;
}
} else {
skip_acquire_swapchain = false;
}
image_acquired = true;
return true;
}
bool OpenXRAPI::OpenXRSwapChainInfo::release() {
if (!image_acquired) {
// Already released or never acquired.
return true;
}
image_acquired = false; // Regardless if we succeed or not, consider this released.
OpenXRAPI *openxr_api = OpenXRAPI::get_singleton();
ERR_FAIL_NULL_V(openxr_api, false);
XrSwapchainImageReleaseInfo swapchain_image_release_info = {
XR_TYPE_SWAPCHAIN_IMAGE_RELEASE_INFO, // type
nullptr // next
};
XrResult result = openxr_api->xrReleaseSwapchainImage(swapchain, &swapchain_image_release_info);
if (XR_FAILED(result)) {
print_line("OpenXR: failed to release swapchain image! [", openxr_api->get_error_string(result), "]");
return false;
}
return true;
}
RID OpenXRAPI::OpenXRSwapChainInfo::get_image() {
OpenXRAPI *openxr_api = OpenXRAPI::get_singleton();
if (image_acquired && openxr_api && openxr_api->get_graphics_extension()) {
return OpenXRAPI::get_singleton()->get_graphics_extension()->get_texture(swapchain_graphics_data, image_index);
} else {
return RID();
}
}
////////////////////////////////////
// OpenXRAPI
2021-12-14 02:44:12 +01:00
OpenXRAPI *OpenXRAPI::singleton = nullptr;
Vector<OpenXRExtensionWrapper *> OpenXRAPI::registered_extension_wrappers;
2021-12-14 02:44:12 +01:00
2022-03-17 00:24:47 +01:00
bool OpenXRAPI::openxr_is_enabled(bool p_check_run_in_editor) {
2021-12-14 02:44:12 +01:00
// @TODO we need an overrule switch so we can force enable openxr, i.e run "godot --openxr_enabled"
2022-03-17 00:24:47 +01:00
if (Engine::get_singleton()->is_editor_hint() && p_check_run_in_editor) {
2021-12-14 02:44:12 +01:00
// Disabled for now, using XR inside of the editor we'll be working on during the coming months.
return false;
} else {
if (XRServer::get_xr_mode() == XRServer::XRMODE_DEFAULT) {
return GLOBAL_GET("xr/openxr/enabled");
} else {
return XRServer::get_xr_mode() == XRServer::XRMODE_ON;
}
2021-12-14 02:44:12 +01:00
}
}
String OpenXRAPI::get_default_action_map_resource_name() {
String name = GLOBAL_GET("xr/openxr/default_action_map");
return name;
}
String OpenXRAPI::get_error_string(XrResult result) const {
2021-12-14 02:44:12 +01:00
if (XR_SUCCEEDED(result)) {
return String("Succeeded");
}
if (instance == XR_NULL_HANDLE) {
Array args;
args.push_back(Variant(result));
return String("Error code {0}").format(args);
}
char resultString[XR_MAX_RESULT_STRING_SIZE];
xrResultToString(instance, result, resultString);
return String(resultString);
}
String OpenXRAPI::get_swapchain_format_name(int64_t p_swapchain_format) const {
// This is rendering engine dependent...
2021-12-14 02:44:12 +01:00
if (graphics_extension) {
return graphics_extension->get_swapchain_format_name(p_swapchain_format);
}
return String("Swapchain format ") + String::num_int64(int64_t(p_swapchain_format));
}
void OpenXRAPI::set_object_name(XrObjectType p_object_type, uint64_t p_object_handle, const String &p_object_name) {
OpenXRDebugUtilsExtension *debug_utils = OpenXRDebugUtilsExtension::get_singleton();
if (!debug_utils || !debug_utils->get_active()) {
// Not enabled/active? Ignore.
return;
}
debug_utils->set_object_name(p_object_type, p_object_handle, p_object_name.utf8().get_data());
}
void OpenXRAPI::begin_debug_label_region(const String &p_label_name) {
OpenXRDebugUtilsExtension *debug_utils = OpenXRDebugUtilsExtension::get_singleton();
if (!debug_utils || !debug_utils->get_active()) {
// Not enabled/active? Ignore.
return;
}
debug_utils->begin_debug_label_region(p_label_name.utf8().get_data());
}
void OpenXRAPI::end_debug_label_region() {
OpenXRDebugUtilsExtension *debug_utils = OpenXRDebugUtilsExtension::get_singleton();
if (!debug_utils || !debug_utils->get_active()) {
// Not enabled/active? Ignore.
return;
}
debug_utils->end_debug_label_region();
}
void OpenXRAPI::insert_debug_label(const String &p_label_name) {
OpenXRDebugUtilsExtension *debug_utils = OpenXRDebugUtilsExtension::get_singleton();
if (!debug_utils || !debug_utils->get_active()) {
// Not enabled/active? Ignore.
return;
}
debug_utils->insert_debug_label(p_label_name.utf8().get_data());
}
2021-12-14 02:44:12 +01:00
bool OpenXRAPI::load_layer_properties() {
// This queries additional layers that are available and can be initialized when we create our OpenXR instance
2021-12-14 02:44:12 +01:00
if (layer_properties != nullptr) {
// already retrieved this
return true;
}
// Note, instance is not yet setup so we can't use get_error_string to retrieve our error
XrResult result = xrEnumerateApiLayerProperties(0, &num_layer_properties, nullptr);
ERR_FAIL_COND_V_MSG(XR_FAILED(result), false, "OpenXR: Failed to enumerate number of api layer properties");
layer_properties = (XrApiLayerProperties *)memalloc(sizeof(XrApiLayerProperties) * num_layer_properties);
ERR_FAIL_NULL_V(layer_properties, false);
for (uint32_t i = 0; i < num_layer_properties; i++) {
layer_properties[i].type = XR_TYPE_API_LAYER_PROPERTIES;
layer_properties[i].next = nullptr;
}
result = xrEnumerateApiLayerProperties(num_layer_properties, &num_layer_properties, layer_properties);
ERR_FAIL_COND_V_MSG(XR_FAILED(result), false, "OpenXR: Failed to enumerate api layer properties");
for (uint32_t i = 0; i < num_layer_properties; i++) {
print_verbose(String("OpenXR: Found OpenXR layer ") + layer_properties[i].layerName);
2021-12-14 02:44:12 +01:00
}
return true;
}
bool OpenXRAPI::load_supported_extensions() {
// This queries supported extensions that are available and can be initialized when we create our OpenXR instance
2021-12-14 02:44:12 +01:00
if (supported_extensions != nullptr) {
// already retrieved this
return true;
}
// Note, instance is not yet setup so we can't use get_error_string to retrieve our error
XrResult result = xrEnumerateInstanceExtensionProperties(nullptr, 0, &num_supported_extensions, nullptr);
ERR_FAIL_COND_V_MSG(XR_FAILED(result), false, "OpenXR: Failed to enumerate number of extension properties");
supported_extensions = (XrExtensionProperties *)memalloc(sizeof(XrExtensionProperties) * num_supported_extensions);
ERR_FAIL_NULL_V(supported_extensions, false);
// set our types
for (uint32_t i = 0; i < num_supported_extensions; i++) {
supported_extensions[i].type = XR_TYPE_EXTENSION_PROPERTIES;
supported_extensions[i].next = nullptr;
}
result = xrEnumerateInstanceExtensionProperties(nullptr, num_supported_extensions, &num_supported_extensions, supported_extensions);
ERR_FAIL_COND_V_MSG(XR_FAILED(result), false, "OpenXR: Failed to enumerate extension properties");
for (uint32_t i = 0; i < num_supported_extensions; i++) {
print_verbose(String("OpenXR: Found OpenXR extension ") + supported_extensions[i].extensionName);
2021-12-14 02:44:12 +01:00
}
return true;
}
bool OpenXRAPI::is_extension_supported(const String &p_extension) const {
2021-12-14 02:44:12 +01:00
for (uint32_t i = 0; i < num_supported_extensions; i++) {
if (supported_extensions[i].extensionName == p_extension) {
2021-12-14 02:44:12 +01:00
return true;
}
}
return false;
}
bool OpenXRAPI::is_extension_enabled(const String &p_extension) const {
CharString extension = p_extension.ascii();
2022-09-23 05:20:12 +02:00
for (int i = 0; i < enabled_extensions.size(); i++) {
if (strcmp(enabled_extensions[i].ptr(), extension.ptr()) == 0) {
return true;
2022-09-23 05:20:12 +02:00
}
}
return false;
}
bool OpenXRAPI::is_top_level_path_supported(const String &p_toplevel_path) {
String required_extension = OpenXRInteractionProfileMetadata::get_singleton()->get_top_level_extension(p_toplevel_path);
// If unsupported is returned we likely have a misspelled interaction profile path in our action map. Always output that as an error.
ERR_FAIL_COND_V_MSG(required_extension == XR_PATH_UNSUPPORTED_NAME, false, "OpenXR: Unsupported toplevel path " + p_toplevel_path);
if (required_extension == "") {
// no extension needed, core top level are always "supported", they just won't be used if not really supported
return true;
}
if (!is_extension_enabled(required_extension)) {
// It is very likely we have top level paths for which the extension is not available so don't flood the logs with unnecessary spam.
print_verbose("OpenXR: Top level path " + p_toplevel_path + " requires extension " + required_extension);
return false;
}
return true;
}
bool OpenXRAPI::is_interaction_profile_supported(const String &p_ip_path) {
String required_extension = OpenXRInteractionProfileMetadata::get_singleton()->get_interaction_profile_extension(p_ip_path);
// If unsupported is returned we likely have a misspelled interaction profile path in our action map. Always output that as an error.
ERR_FAIL_COND_V_MSG(required_extension == XR_PATH_UNSUPPORTED_NAME, false, "OpenXR: Unsupported interaction profile " + p_ip_path);
if (required_extension == "") {
// no extension needed, core interaction profiles are always "supported", they just won't be used if not really supported
return true;
}
if (!is_extension_enabled(required_extension)) {
// It is very likely we have interaction profiles for which the extension is not available so don't flood the logs with unnecessary spam.
print_verbose("OpenXR: Interaction profile " + p_ip_path + " requires extension " + required_extension);
return false;
}
return true;
}
bool OpenXRAPI::interaction_profile_supports_io_path(const String &p_ip_path, const String &p_io_path) {
if (!is_interaction_profile_supported(p_ip_path)) {
return false;
}
const OpenXRInteractionProfileMetadata::IOPath *io_path = OpenXRInteractionProfileMetadata::get_singleton()->get_io_path(p_ip_path, p_io_path);
// If the io_path is not part of our metadata we've likely got a misspelled name or a bad action map, report
ERR_FAIL_NULL_V_MSG(io_path, false, "OpenXR: Unsupported io path " + String(p_ip_path) + String(p_io_path));
if (io_path->openxr_extension_name == "") {
// no extension needed, core io paths are always "supported", they just won't be used if not really supported
return true;
}
if (!is_extension_enabled(io_path->openxr_extension_name)) {
// It is very likely we have io paths for which the extension is not available so don't flood the logs with unnecessary spam.
print_verbose("OpenXR: IO path " + String(p_ip_path) + String(p_io_path) + " requires extension " + io_path->openxr_extension_name);
return false;
}
2022-09-23 05:20:12 +02:00
return true;
}
2021-12-14 02:44:12 +01:00
void OpenXRAPI::copy_string_to_char_buffer(const String p_string, char *p_buffer, int p_buffer_len) {
CharString char_string = p_string.utf8();
int len = char_string.length();
if (len < p_buffer_len - 1) {
// was having weird CI issues with strcpy so....
memcpy(p_buffer, char_string.get_data(), len);
p_buffer[len] = '\0';
} else {
memcpy(p_buffer, char_string.get_data(), p_buffer_len - 1);
p_buffer[p_buffer_len - 1] = '\0';
}
}
bool OpenXRAPI::create_instance() {
// Create our OpenXR instance, this will query any registered extension wrappers for extensions we need to enable.
// Append the extensions requested by the registered extension wrappers.
HashMap<String, bool *> requested_extensions;
2021-12-14 02:44:12 +01:00
for (OpenXRExtensionWrapper *wrapper : registered_extension_wrappers) {
const HashMap<String, bool *> &wrapper_request_extensions = wrapper->get_requested_extensions();
2021-12-14 02:44:12 +01:00
for (const KeyValue<String, bool *> &requested_extension : wrapper_request_extensions) {
2021-12-14 02:44:12 +01:00
requested_extensions[requested_extension.key] = requested_extension.value;
}
}
// Check which extensions are supported.
2021-12-14 02:44:12 +01:00
enabled_extensions.clear();
for (KeyValue<String, bool *> &requested_extension : requested_extensions) {
2021-12-14 02:44:12 +01:00
if (!is_extension_supported(requested_extension.key)) {
if (requested_extension.value == nullptr) {
// Null means this is a mandatory extension so we fail.
ERR_FAIL_V_MSG(false, String("OpenXR: OpenXR Runtime does not support ") + requested_extension.key + String(" extension!"));
2021-12-14 02:44:12 +01:00
} else {
// Set this extension as not supported.
2021-12-14 02:44:12 +01:00
*requested_extension.value = false;
}
} else if (requested_extension.value != nullptr) {
// Set this extension as supported.
2021-12-14 02:44:12 +01:00
*requested_extension.value = true;
// And record that we want to enable it.
enabled_extensions.push_back(requested_extension.key.ascii());
2021-12-14 02:44:12 +01:00
} else {
// Record that we want to enable this.
enabled_extensions.push_back(requested_extension.key.ascii());
2021-12-14 02:44:12 +01:00
}
}
Vector<const char *> extension_ptrs;
for (int i = 0; i < enabled_extensions.size(); i++) {
print_verbose(String("OpenXR: Enabling extension ") + String(enabled_extensions[i]));
extension_ptrs.push_back(enabled_extensions[i].get_data());
}
2021-12-14 02:44:12 +01:00
// Get our project name
String project_name = GLOBAL_GET("application/config/name");
// Create our OpenXR instance
XrApplicationInfo application_info{
"", // applicationName, we'll set this down below
1, // applicationVersion, we don't currently have this
"Godot Game Engine", // engineName
VERSION_MAJOR * 10000 + VERSION_MINOR * 100 + VERSION_PATCH, // engineVersion 4.0 -> 40000, 4.0.1 -> 40001, 4.1 -> 40100, etc.
XR_API_VERSION_1_0 // apiVersion
2021-12-14 02:44:12 +01:00
};
void *next_pointer = nullptr;
for (OpenXRExtensionWrapper *wrapper : registered_extension_wrappers) {
void *np = wrapper->set_instance_create_info_and_get_next_pointer(next_pointer);
if (np != nullptr) {
next_pointer = np;
}
}
2021-12-14 02:44:12 +01:00
XrInstanceCreateInfo instance_create_info = {
XR_TYPE_INSTANCE_CREATE_INFO, // type
next_pointer, // next
2021-12-14 02:44:12 +01:00
0, // createFlags
application_info, // applicationInfo
0, // enabledApiLayerCount, need to find out if we need support for this?
nullptr, // enabledApiLayerNames
uint32_t(extension_ptrs.size()), // enabledExtensionCount
extension_ptrs.ptr() // enabledExtensionNames
2021-12-14 02:44:12 +01:00
};
copy_string_to_char_buffer(project_name, instance_create_info.applicationInfo.applicationName, XR_MAX_APPLICATION_NAME_SIZE);
XrResult result = xrCreateInstance(&instance_create_info, &instance);
ERR_FAIL_COND_V_MSG(XR_FAILED(result), false, "Failed to create XR instance.");
// from this point on we can use get_error_string to get more info about our errors...
XrInstanceProperties instanceProps = {
XR_TYPE_INSTANCE_PROPERTIES, // type;
nullptr, // next
0, // runtimeVersion, from here will be set by our get call
"" // runtimeName
};
OPENXR_API_INIT_XR_FUNC_V(xrGetInstanceProperties);
2021-12-14 02:44:12 +01:00
result = xrGetInstanceProperties(instance, &instanceProps);
if (XR_FAILED(result)) {
// not fatal probably
print_line("OpenXR: Failed to get XR instance properties [", get_error_string(result), "]");
runtime_name = "";
runtime_version = "";
2021-12-14 02:44:12 +01:00
} else {
runtime_name = instanceProps.runtimeName;
runtime_version = OpenXRUtil::make_xr_version_string(instanceProps.runtimeVersion);
print_line("OpenXR: Running on OpenXR runtime: ", runtime_name, " ", runtime_version);
2021-12-14 02:44:12 +01:00
}
for (OpenXRExtensionWrapper *wrapper : registered_extension_wrappers) {
wrapper->on_instance_created(instance);
}
return true;
}
bool OpenXRAPI::get_system_info() {
// Retrieve basic OpenXR system info based on the form factor we desire
// Retrieve the system for our form factor, fails if form factor is not available
XrSystemGetInfo system_get_info = {
XR_TYPE_SYSTEM_GET_INFO, // type;
nullptr, // next
form_factor // formFactor
};
XrResult result = xrGetSystem(instance, &system_get_info, &system_id);
if (XR_FAILED(result)) {
print_line("OpenXR: Failed to get system for our form factor [", get_error_string(result), "]");
return false;
}
// obtain info about our system, writing this out completely to make CI on Linux happy..
void *next_pointer = nullptr;
for (OpenXRExtensionWrapper *wrapper : registered_extension_wrappers) {
void *np = wrapper->set_system_properties_and_get_next_pointer(next_pointer);
if (np != nullptr) {
next_pointer = np;
}
}
XrSystemProperties system_properties = {
XR_TYPE_SYSTEM_PROPERTIES, // type
next_pointer, // next
0, // systemId, from here will be set by our get call
0, // vendorId
"", // systemName
{
0, // maxSwapchainImageHeight
0, // maxSwapchainImageWidth
0, // maxLayerCount
}, // graphicsProperties
{
false, // orientationTracking
false // positionTracking
} // trackingProperties
};
result = xrGetSystemProperties(instance, system_id, &system_properties);
if (XR_FAILED(result)) {
print_line("OpenXR: Failed to get System properties [", get_error_string(result), "]");
return false;
}
// remember this state, we'll use it later
system_name = String(system_properties.systemName);
vendor_id = system_properties.vendorId;
graphics_properties = system_properties.graphicsProperties;
tracking_properties = system_properties.trackingProperties;
return true;
}
bool OpenXRAPI::load_supported_view_configuration_types() {
// This queries the supported configuration types, likely there will only be one choosing between Mono (phone AR) and Stereo (HMDs)
2021-12-14 02:44:12 +01:00
ERR_FAIL_COND_V(instance == XR_NULL_HANDLE, false);
if (supported_view_configuration_types != nullptr) {
// free previous results
memfree(supported_view_configuration_types);
supported_view_configuration_types = nullptr;
}
XrResult result = xrEnumerateViewConfigurations(instance, system_id, 0, &num_view_configuration_types, nullptr);
if (XR_FAILED(result)) {
print_line("OpenXR: Failed to get view configuration count [", get_error_string(result), "]");
return false;
}
supported_view_configuration_types = (XrViewConfigurationType *)memalloc(sizeof(XrViewConfigurationType) * num_view_configuration_types);
ERR_FAIL_NULL_V(supported_view_configuration_types, false);
result = xrEnumerateViewConfigurations(instance, system_id, num_view_configuration_types, &num_view_configuration_types, supported_view_configuration_types);
ERR_FAIL_COND_V_MSG(XR_FAILED(result), false, "OpenXR: Failed to enumerateview configurations");
ERR_FAIL_COND_V_MSG(num_view_configuration_types == 0, false, "OpenXR: Failed to enumerateview configurations"); // JIC there should be at least 1!
2021-12-14 02:44:12 +01:00
for (uint32_t i = 0; i < num_view_configuration_types; i++) {
print_verbose(String("OpenXR: Found supported view configuration ") + OpenXRUtil::get_view_configuration_name(supported_view_configuration_types[i]));
2021-12-14 02:44:12 +01:00
}
// Check value we loaded at startup...
if (!is_view_configuration_supported(view_configuration)) {
print_verbose(String("OpenXR: ") + OpenXRUtil::get_view_configuration_name(view_configuration) + String(" isn't supported, defaulting to ") + OpenXRUtil::get_view_configuration_name(supported_view_configuration_types[0]));
view_configuration = supported_view_configuration_types[0];
}
2021-12-14 02:44:12 +01:00
return true;
}
2023-02-02 15:14:16 +01:00
bool OpenXRAPI::load_supported_environmental_blend_modes() {
// This queries the supported environmental blend modes.
ERR_FAIL_COND_V(instance == XR_NULL_HANDLE, false);
if (supported_environment_blend_modes != nullptr) {
// free previous results
memfree(supported_environment_blend_modes);
supported_environment_blend_modes = nullptr;
num_supported_environment_blend_modes = 0;
}
XrResult result = xrEnumerateEnvironmentBlendModes(instance, system_id, view_configuration, 0, &num_supported_environment_blend_modes, nullptr);
if (XR_FAILED(result)) {
print_line("OpenXR: Failed to get supported environmental blend mode count [", get_error_string(result), "]");
return false;
}
supported_environment_blend_modes = (XrEnvironmentBlendMode *)memalloc(sizeof(XrEnvironmentBlendMode) * num_supported_environment_blend_modes);
ERR_FAIL_NULL_V(supported_environment_blend_modes, false);
result = xrEnumerateEnvironmentBlendModes(instance, system_id, view_configuration, num_supported_environment_blend_modes, &num_supported_environment_blend_modes, supported_environment_blend_modes);
ERR_FAIL_COND_V_MSG(XR_FAILED(result), false, "OpenXR: Failed to enumerate environmental blend modes");
ERR_FAIL_COND_V_MSG(num_supported_environment_blend_modes == 0, false, "OpenXR: Failed to enumerate environmental blend modes"); // JIC there should be at least 1!
2023-02-02 15:14:16 +01:00
for (uint32_t i = 0; i < num_supported_environment_blend_modes; i++) {
print_verbose(String("OpenXR: Found environmental blend mode ") + OpenXRUtil::get_environment_blend_mode_name(supported_environment_blend_modes[i]));
}
return true;
}
2021-12-14 02:44:12 +01:00
bool OpenXRAPI::is_view_configuration_supported(XrViewConfigurationType p_configuration_type) const {
ERR_FAIL_NULL_V(supported_view_configuration_types, false);
for (uint32_t i = 0; i < num_view_configuration_types; i++) {
if (supported_view_configuration_types[i] == p_configuration_type) {
return true;
}
}
return false;
}
bool OpenXRAPI::load_supported_view_configuration_views(XrViewConfigurationType p_configuration_type) {
// This loads our view configuration for each view so for a stereo HMD, we'll get two entries (that are likely identical)
// The returned data supplies us with the recommended render target size
if (!is_view_configuration_supported(p_configuration_type)) {
print_line("OpenXR: View configuration ", OpenXRUtil::get_view_configuration_name(view_configuration), " is not supported.");
return false;
}
if (view_configuration_views != nullptr) {
// free previous results
memfree(view_configuration_views);
view_configuration_views = nullptr;
}
XrResult result = xrEnumerateViewConfigurationViews(instance, system_id, p_configuration_type, 0, &view_count, nullptr);
if (XR_FAILED(result)) {
print_line("OpenXR: Failed to get view configuration count [", get_error_string(result), "]");
return false;
}
view_configuration_views = (XrViewConfigurationView *)memalloc(sizeof(XrViewConfigurationView) * view_count);
ERR_FAIL_NULL_V(view_configuration_views, false);
for (uint32_t i = 0; i < view_count; i++) {
view_configuration_views[i].type = XR_TYPE_VIEW_CONFIGURATION_VIEW;
view_configuration_views[i].next = nullptr;
2021-12-14 02:44:12 +01:00
}
result = xrEnumerateViewConfigurationViews(instance, system_id, p_configuration_type, view_count, &view_count, view_configuration_views);
ERR_FAIL_COND_V_MSG(XR_FAILED(result), false, "OpenXR: Failed to enumerate view configurations");
for (uint32_t i = 0; i < view_count; i++) {
print_verbose("OpenXR: Found supported view configuration view");
2022-10-18 04:55:39 +02:00
print_verbose(String(" - width: ") + itos(view_configuration_views[i].maxImageRectWidth));
print_verbose(String(" - height: ") + itos(view_configuration_views[i].maxImageRectHeight));
print_verbose(String(" - sample count: ") + itos(view_configuration_views[i].maxSwapchainSampleCount));
print_verbose(String(" - recommended render width: ") + itos(view_configuration_views[i].recommendedImageRectWidth));
print_verbose(String(" - recommended render height: ") + itos(view_configuration_views[i].recommendedImageRectHeight));
print_verbose(String(" - recommended render sample count: ") + itos(view_configuration_views[i].recommendedSwapchainSampleCount));
2021-12-14 02:44:12 +01:00
}
return true;
}
void OpenXRAPI::destroy_instance() {
if (view_configuration_views != nullptr) {
memfree(view_configuration_views);
view_configuration_views = nullptr;
}
if (supported_view_configuration_types != nullptr) {
memfree(supported_view_configuration_types);
supported_view_configuration_types = nullptr;
}
2023-02-02 15:14:16 +01:00
if (supported_environment_blend_modes != nullptr) {
memfree(supported_environment_blend_modes);
supported_environment_blend_modes = nullptr;
num_supported_environment_blend_modes = 0;
}
2021-12-14 02:44:12 +01:00
if (instance != XR_NULL_HANDLE) {
for (OpenXRExtensionWrapper *wrapper : registered_extension_wrappers) {
wrapper->on_instance_destroyed();
}
xrDestroyInstance(instance);
instance = XR_NULL_HANDLE;
}
enabled_extensions.clear();
if (graphics_extension != nullptr) {
unregister_extension_wrapper(graphics_extension);
memdelete(graphics_extension);
graphics_extension = nullptr;
}
2021-12-14 02:44:12 +01:00
}
bool OpenXRAPI::create_session() {
ERR_FAIL_COND_V(instance == XR_NULL_HANDLE, false);
ERR_FAIL_COND_V(session != XR_NULL_HANDLE, false);
void *next_pointer = nullptr;
for (OpenXRExtensionWrapper *wrapper : registered_extension_wrappers) {
void *np = wrapper->set_session_create_and_get_next_pointer(next_pointer);
if (np != nullptr) {
next_pointer = np;
}
}
XrSessionCreateInfo session_create_info = {
XR_TYPE_SESSION_CREATE_INFO, // type
next_pointer, // next
0, // createFlags
system_id // systemId
};
XrResult result = xrCreateSession(instance, &session_create_info, &session);
if (XR_FAILED(result)) {
print_line("OpenXR: Failed to create session [", get_error_string(result), "]");
return false;
}
set_object_name(XR_OBJECT_TYPE_SESSION, uint64_t(session), "Main Godot OpenXR Session");
begin_debug_label_region("Godot session active");
2021-12-14 02:44:12 +01:00
for (OpenXRExtensionWrapper *wrapper : registered_extension_wrappers) {
wrapper->on_session_created(session);
}
// Check our environment blend mode. This needs to happen after we call `on_session_created()`
// on the extension wrappers, so they can emulate alpha blend mode.
if (!set_environment_blend_mode(environment_blend_mode)) {
print_verbose(String("OpenXR: ") + OpenXRUtil::get_environment_blend_mode_name(environment_blend_mode) + String(" isn't supported, defaulting to ") + OpenXRUtil::get_environment_blend_mode_name(supported_environment_blend_modes[0]));
set_environment_blend_mode(supported_environment_blend_modes[0]);
}
2021-12-14 02:44:12 +01:00
return true;
}
bool OpenXRAPI::load_supported_reference_spaces() {
// loads the supported reference spaces for our OpenXR session
ERR_FAIL_COND_V(session == XR_NULL_HANDLE, false);
if (supported_reference_spaces != nullptr) {
// free previous results
memfree(supported_reference_spaces);
supported_reference_spaces = nullptr;
}
XrResult result = xrEnumerateReferenceSpaces(session, 0, &num_reference_spaces, nullptr);
if (XR_FAILED(result)) {
print_line("OpenXR: Failed to get reference space count [", get_error_string(result), "]");
return false;
}
supported_reference_spaces = (XrReferenceSpaceType *)memalloc(sizeof(XrReferenceSpaceType) * num_reference_spaces);
ERR_FAIL_NULL_V(supported_reference_spaces, false);
result = xrEnumerateReferenceSpaces(session, num_reference_spaces, &num_reference_spaces, supported_reference_spaces);
ERR_FAIL_COND_V_MSG(XR_FAILED(result), false, "OpenXR: Failed to enumerate reference spaces");
ERR_FAIL_COND_V_MSG(num_reference_spaces == 0, false, "OpenXR: Failed to enumerate reference spaces");
2021-12-14 02:44:12 +01:00
for (uint32_t i = 0; i < num_reference_spaces; i++) {
print_verbose(String("OpenXR: Found supported reference space ") + OpenXRUtil::get_reference_space_name(supported_reference_spaces[i]));
2021-12-14 02:44:12 +01:00
}
return true;
}
bool OpenXRAPI::is_reference_space_supported(XrReferenceSpaceType p_reference_space) {
ERR_FAIL_NULL_V(supported_reference_spaces, false);
for (uint32_t i = 0; i < num_reference_spaces; i++) {
if (supported_reference_spaces[i] == p_reference_space) {
return true;
}
}
return false;
}
bool OpenXRAPI::setup_play_space() {
ERR_FAIL_COND_V(session == XR_NULL_HANDLE, false);
2021-12-14 02:44:12 +01:00
XrPosef identityPose = {
{ 0.0, 0.0, 0.0, 1.0 },
{ 0.0, 0.0, 0.0 }
};
XrReferenceSpaceType new_reference_space;
XrSpace new_play_space = XR_NULL_HANDLE;
bool will_emulate_local_floor = false;
2021-12-14 02:44:12 +01:00
if (is_reference_space_supported(requested_reference_space)) {
new_reference_space = requested_reference_space;
} else if (requested_reference_space == XR_REFERENCE_SPACE_TYPE_LOCAL_FLOOR_EXT && is_reference_space_supported(XR_REFERENCE_SPACE_TYPE_STAGE)) {
print_verbose("OpenXR: LOCAL_FLOOR space isn't supported, emulating using STAGE and LOCAL spaces.");
new_reference_space = XR_REFERENCE_SPACE_TYPE_LOCAL;
will_emulate_local_floor = true;
if (local_floor_emulation.local_space == XR_NULL_HANDLE) {
XrReferenceSpaceCreateInfo create_info = {
XR_TYPE_REFERENCE_SPACE_CREATE_INFO, // type
nullptr, // next
XR_REFERENCE_SPACE_TYPE_LOCAL, // referenceSpaceType
identityPose, // poseInReferenceSpace
};
XrResult result = xrCreateReferenceSpace(session, &create_info, &local_floor_emulation.local_space);
if (XR_FAILED(result)) {
print_line("OpenXR: Failed to create LOCAL space in order to emulate LOCAL_FLOOR [", get_error_string(result), "]");
will_emulate_local_floor = false;
}
set_object_name(XR_OBJECT_TYPE_SPACE, uint64_t(local_floor_emulation.local_space), "Emulation local space");
}
if (local_floor_emulation.stage_space == XR_NULL_HANDLE) {
XrReferenceSpaceCreateInfo create_info = {
XR_TYPE_REFERENCE_SPACE_CREATE_INFO, // type
nullptr, // next
XR_REFERENCE_SPACE_TYPE_STAGE, // referenceSpaceType
identityPose, // poseInReferenceSpace
};
XrResult result = xrCreateReferenceSpace(session, &create_info, &local_floor_emulation.stage_space);
if (XR_FAILED(result)) {
print_line("OpenXR: Failed to create STAGE space in order to emulate LOCAL_FLOOR [", get_error_string(result), "]");
will_emulate_local_floor = false;
}
set_object_name(XR_OBJECT_TYPE_SPACE, uint64_t(local_floor_emulation.stage_space), "Emulation stage space");
}
if (!will_emulate_local_floor) {
if (local_floor_emulation.local_space != XR_NULL_HANDLE) {
xrDestroySpace(local_floor_emulation.local_space);
local_floor_emulation.local_space = XR_NULL_HANDLE;
}
if (local_floor_emulation.stage_space != XR_NULL_HANDLE) {
xrDestroySpace(local_floor_emulation.stage_space);
local_floor_emulation.stage_space = XR_NULL_HANDLE;
}
}
} else {
// Fallback on LOCAL, which all OpenXR runtimes are required to support.
print_verbose(String("OpenXR: ") + OpenXRUtil::get_reference_space_name(requested_reference_space) + String(" isn't supported, defaulting to LOCAL space."));
new_reference_space = XR_REFERENCE_SPACE_TYPE_LOCAL;
}
XrReferenceSpaceCreateInfo play_space_create_info = {
XR_TYPE_REFERENCE_SPACE_CREATE_INFO, // type
nullptr, // next
new_reference_space, // referenceSpaceType
identityPose, // poseInReferenceSpace
};
XrResult result = xrCreateReferenceSpace(session, &play_space_create_info, &new_play_space);
if (XR_FAILED(result)) {
print_line("OpenXR: Failed to create play space [", get_error_string(result), "]");
return false;
}
// If we've previously created a play space, clean it up first.
if (play_space != XR_NULL_HANDLE) {
// TODO Investigate if destroying our play space here is safe,
// it may still be used in the rendering thread.
xrDestroySpace(play_space);
}
play_space = new_play_space;
reference_space = new_reference_space;
2021-12-14 02:44:12 +01:00
set_object_name(XR_OBJECT_TYPE_SPACE, uint64_t(play_space), "Play space");
local_floor_emulation.enabled = will_emulate_local_floor;
local_floor_emulation.should_reset_floor_height = will_emulate_local_floor;
// Update render state so this play space is used rendering the upcoming frame.
set_render_play_space(play_space);
2021-12-14 02:44:12 +01:00
return true;
}
bool OpenXRAPI::setup_view_space() {
ERR_FAIL_COND_V(session == XR_NULL_HANDLE, false);
if (!is_reference_space_supported(XR_REFERENCE_SPACE_TYPE_VIEW)) {
print_line("OpenXR: reference space XR_REFERENCE_SPACE_TYPE_VIEW is not supported.");
return false;
2021-12-14 02:44:12 +01:00
}
XrPosef identityPose = {
{ 0.0, 0.0, 0.0, 1.0 },
{ 0.0, 0.0, 0.0 }
};
2021-12-14 02:44:12 +01:00
XrReferenceSpaceCreateInfo view_space_create_info = {
XR_TYPE_REFERENCE_SPACE_CREATE_INFO, // type
nullptr, // next
XR_REFERENCE_SPACE_TYPE_VIEW, // referenceSpaceType
identityPose // poseInReferenceSpace
};
2021-12-14 02:44:12 +01:00
XrResult result = xrCreateReferenceSpace(session, &view_space_create_info, &view_space);
if (XR_FAILED(result)) {
print_line("OpenXR: Failed to create view space [", get_error_string(result), "]");
return false;
2021-12-14 02:44:12 +01:00
}
set_object_name(XR_OBJECT_TYPE_SPACE, uint64_t(view_space), "View space");
2021-12-14 02:44:12 +01:00
return true;
}
bool OpenXRAPI::reset_emulated_floor_height() {
ERR_FAIL_COND_V(!local_floor_emulation.enabled, false);
ERR_FAIL_COND_V(local_floor_emulation.local_space == XR_NULL_HANDLE, false);
ERR_FAIL_COND_V(local_floor_emulation.stage_space == XR_NULL_HANDLE, false);
XrResult result;
XrSpaceLocation stage_location = {
XR_TYPE_SPACE_LOCATION, // type
nullptr, // next
0, // locationFlags
{ { 0.0, 0.0, 0.0, 1.0 }, { 0.0, 0.0, 0.0 } }, // pose
};
result = xrLocateSpace(local_floor_emulation.stage_space, local_floor_emulation.local_space, get_predicted_display_time(), &stage_location);
if (XR_FAILED(result)) {
print_line("OpenXR: Failed to locate STAGE space in LOCAL space, in order to emulate LOCAL_FLOOR [", get_error_string(result), "]");
return false;
}
XrPosef pose = {
{ 0.0, 0.0, 0.0, 1.0 },
{ 0.0, stage_location.pose.position.y, 0.0 }
};
XrReferenceSpaceCreateInfo create_info = {
XR_TYPE_REFERENCE_SPACE_CREATE_INFO, // type
nullptr, // next
XR_REFERENCE_SPACE_TYPE_LOCAL, // referenceSpaceType
pose, // poseInReferenceSpace
};
XrSpace new_play_space;
result = xrCreateReferenceSpace(session, &create_info, &new_play_space);
if (XR_FAILED(result)) {
print_line("OpenXR: Failed to recreate emulated LOCAL_FLOOR play space with latest floor estimate [", get_error_string(result), "]");
return false;
}
xrDestroySpace(play_space);
play_space = new_play_space;
// If we've made it this far, it means we can properly emulate LOCAL_FLOOR, so we'll
// report that as the reference space to the outside world.
reference_space = XR_REFERENCE_SPACE_TYPE_LOCAL_FLOOR_EXT;
// Update render state so this play space is used rendering the upcoming frame.
set_render_play_space(play_space);
return true;
}
2021-12-14 02:44:12 +01:00
bool OpenXRAPI::load_supported_swapchain_formats() {
ERR_FAIL_COND_V(session == XR_NULL_HANDLE, false);
if (supported_swapchain_formats != nullptr) {
// free previous results
memfree(supported_swapchain_formats);
supported_swapchain_formats = nullptr;
}
XrResult result = xrEnumerateSwapchainFormats(session, 0, &num_swapchain_formats, nullptr);
if (XR_FAILED(result)) {
print_line("OpenXR: Failed to get swapchain format count [", get_error_string(result), "]");
return false;
}
supported_swapchain_formats = (int64_t *)memalloc(sizeof(int64_t) * num_swapchain_formats);
ERR_FAIL_NULL_V(supported_swapchain_formats, false);
result = xrEnumerateSwapchainFormats(session, num_swapchain_formats, &num_swapchain_formats, supported_swapchain_formats);
ERR_FAIL_COND_V_MSG(XR_FAILED(result), false, "OpenXR: Failed to enumerate swapchain formats");
for (uint32_t i = 0; i < num_swapchain_formats; i++) {
print_verbose(String("OpenXR: Found supported swapchain format ") + get_swapchain_format_name(supported_swapchain_formats[i]));
2021-12-14 02:44:12 +01:00
}
return true;
}
bool OpenXRAPI::is_swapchain_format_supported(int64_t p_swapchain_format) {
ERR_FAIL_NULL_V(supported_swapchain_formats, false);
for (uint32_t i = 0; i < num_swapchain_formats; i++) {
if (supported_swapchain_formats[i] == p_swapchain_format) {
return true;
}
}
return false;
}
bool OpenXRAPI::obtain_swapchain_formats() {
2021-12-14 02:44:12 +01:00
ERR_FAIL_NULL_V(graphics_extension, false);
ERR_FAIL_COND_V(session == XR_NULL_HANDLE, false);
{
// Build a vector with swapchain formats we want to use, from best fit to worst
Vector<int64_t> usable_swapchain_formats;
color_swapchain_format = 0;
2021-12-14 02:44:12 +01:00
graphics_extension->get_usable_swapchain_formats(usable_swapchain_formats);
2021-12-14 02:44:12 +01:00
// now find out which one is supported
for (int i = 0; i < usable_swapchain_formats.size() && color_swapchain_format == 0; i++) {
if (is_swapchain_format_supported(usable_swapchain_formats[i])) {
color_swapchain_format = usable_swapchain_formats[i];
}
}
2021-12-14 02:44:12 +01:00
if (color_swapchain_format == 0) {
color_swapchain_format = usable_swapchain_formats[0]; // just use the first one and hope for the best...
print_line("Couldn't find usable color swap chain format, using", get_swapchain_format_name(color_swapchain_format), "instead.");
} else {
print_verbose(String("Using color swap chain format:") + get_swapchain_format_name(color_swapchain_format));
}
2021-12-14 02:44:12 +01:00
}
{
// Build a vector with swapchain formats we want to use, from best fit to worst
Vector<int64_t> usable_swapchain_formats;
depth_swapchain_format = 0;
graphics_extension->get_usable_depth_formats(usable_swapchain_formats);
// now find out which one is supported
for (int i = 0; i < usable_swapchain_formats.size() && depth_swapchain_format == 0; i++) {
if (is_swapchain_format_supported(usable_swapchain_formats[i])) {
depth_swapchain_format = usable_swapchain_formats[i];
}
}
if (depth_swapchain_format == 0) {
WARN_PRINT_ONCE("Couldn't find usable depth swap chain format, depth buffer will not be submitted if requested.");
} else {
print_verbose(String("Using depth swap chain format:") + get_swapchain_format_name(depth_swapchain_format));
}
}
return true;
}
bool OpenXRAPI::create_main_swapchains(Size2i p_size) {
ERR_NOT_ON_RENDER_THREAD_V(false);
ERR_FAIL_NULL_V(graphics_extension, false);
ERR_FAIL_COND_V(session == XR_NULL_HANDLE, false);
/*
TODO: We need to improve on this, for now we're taking our old approach of creating our main swapchains and substituting
those for the ones Godot normally creates.
This however means we can only use swapchains for our main XR view.
It would have been nicer if we could override the swapchain creation in Godot with ours but we have a timing issue here.
We can't create XR swapchains until after our XR session is fully instantiated, yet Godot creates its swapchain much earlier.
We only creates a swapchain for the main output here.
Additional swapchains may be created through our composition layer extension.
Finally an area we need to expand upon is that Foveated rendering is only enabled for the swap chain we create,
as we render 3D content into internal buffers that are copied into the swapchain, we do now have (basic) VRS support
*/
render_state.main_swapchain_size = p_size;
uint32_t sample_count = 1;
// We start with our color swapchain...
if (color_swapchain_format != 0) {
if (!render_state.main_swapchains[OPENXR_SWAPCHAIN_COLOR].create(0, XR_SWAPCHAIN_USAGE_SAMPLED_BIT | XR_SWAPCHAIN_USAGE_COLOR_ATTACHMENT_BIT | XR_SWAPCHAIN_USAGE_MUTABLE_FORMAT_BIT, color_swapchain_format, render_state.main_swapchain_size.width, render_state.main_swapchain_size.height, sample_count, view_count)) {
return false;
}
set_object_name(XR_OBJECT_TYPE_SWAPCHAIN, uint64_t(render_state.main_swapchains[OPENXR_SWAPCHAIN_COLOR].get_swapchain()), "Main color swapchain");
}
// We create our depth swapchain if:
// - we've enabled submitting depth buffer
// - we support our depth layer extension
// - we have our spacewarp extension (not yet implemented)
if (depth_swapchain_format != 0 && submit_depth_buffer && OpenXRCompositionLayerDepthExtension::get_singleton()->is_available()) {
if (!render_state.main_swapchains[OPENXR_SWAPCHAIN_DEPTH].create(0, XR_SWAPCHAIN_USAGE_SAMPLED_BIT | XR_SWAPCHAIN_USAGE_DEPTH_STENCIL_ATTACHMENT_BIT, depth_swapchain_format, render_state.main_swapchain_size.width, render_state.main_swapchain_size.height, sample_count, view_count)) {
return false;
}
set_object_name(XR_OBJECT_TYPE_SWAPCHAIN, uint64_t(render_state.main_swapchains[OPENXR_SWAPCHAIN_COLOR].get_swapchain()), "Main depth swapchain");
}
// We create our velocity swapchain if:
// - we have our spacewarp extension (not yet implemented)
{
// TBD
}
for (uint32_t i = 0; i < render_state.view_count; i++) {
render_state.views[i].type = XR_TYPE_VIEW;
render_state.views[i].next = nullptr;
render_state.projection_views[i].type = XR_TYPE_COMPOSITION_LAYER_PROJECTION_VIEW;
render_state.projection_views[i].next = nullptr;
render_state.projection_views[i].subImage.swapchain = render_state.main_swapchains[OPENXR_SWAPCHAIN_COLOR].get_swapchain();
render_state.projection_views[i].subImage.imageArrayIndex = i;
render_state.projection_views[i].subImage.imageRect.offset.x = 0;
render_state.projection_views[i].subImage.imageRect.offset.y = 0;
render_state.projection_views[i].subImage.imageRect.extent.width = render_state.main_swapchain_size.width;
render_state.projection_views[i].subImage.imageRect.extent.height = render_state.main_swapchain_size.height;
if (render_state.submit_depth_buffer && OpenXRCompositionLayerDepthExtension::get_singleton()->is_available() && render_state.depth_views) {
render_state.projection_views[i].next = &render_state.depth_views[i];
render_state.depth_views[i].type = XR_TYPE_COMPOSITION_LAYER_DEPTH_INFO_KHR;
render_state.depth_views[i].next = nullptr;
render_state.depth_views[i].subImage.swapchain = render_state.main_swapchains[OPENXR_SWAPCHAIN_DEPTH].get_swapchain();
render_state.depth_views[i].subImage.imageArrayIndex = i;
render_state.depth_views[i].subImage.imageRect.offset.x = 0;
render_state.depth_views[i].subImage.imageRect.offset.y = 0;
render_state.depth_views[i].subImage.imageRect.extent.width = render_state.main_swapchain_size.width;
render_state.depth_views[i].subImage.imageRect.extent.height = render_state.main_swapchain_size.height;
2024-04-09 08:15:42 +02:00
// OpenXR spec says that: minDepth < maxDepth.
render_state.depth_views[i].minDepth = 0.0;
render_state.depth_views[i].maxDepth = 1.0;
2024-04-09 08:15:42 +02:00
// But we can reverse near and far for reverse-Z.
render_state.depth_views[i].nearZ = 100.0; // Near and far Z will be set to the correct values in fill_projection_matrix
render_state.depth_views[i].farZ = 0.01;
}
2021-12-14 02:44:12 +01:00
};
for (OpenXRExtensionWrapper *wrapper : registered_extension_wrappers) {
wrapper->on_main_swapchains_created();
}
2021-12-14 02:44:12 +01:00
return true;
};
void OpenXRAPI::destroy_session() {
// TODO need to figure out if we're still rendering our current frame
// in a separate rendering thread and if so,
// if we need to wait for completion.
// We could be pulling the rug from underneath rendering...
if (running) {
if (session != XR_NULL_HANDLE) {
xrEndSession(session);
}
running = false;
render_state.running = false;
2021-12-14 02:44:12 +01:00
}
if (render_state.views != nullptr) {
memfree(render_state.views);
render_state.views = nullptr;
2021-12-14 02:44:12 +01:00
}
if (render_state.projection_views != nullptr) {
memfree(render_state.projection_views);
render_state.projection_views = nullptr;
2021-12-14 02:44:12 +01:00
}
if (render_state.depth_views != nullptr) {
memfree(render_state.depth_views);
render_state.depth_views = nullptr;
}
free_main_swapchains();
OpenXRSwapChainInfo::free_queued();
2021-12-14 02:44:12 +01:00
if (supported_swapchain_formats != nullptr) {
memfree(supported_swapchain_formats);
supported_swapchain_formats = nullptr;
}
// destroy our spaces
if (play_space != XR_NULL_HANDLE) {
xrDestroySpace(play_space);
play_space = XR_NULL_HANDLE;
render_state.play_space = XR_NULL_HANDLE;
2021-12-14 02:44:12 +01:00
}
if (view_space != XR_NULL_HANDLE) {
xrDestroySpace(view_space);
view_space = XR_NULL_HANDLE;
}
if (local_floor_emulation.local_space != XR_NULL_HANDLE) {
xrDestroySpace(local_floor_emulation.local_space);
local_floor_emulation.local_space = XR_NULL_HANDLE;
}
if (local_floor_emulation.stage_space != XR_NULL_HANDLE) {
xrDestroySpace(local_floor_emulation.stage_space);
local_floor_emulation.stage_space = XR_NULL_HANDLE;
}
local_floor_emulation.enabled = false;
local_floor_emulation.should_reset_floor_height = false;
2021-12-14 02:44:12 +01:00
if (supported_reference_spaces != nullptr) {
// free previous results
memfree(supported_reference_spaces);
supported_reference_spaces = nullptr;
}
if (session != XR_NULL_HANDLE) {
for (OpenXRExtensionWrapper *wrapper : registered_extension_wrappers) {
wrapper->on_session_destroyed();
}
end_debug_label_region();
2021-12-14 02:44:12 +01:00
xrDestroySession(session);
session = XR_NULL_HANDLE;
}
}
bool OpenXRAPI::on_state_idle() {
print_verbose("On state idle");
2021-12-14 02:44:12 +01:00
for (OpenXRExtensionWrapper *wrapper : registered_extension_wrappers) {
wrapper->on_state_idle();
}
return true;
}
bool OpenXRAPI::on_state_ready() {
print_verbose("On state ready");
2021-12-14 02:44:12 +01:00
// begin session
XrSessionBeginInfo session_begin_info = {
XR_TYPE_SESSION_BEGIN_INFO, // type
nullptr, // next
view_configuration // primaryViewConfigurationType
};
XrResult result = xrBeginSession(session, &session_begin_info);
if (XR_FAILED(result)) {
print_line("OpenXR: Failed to begin session [", get_error_string(result), "]");
return false;
}
// we're running
running = true;
set_render_session_running(true);
2021-12-14 02:44:12 +01:00
for (OpenXRExtensionWrapper *wrapper : registered_extension_wrappers) {
wrapper->on_state_ready();
}
if (xr_interface) {
xr_interface->on_state_ready();
}
2021-12-14 02:44:12 +01:00
return true;
}
bool OpenXRAPI::on_state_synchronized() {
print_verbose("On state synchronized");
2021-12-14 02:44:12 +01:00
// Just in case, see if we already have active trackers...
List<RID> trackers;
tracker_owner.get_owned_list(&trackers);
for (const RID &tracker : trackers) {
tracker_check_profile(tracker);
}
2021-12-14 02:44:12 +01:00
for (OpenXRExtensionWrapper *wrapper : registered_extension_wrappers) {
wrapper->on_state_synchronized();
}
return true;
}
bool OpenXRAPI::on_state_visible() {
print_verbose("On state visible");
2021-12-14 02:44:12 +01:00
for (OpenXRExtensionWrapper *wrapper : registered_extension_wrappers) {
wrapper->on_state_visible();
}
if (xr_interface) {
xr_interface->on_state_visible();
}
2021-12-14 02:44:12 +01:00
return true;
}
bool OpenXRAPI::on_state_focused() {
print_verbose("On state focused");
2021-12-14 02:44:12 +01:00
for (OpenXRExtensionWrapper *wrapper : registered_extension_wrappers) {
wrapper->on_state_focused();
}
if (xr_interface) {
xr_interface->on_state_focused();
}
2021-12-14 02:44:12 +01:00
return true;
}
bool OpenXRAPI::on_state_stopping() {
print_verbose("On state stopping");
2021-12-14 02:44:12 +01:00
if (xr_interface) {
xr_interface->on_state_stopping();
}
2021-12-14 02:44:12 +01:00
for (OpenXRExtensionWrapper *wrapper : registered_extension_wrappers) {
wrapper->on_state_stopping();
}
if (running) {
XrResult result = xrEndSession(session);
if (XR_FAILED(result)) {
// we only report this..
print_line("OpenXR: Failed to end session [", get_error_string(result), "]");
}
running = false;
set_render_session_running(false);
2021-12-14 02:44:12 +01:00
}
return true;
}
bool OpenXRAPI::on_state_loss_pending() {
print_verbose("On state loss pending");
2021-12-14 02:44:12 +01:00
if (xr_interface) {
xr_interface->on_state_loss_pending();
}
2021-12-14 02:44:12 +01:00
for (OpenXRExtensionWrapper *wrapper : registered_extension_wrappers) {
wrapper->on_state_loss_pending();
}
return true;
}
bool OpenXRAPI::on_state_exiting() {
print_verbose("On state existing");
2021-12-14 02:44:12 +01:00
if (xr_interface) {
xr_interface->on_state_exiting();
}
2021-12-14 02:44:12 +01:00
for (OpenXRExtensionWrapper *wrapper : registered_extension_wrappers) {
wrapper->on_state_exiting();
}
return true;
}
void OpenXRAPI::set_form_factor(XrFormFactor p_form_factor) {
ERR_FAIL_COND(is_initialized());
form_factor = p_form_factor;
}
uint32_t OpenXRAPI::get_view_count() {
return view_count;
}
void OpenXRAPI::set_view_configuration(XrViewConfigurationType p_view_configuration) {
ERR_FAIL_COND(is_initialized());
view_configuration = p_view_configuration;
}
bool OpenXRAPI::set_requested_reference_space(XrReferenceSpaceType p_requested_reference_space) {
requested_reference_space = p_requested_reference_space;
play_space_is_dirty = true;
return true;
}
void OpenXRAPI::set_submit_depth_buffer(bool p_submit_depth_buffer) {
ERR_FAIL_COND(is_initialized());
submit_depth_buffer = p_submit_depth_buffer;
}
2021-12-14 02:44:12 +01:00
bool OpenXRAPI::is_initialized() {
return (instance != XR_NULL_HANDLE);
}
bool OpenXRAPI::is_running() {
if (instance == XR_NULL_HANDLE) {
return false;
}
if (session == XR_NULL_HANDLE) {
return false;
}
return running;
}
bool OpenXRAPI::openxr_loader_init() {
#ifdef ANDROID_ENABLED
ERR_FAIL_COND_V_MSG(openxr_loader_library_handle != nullptr, false, "OpenXR Loader library is already loaded.");
{
Error error_code = OS::get_singleton()->open_dynamic_library(OPENXR_LOADER_NAME, openxr_loader_library_handle);
ERR_FAIL_COND_V_MSG(error_code != OK, false, "OpenXR loader not found.");
}
{
Error error_code = OS::get_singleton()->get_dynamic_library_symbol_handle(openxr_loader_library_handle, "xrGetInstanceProcAddr", (void *&)xrGetInstanceProcAddr);
ERR_FAIL_COND_V_MSG(error_code != OK, false, "Symbol xrGetInstanceProcAddr not found in OpenXR Loader library.");
}
#endif
// Resolve the symbols that don't require an instance
OPENXR_API_INIT_XR_FUNC_V(xrCreateInstance);
OPENXR_API_INIT_XR_FUNC_V(xrEnumerateApiLayerProperties);
OPENXR_API_INIT_XR_FUNC_V(xrEnumerateInstanceExtensionProperties);
return true;
}
bool OpenXRAPI::resolve_instance_openxr_symbols() {
ERR_FAIL_COND_V(instance == XR_NULL_HANDLE, false);
OPENXR_API_INIT_XR_FUNC_V(xrAcquireSwapchainImage);
OPENXR_API_INIT_XR_FUNC_V(xrApplyHapticFeedback);
OPENXR_API_INIT_XR_FUNC_V(xrAttachSessionActionSets);
OPENXR_API_INIT_XR_FUNC_V(xrBeginFrame);
OPENXR_API_INIT_XR_FUNC_V(xrBeginSession);
OPENXR_API_INIT_XR_FUNC_V(xrCreateAction);
OPENXR_API_INIT_XR_FUNC_V(xrCreateActionSet);
OPENXR_API_INIT_XR_FUNC_V(xrCreateActionSpace);
OPENXR_API_INIT_XR_FUNC_V(xrCreateReferenceSpace);
OPENXR_API_INIT_XR_FUNC_V(xrCreateSession);
OPENXR_API_INIT_XR_FUNC_V(xrCreateSwapchain);
OPENXR_API_INIT_XR_FUNC_V(xrDestroyAction);
OPENXR_API_INIT_XR_FUNC_V(xrDestroyActionSet);
OPENXR_API_INIT_XR_FUNC_V(xrDestroyInstance);
OPENXR_API_INIT_XR_FUNC_V(xrDestroySession);
OPENXR_API_INIT_XR_FUNC_V(xrDestroySpace);
OPENXR_API_INIT_XR_FUNC_V(xrDestroySwapchain);
OPENXR_API_INIT_XR_FUNC_V(xrEndFrame);
OPENXR_API_INIT_XR_FUNC_V(xrEndSession);
2023-02-02 15:14:16 +01:00
OPENXR_API_INIT_XR_FUNC_V(xrEnumerateEnvironmentBlendModes);
OPENXR_API_INIT_XR_FUNC_V(xrEnumerateReferenceSpaces);
OPENXR_API_INIT_XR_FUNC_V(xrEnumerateSwapchainFormats);
OPENXR_API_INIT_XR_FUNC_V(xrEnumerateViewConfigurations);
OPENXR_API_INIT_XR_FUNC_V(xrEnumerateViewConfigurationViews);
OPENXR_API_INIT_XR_FUNC_V(xrGetActionStateBoolean);
OPENXR_API_INIT_XR_FUNC_V(xrGetActionStateFloat);
OPENXR_API_INIT_XR_FUNC_V(xrGetActionStateVector2f);
OPENXR_API_INIT_XR_FUNC_V(xrGetCurrentInteractionProfile);
OPENXR_API_INIT_XR_FUNC_V(xrGetReferenceSpaceBoundsRect);
OPENXR_API_INIT_XR_FUNC_V(xrGetSystem);
OPENXR_API_INIT_XR_FUNC_V(xrGetSystemProperties);
OPENXR_API_INIT_XR_FUNC_V(xrLocateViews);
OPENXR_API_INIT_XR_FUNC_V(xrLocateSpace);
OPENXR_API_INIT_XR_FUNC_V(xrPathToString);
OPENXR_API_INIT_XR_FUNC_V(xrPollEvent);
OPENXR_API_INIT_XR_FUNC_V(xrReleaseSwapchainImage);
OPENXR_API_INIT_XR_FUNC_V(xrResultToString);
OPENXR_API_INIT_XR_FUNC_V(xrStringToPath);
OPENXR_API_INIT_XR_FUNC_V(xrSuggestInteractionProfileBindings);
OPENXR_API_INIT_XR_FUNC_V(xrSyncActions);
OPENXR_API_INIT_XR_FUNC_V(xrWaitFrame);
OPENXR_API_INIT_XR_FUNC_V(xrWaitSwapchainImage);
return true;
}
XrResult OpenXRAPI::try_get_instance_proc_addr(const char *p_name, PFN_xrVoidFunction *p_addr) {
return xrGetInstanceProcAddr(instance, p_name, p_addr);
}
XrResult OpenXRAPI::get_instance_proc_addr(const char *p_name, PFN_xrVoidFunction *p_addr) {
XrResult result = try_get_instance_proc_addr(p_name, p_addr);
if (result != XR_SUCCESS) {
String error_message = String("Symbol ") + p_name + " not found in OpenXR instance.";
2023-04-18 09:20:48 +02:00
ERR_FAIL_V_MSG(result, error_message.utf8().get_data());
}
return result;
}
bool OpenXRAPI::initialize(const String &p_rendering_driver) {
2021-12-14 02:44:12 +01:00
ERR_FAIL_COND_V_MSG(instance != XR_NULL_HANDLE, false, "OpenXR instance was already created");
if (!openxr_loader_init()) {
return false;
}
2021-12-14 02:44:12 +01:00
if (p_rendering_driver == "vulkan") {
#ifdef VULKAN_ENABLED
graphics_extension = memnew(OpenXRVulkanExtension);
2021-12-14 02:44:12 +01:00
register_extension_wrapper(graphics_extension);
#else
// shouldn't be possible...
ERR_FAIL_V(false);
#endif
} else if (p_rendering_driver == "opengl3") {
2023-07-18 16:02:47 +02:00
#if defined(GLES3_ENABLED) && !defined(MACOS_ENABLED)
graphics_extension = memnew(OpenXROpenGLExtension);
2022-10-23 03:29:15 +02:00
register_extension_wrapper(graphics_extension);
2021-12-14 02:44:12 +01:00
#else
// shouldn't be possible...
ERR_FAIL_V(false);
#endif
} else {
ERR_FAIL_V_MSG(false, "OpenXR: Unsupported rendering device.");
}
// Also register our rendering extensions
register_extension_wrapper(memnew(OpenXRFBUpdateSwapchainExtension(p_rendering_driver)));
register_extension_wrapper(memnew(OpenXRFBFoveationExtension(p_rendering_driver)));
// initialize
for (OpenXRExtensionWrapper *wrapper : registered_extension_wrappers) {
wrapper->on_before_instance_created();
}
2021-12-14 02:44:12 +01:00
if (!load_layer_properties()) {
destroy_instance();
return false;
}
if (!load_supported_extensions()) {
destroy_instance();
return false;
}
if (!create_instance()) {
destroy_instance();
return false;
}
if (!resolve_instance_openxr_symbols()) {
destroy_instance();
return false;
}
2021-12-14 02:44:12 +01:00
if (!get_system_info()) {
destroy_instance();
return false;
}
if (!load_supported_view_configuration_types()) {
destroy_instance();
return false;
}
if (!load_supported_view_configuration_views(view_configuration)) {
destroy_instance();
return false;
}
2023-02-02 15:14:16 +01:00
if (!load_supported_environmental_blend_modes()) {
destroy_instance();
return false;
}
2021-12-14 02:44:12 +01:00
return true;
}
bool OpenXRAPI::initialize_session() {
2021-12-14 02:44:12 +01:00
if (!create_session()) {
destroy_session();
return false;
}
if (!load_supported_reference_spaces()) {
destroy_session();
return false;
}
if (!setup_view_space()) {
2021-12-14 02:44:12 +01:00
destroy_session();
return false;
}
if (!load_supported_swapchain_formats()) {
destroy_session();
return false;
}
if (!obtain_swapchain_formats()) {
destroy_session();
return false;
}
allocate_view_buffers(view_count, submit_depth_buffer);
2021-12-14 02:44:12 +01:00
return true;
}
void OpenXRAPI::finish() {
destroy_session();
destroy_instance();
}
void OpenXRAPI::set_xr_interface(OpenXRInterface *p_xr_interface) {
xr_interface = p_xr_interface;
}
2021-12-14 02:44:12 +01:00
void OpenXRAPI::register_extension_wrapper(OpenXRExtensionWrapper *p_extension_wrapper) {
registered_extension_wrappers.push_back(p_extension_wrapper);
}
void OpenXRAPI::unregister_extension_wrapper(OpenXRExtensionWrapper *p_extension_wrapper) {
registered_extension_wrappers.erase(p_extension_wrapper);
}
const Vector<OpenXRExtensionWrapper *> &OpenXRAPI::get_registered_extension_wrappers() {
return registered_extension_wrappers;
}
void OpenXRAPI::register_extension_metadata() {
for (OpenXRExtensionWrapper *extension_wrapper : registered_extension_wrappers) {
extension_wrapper->on_register_metadata();
}
}
void OpenXRAPI::cleanup_extension_wrappers() {
for (OpenXRExtensionWrapper *extension_wrapper : registered_extension_wrappers) {
// Fix crash when the extension wrapper comes from GDExtension.
OpenXRExtensionWrapperExtension *gdextension_extension_wrapper = dynamic_cast<OpenXRExtensionWrapperExtension *>(extension_wrapper);
if (gdextension_extension_wrapper) {
memdelete(gdextension_extension_wrapper);
} else {
memdelete(extension_wrapper);
}
}
registered_extension_wrappers.clear();
}
2024-02-19 03:06:36 +01:00
XrHandTrackerEXT OpenXRAPI::get_hand_tracker(int p_hand_index) {
ERR_FAIL_INDEX_V(p_hand_index, OpenXRHandTrackingExtension::HandTrackedHands::OPENXR_MAX_TRACKED_HANDS, XR_NULL_HANDLE);
OpenXRHandTrackingExtension *hand_tracking = OpenXRHandTrackingExtension::get_singleton();
ERR_FAIL_NULL_V(hand_tracking, XR_NULL_HANDLE);
2024-02-19 03:06:36 +01:00
OpenXRHandTrackingExtension::HandTrackedHands hand = static_cast<OpenXRHandTrackingExtension::HandTrackedHands>(p_hand_index);
return hand_tracking->get_hand_tracker(hand)->hand_tracker;
2024-02-19 03:06:36 +01:00
}
2021-12-14 02:44:12 +01:00
Size2 OpenXRAPI::get_recommended_target_size() {
RenderingServer *rendering_server = RenderingServer::get_singleton();
2021-12-14 02:44:12 +01:00
ERR_FAIL_NULL_V(view_configuration_views, Size2());
Size2 target_size;
if (rendering_server && rendering_server->is_on_render_thread()) {
target_size.width = view_configuration_views[0].recommendedImageRectWidth * render_state.render_target_size_multiplier;
target_size.height = view_configuration_views[0].recommendedImageRectHeight * render_state.render_target_size_multiplier;
} else {
target_size.width = view_configuration_views[0].recommendedImageRectWidth * render_target_size_multiplier;
target_size.height = view_configuration_views[0].recommendedImageRectHeight * render_target_size_multiplier;
}
2021-12-14 02:44:12 +01:00
return target_size;
}
2022-06-14 16:33:03 +02:00
XRPose::TrackingConfidence OpenXRAPI::get_head_center(Transform3D &r_transform, Vector3 &r_linear_velocity, Vector3 &r_angular_velocity) {
2021-12-14 02:44:12 +01:00
XrResult result;
if (!running) {
return XRPose::XR_TRACKING_CONFIDENCE_NONE;
}
2021-12-14 02:44:12 +01:00
// Get display time
XrTime display_time = get_predicted_display_time();
if (display_time == 0) {
2021-12-14 02:44:12 +01:00
return XRPose::XR_TRACKING_CONFIDENCE_NONE;
}
XrSpaceVelocity velocity = {
XR_TYPE_SPACE_VELOCITY, // type
nullptr, // next
0, // velocityFlags
{ 0.0, 0.0, 0.0 }, // linearVelocity
{ 0.0, 0.0, 0.0 } // angularVelocity
};
XrSpaceLocation location = {
XR_TYPE_SPACE_LOCATION, // type
&velocity, // next
0, // locationFlags
{
{ 0.0, 0.0, 0.0, 0.0 }, // orientation
{ 0.0, 0.0, 0.0 } // position
} // pose
};
result = xrLocateSpace(view_space, play_space, display_time, &location);
if (XR_FAILED(result)) {
print_line("OpenXR: Failed to locate view space in play space [", get_error_string(result), "]");
return XRPose::XR_TRACKING_CONFIDENCE_NONE;
}
XRPose::TrackingConfidence confidence = transform_from_location(location, r_transform);
parse_velocities(velocity, r_linear_velocity, r_angular_velocity);
if (head_pose_confidence != confidence) {
// prevent error spam
head_pose_confidence = confidence;
if (head_pose_confidence == XRPose::XR_TRACKING_CONFIDENCE_NONE) {
print_line("OpenXR head space location not valid (check tracking?)");
} else if (head_pose_confidence == XRPose::XR_TRACKING_CONFIDENCE_LOW) {
print_verbose("OpenVR Head pose now tracking with low confidence");
2021-12-14 02:44:12 +01:00
} else {
print_verbose("OpenVR Head pose now tracking with high confidence");
2021-12-14 02:44:12 +01:00
}
}
return confidence;
}
bool OpenXRAPI::get_view_transform(uint32_t p_view, Transform3D &r_transform) {
ERR_NOT_ON_RENDER_THREAD_V(false);
2021-12-14 02:44:12 +01:00
if (!render_state.running) {
2021-12-14 02:44:12 +01:00
return false;
}
// we don't have valid view info
if (render_state.views == nullptr || !render_state.view_pose_valid) {
2021-12-14 02:44:12 +01:00
return false;
}
// Note, the timing of this is set right before rendering, which is what we need here.
r_transform = transform_from_pose(render_state.views[p_view].pose);
2021-12-14 02:44:12 +01:00
return true;
}
bool OpenXRAPI::get_view_projection(uint32_t p_view, double p_z_near, double p_z_far, Projection &p_camera_matrix) {
ERR_NOT_ON_RENDER_THREAD_V(false);
2021-12-14 02:44:12 +01:00
ERR_FAIL_NULL_V(graphics_extension, false);
if (!render_state.running) {
2021-12-14 02:44:12 +01:00
return false;
}
// we don't have valid view info
if (render_state.views == nullptr || !render_state.view_pose_valid) {
2021-12-14 02:44:12 +01:00
return false;
}
// if we're using depth views, make sure we update our near and far there...
if (render_state.depth_views != nullptr) {
for (uint32_t i = 0; i < render_state.view_count; i++) {
2024-04-09 08:15:42 +02:00
// As we are using reverse-Z these need to be flipped.
render_state.depth_views[i].nearZ = p_z_far;
render_state.depth_views[i].farZ = p_z_near;
}
}
// now update our projection
return graphics_extension->create_projection_fov(render_state.views[p_view].fov, p_z_near, p_z_far, p_camera_matrix);
2021-12-14 02:44:12 +01:00
}
2024-03-26 02:57:26 +01:00
Vector2 OpenXRAPI::get_eye_focus(uint32_t p_view, float p_aspect) {
ERR_FAIL_NULL_V(graphics_extension, Vector2());
if (!render_state.running) {
return Vector2();
}
// xrWaitFrame not run yet
if (render_state.predicted_display_time == 0) {
return Vector2();
}
// we don't have valid view info
if (render_state.views == nullptr || !render_state.view_pose_valid) {
return Vector2();
}
Projection cm;
if (!graphics_extension->create_projection_fov(render_state.views[p_view].fov, 0.1, 1000.0, cm)) {
return Vector2();
}
// Default focus to center...
Vector3 focus = cm.xform(Vector3(0.0, 0.0, 999.9));
// Lets check for eye tracking...
OpenXREyeGazeInteractionExtension *eye_gaze_interaction = OpenXREyeGazeInteractionExtension::get_singleton();
if (eye_gaze_interaction && eye_gaze_interaction->supports_eye_gaze_interaction()) {
Vector3 eye_gaze_pose;
if (eye_gaze_interaction->get_eye_gaze_pose(1.0, eye_gaze_pose)) {
Transform3D view_transform = transform_from_pose(render_state.views[p_view].pose);
eye_gaze_pose = view_transform.xform_inv(eye_gaze_pose);
focus = cm.xform(eye_gaze_pose);
}
}
return Vector2(focus.x, focus.y);
}
2021-12-14 02:44:12 +01:00
bool OpenXRAPI::poll_events() {
ERR_FAIL_COND_V(instance == XR_NULL_HANDLE, false);
XrEventDataBuffer runtimeEvent;
runtimeEvent.type = XR_TYPE_EVENT_DATA_BUFFER;
runtimeEvent.next = nullptr;
// runtimeEvent.varying = ...
XrResult pollResult = xrPollEvent(instance, &runtimeEvent);
while (pollResult == XR_SUCCESS) {
bool handled = false;
for (OpenXRExtensionWrapper *wrapper : registered_extension_wrappers) {
handled |= wrapper->on_event_polled(runtimeEvent);
}
switch (runtimeEvent.type) {
case XR_TYPE_EVENT_DATA_EVENTS_LOST: {
XrEventDataEventsLost *event = (XrEventDataEventsLost *)&runtimeEvent;
// We probably didn't poll fast enough, just output warning
WARN_PRINT("OpenXR EVENT: " + itos(event->lostEventCount) + " event data lost!");
} break;
case XR_TYPE_EVENT_DATA_INSTANCE_LOSS_PENDING: {
XrEventDataInstanceLossPending *event = (XrEventDataInstanceLossPending *)&runtimeEvent;
// TODO We get this event if we're about to loose our OpenXR instance.
// We should queue exiting Godot at this point.
print_verbose(String("OpenXR EVENT: instance loss pending at ") + itos(event->lossTime));
return false;
} break;
2021-12-14 02:44:12 +01:00
case XR_TYPE_EVENT_DATA_SESSION_STATE_CHANGED: {
XrEventDataSessionStateChanged *event = (XrEventDataSessionStateChanged *)&runtimeEvent;
session_state = event->state;
if (session_state >= XR_SESSION_STATE_MAX_ENUM) {
print_verbose(String("OpenXR EVENT: session state changed to UNKNOWN - ") + itos(session_state));
2021-12-14 02:44:12 +01:00
} else {
print_verbose(String("OpenXR EVENT: session state changed to ") + OpenXRUtil::get_session_state_name(session_state));
2021-12-14 02:44:12 +01:00
switch (session_state) {
case XR_SESSION_STATE_IDLE:
on_state_idle();
break;
case XR_SESSION_STATE_READY:
on_state_ready();
break;
case XR_SESSION_STATE_SYNCHRONIZED:
on_state_synchronized();
break;
case XR_SESSION_STATE_VISIBLE:
on_state_visible();
break;
case XR_SESSION_STATE_FOCUSED:
on_state_focused();
break;
case XR_SESSION_STATE_STOPPING:
on_state_stopping();
break;
case XR_SESSION_STATE_LOSS_PENDING:
on_state_loss_pending();
break;
case XR_SESSION_STATE_EXITING:
on_state_exiting();
break;
default:
break;
}
}
} break;
case XR_TYPE_EVENT_DATA_REFERENCE_SPACE_CHANGE_PENDING: {
XrEventDataReferenceSpaceChangePending *event = (XrEventDataReferenceSpaceChangePending *)&runtimeEvent;
print_verbose(String("OpenXR EVENT: reference space type ") + OpenXRUtil::get_reference_space_name(event->referenceSpaceType) + " change pending!");
if (local_floor_emulation.enabled) {
local_floor_emulation.should_reset_floor_height = true;
}
if (event->poseValid && xr_interface) {
xr_interface->on_pose_recentered();
}
} break;
case XR_TYPE_EVENT_DATA_INTERACTION_PROFILE_CHANGED: {
print_verbose("OpenXR EVENT: interaction profile changed!");
XrEventDataInteractionProfileChanged *event = (XrEventDataInteractionProfileChanged *)&runtimeEvent;
List<RID> trackers;
tracker_owner.get_owned_list(&trackers);
for (const RID &tracker : trackers) {
tracker_check_profile(tracker, event->session);
}
} break;
2021-12-14 02:44:12 +01:00
default:
if (!handled) {
print_verbose(String("OpenXR Unhandled event type ") + OpenXRUtil::get_structure_type_name(runtimeEvent.type));
2021-12-14 02:44:12 +01:00
}
break;
}
runtimeEvent.type = XR_TYPE_EVENT_DATA_BUFFER;
pollResult = xrPollEvent(instance, &runtimeEvent);
}
if (pollResult == XR_EVENT_UNAVAILABLE) {
// processed all events in the queue
return true;
} else {
ERR_FAIL_V_MSG(false, "OpenXR: Failed to poll events!");
}
}
void OpenXRAPI::_allocate_view_buffers(uint32_t p_view_count, bool p_submit_depth_buffer) {
// Must be called from rendering thread!
ERR_NOT_ON_RENDER_THREAD;
2021-12-14 02:44:12 +01:00
OpenXRAPI *openxr_api = OpenXRAPI::get_singleton();
ERR_FAIL_NULL(openxr_api);
2021-12-14 02:44:12 +01:00
openxr_api->render_state.view_count = p_view_count;
openxr_api->render_state.submit_depth_buffer = p_submit_depth_buffer;
2021-12-14 02:44:12 +01:00
// Allocate buffers we'll be populating with view information.
openxr_api->render_state.views = (XrView *)memalloc(sizeof(XrView) * p_view_count);
ERR_FAIL_NULL_MSG(openxr_api->render_state.views, "OpenXR Couldn't allocate memory for views");
memset(openxr_api->render_state.views, 0, sizeof(XrView) * p_view_count);
openxr_api->render_state.projection_views = (XrCompositionLayerProjectionView *)memalloc(sizeof(XrCompositionLayerProjectionView) * p_view_count);
ERR_FAIL_NULL_MSG(openxr_api->render_state.projection_views, "OpenXR Couldn't allocate memory for projection views");
memset(openxr_api->render_state.projection_views, 0, sizeof(XrCompositionLayerProjectionView) * p_view_count);
if (p_submit_depth_buffer && OpenXRCompositionLayerDepthExtension::get_singleton()->is_available()) {
openxr_api->render_state.depth_views = (XrCompositionLayerDepthInfoKHR *)memalloc(sizeof(XrCompositionLayerDepthInfoKHR) * p_view_count);
ERR_FAIL_NULL_MSG(openxr_api->render_state.depth_views, "OpenXR Couldn't allocate memory for depth views");
memset(openxr_api->render_state.depth_views, 0, sizeof(XrCompositionLayerDepthInfoKHR) * p_view_count);
2021-12-14 02:44:12 +01:00
}
}
2021-12-14 02:44:12 +01:00
void OpenXRAPI::_set_render_session_running(bool p_is_running) {
// Must be called from rendering thread!
ERR_NOT_ON_RENDER_THREAD;
OpenXRAPI *openxr_api = OpenXRAPI::get_singleton();
ERR_FAIL_NULL(openxr_api);
openxr_api->render_state.running = p_is_running;
2021-12-14 02:44:12 +01:00
}
void OpenXRAPI::_set_render_display_info(XrTime p_predicted_display_time, bool p_should_render) {
// Must be called from rendering thread!
ERR_NOT_ON_RENDER_THREAD;
OpenXRAPI *openxr_api = OpenXRAPI::get_singleton();
ERR_FAIL_NULL(openxr_api);
openxr_api->render_state.predicted_display_time = p_predicted_display_time;
openxr_api->render_state.should_render = p_should_render;
2021-12-14 02:44:12 +01:00
}
void OpenXRAPI::_set_render_play_space(uint64_t p_play_space) {
// Must be called from rendering thread!
ERR_NOT_ON_RENDER_THREAD;
2021-12-14 02:44:12 +01:00
OpenXRAPI *openxr_api = OpenXRAPI::get_singleton();
ERR_FAIL_NULL(openxr_api);
openxr_api->render_state.play_space = XrSpace(p_play_space);
}
2021-12-14 02:44:12 +01:00
void OpenXRAPI::_set_render_state_multiplier(double p_render_target_size_multiplier) {
// Must be called from rendering thread!
ERR_NOT_ON_RENDER_THREAD;
OpenXRAPI *openxr_api = OpenXRAPI::get_singleton();
ERR_FAIL_NULL(openxr_api);
openxr_api->render_state.render_target_size_multiplier = p_render_target_size_multiplier;
}
bool OpenXRAPI::process() {
ERR_FAIL_COND_V(instance == XR_NULL_HANDLE, false);
if (!poll_events()) {
return false;
}
if (!running) {
return false;
}
// We call xrWaitFrame as early as possible, this will allow OpenXR to get
// proper timing info between this point, and when we're ready to start rendering.
// As the name suggests, OpenXR can pause the thread to minimize the time between
// retrieving tracking data and using that tracking data to render.
// OpenXR thus works best if rendering is performed on a separate thread.
2021-12-14 02:44:12 +01:00
XrFrameWaitInfo frame_wait_info = { XR_TYPE_FRAME_WAIT_INFO, nullptr };
2022-09-23 05:20:12 +02:00
frame_state.predictedDisplayTime = 0;
frame_state.predictedDisplayPeriod = 0;
frame_state.shouldRender = false;
2021-12-14 02:44:12 +01:00
XrResult result = xrWaitFrame(session, &frame_wait_info, &frame_state);
if (XR_FAILED(result)) {
print_line("OpenXR: xrWaitFrame() was not successful [", get_error_string(result), "]");
// reset just in case
frame_state.predictedDisplayTime = 0;
frame_state.predictedDisplayPeriod = 0;
frame_state.shouldRender = false;
set_render_display_info(0, false);
return false;
2021-12-14 02:44:12 +01:00
}
if (frame_state.predictedDisplayPeriod > 500000000) {
// display period more then 0.5 seconds? must be wrong data
print_verbose(String("OpenXR resetting invalid display period ") + rtos(frame_state.predictedDisplayPeriod));
frame_state.predictedDisplayPeriod = 0;
}
set_render_display_info(frame_state.predictedDisplayTime, frame_state.shouldRender);
// This is before setup_play_space() to ensure that it happens on the frame after
// the play space has been created.
if (unlikely(local_floor_emulation.should_reset_floor_height && !play_space_is_dirty)) {
reset_emulated_floor_height();
local_floor_emulation.should_reset_floor_height = false;
}
if (unlikely(play_space_is_dirty)) {
setup_play_space();
play_space_is_dirty = false;
}
for (OpenXRExtensionWrapper *wrapper : registered_extension_wrappers) {
wrapper->on_process();
}
return true;
}
void OpenXRAPI::free_main_swapchains() {
for (int i = 0; i < OPENXR_SWAPCHAIN_MAX; i++) {
render_state.main_swapchains[i].queue_free();
}
}
void OpenXRAPI::pre_render() {
ERR_FAIL_COND(session == XR_NULL_HANDLE);
// Must be called from rendering thread!
ERR_NOT_ON_RENDER_THREAD;
if (!render_state.running) {
return;
}
// Process any swapchains that were queued to be freed
OpenXRSwapChainInfo::free_queued();
Size2i swapchain_size = get_recommended_target_size();
if (swapchain_size != render_state.main_swapchain_size) {
// Out with the old.
free_main_swapchains();
// In with the new.
create_main_swapchains(swapchain_size);
}
2021-12-14 02:44:12 +01:00
for (OpenXRExtensionWrapper *wrapper : registered_extension_wrappers) {
wrapper->on_pre_render();
}
// Get our view info for the frame we're about to render, note from the OpenXR manual:
// "Repeatedly calling xrLocateViews with the same time may not necessarily return the same result. Instead the prediction gets increasingly accurate as the function is called closer to the given time for which a prediction is made"
// We're calling this "relatively" early, the positioning we're obtaining here will be used to do our frustum culling,
// occlusion culling, etc. There is however a technique that we can investigate in the future where after our entire
// Vulkan command buffer is build, but right before vkSubmitQueue is called, we call xrLocateViews one more time and
// update the view and projection matrix once more with a slightly more accurate predication and then submit the
// command queues.
// That is not possible yet but worth investigating in the future.
XrViewLocateInfo view_locate_info = {
XR_TYPE_VIEW_LOCATE_INFO, // type
nullptr, // next
view_configuration, // viewConfigurationType
render_state.predicted_display_time, // displayTime
render_state.play_space // space
2021-12-14 02:44:12 +01:00
};
XrViewState view_state = {
XR_TYPE_VIEW_STATE, // type
nullptr, // next
0 // viewStateFlags
};
uint32_t view_count_output;
XrResult result = xrLocateViews(session, &view_locate_info, &view_state, render_state.view_count, &view_count_output, render_state.views);
2021-12-14 02:44:12 +01:00
if (XR_FAILED(result)) {
print_line("OpenXR: Couldn't locate views [", get_error_string(result), "]");
return;
}
bool pose_valid = true;
for (uint64_t i = 0; i < view_count_output; i++) {
if ((view_state.viewStateFlags & XR_VIEW_STATE_ORIENTATION_VALID_BIT) == 0 ||
(view_state.viewStateFlags & XR_VIEW_STATE_POSITION_VALID_BIT) == 0) {
pose_valid = false;
}
}
if (render_state.view_pose_valid != pose_valid) {
render_state.view_pose_valid = pose_valid;
if (!render_state.view_pose_valid) {
print_verbose("OpenXR View pose became invalid");
2021-12-14 02:44:12 +01:00
} else {
print_verbose("OpenXR View pose became valid");
2021-12-14 02:44:12 +01:00
}
}
// We should get our frame no from the rendering server, but this will do.
begin_debug_label_region(String("Session Frame ") + String::num_uint64(++render_state.frame));
2021-12-14 02:44:12 +01:00
// let's start our frame..
XrFrameBeginInfo frame_begin_info = {
XR_TYPE_FRAME_BEGIN_INFO, // type
nullptr // next
};
result = xrBeginFrame(session, &frame_begin_info);
if (XR_FAILED(result)) {
print_line("OpenXR: failed to being frame [", get_error_string(result), "]");
return;
}
// Reset this, we haven't found a viewport for output yet
render_state.has_xr_viewport = false;
2021-12-14 02:44:12 +01:00
}
bool OpenXRAPI::pre_draw_viewport(RID p_render_target) {
// Must be called from rendering thread!
ERR_NOT_ON_RENDER_THREAD_V(false);
// We found an XR viewport!
render_state.has_xr_viewport = true;
if (instance == XR_NULL_HANDLE || session == XR_NULL_HANDLE || !render_state.running || !render_state.view_pose_valid || !render_state.should_render) {
2021-12-14 02:44:12 +01:00
return false;
}
// Acquire our images
for (int i = 0; i < OPENXR_SWAPCHAIN_MAX; i++) {
if (!render_state.main_swapchains[i].is_image_acquired() && render_state.main_swapchains[i].get_swapchain() != XR_NULL_HANDLE) {
if (!render_state.main_swapchains[i].acquire(render_state.should_render)) {
return false;
}
}
}
for (OpenXRExtensionWrapper *wrapper : registered_extension_wrappers) {
wrapper->on_pre_draw_viewport(p_render_target);
}
2021-12-14 02:44:12 +01:00
return true;
}
XrSwapchain OpenXRAPI::get_color_swapchain() {
ERR_NOT_ON_RENDER_THREAD_V(XR_NULL_HANDLE);
return render_state.main_swapchains[OPENXR_SWAPCHAIN_COLOR].get_swapchain();
}
RID OpenXRAPI::get_color_texture() {
ERR_NOT_ON_RENDER_THREAD_V(RID());
return render_state.main_swapchains[OPENXR_SWAPCHAIN_COLOR].get_image();
}
RID OpenXRAPI::get_depth_texture() {
ERR_NOT_ON_RENDER_THREAD_V(RID());
// Note, image will not be acquired if we didn't have a suitable swap chain format.
if (render_state.submit_depth_buffer && render_state.main_swapchains[OPENXR_SWAPCHAIN_DEPTH].is_image_acquired()) {
return render_state.main_swapchains[OPENXR_SWAPCHAIN_DEPTH].get_image();
} else {
return RID();
}
}
2021-12-14 02:44:12 +01:00
void OpenXRAPI::post_draw_viewport(RID p_render_target) {
// Must be called from rendering thread!
ERR_NOT_ON_RENDER_THREAD;
if (instance == XR_NULL_HANDLE || session == XR_NULL_HANDLE || !render_state.running || !render_state.view_pose_valid || !render_state.should_render) {
2021-12-14 02:44:12 +01:00
return;
}
for (OpenXRExtensionWrapper *wrapper : registered_extension_wrappers) {
wrapper->on_post_draw_viewport(p_render_target);
}
2021-12-14 02:44:12 +01:00
};
void OpenXRAPI::end_frame() {
XrResult result;
ERR_FAIL_COND(session == XR_NULL_HANDLE);
2021-12-14 02:44:12 +01:00
// Must be called from rendering thread!
ERR_NOT_ON_RENDER_THREAD;
if (!render_state.running) {
2021-12-14 02:44:12 +01:00
return;
}
if (render_state.should_render && render_state.view_pose_valid) {
if (!render_state.has_xr_viewport) {
print_line("OpenXR: No viewport was marked with use_xr, there is no rendered output!");
} else if (!render_state.main_swapchains[OPENXR_SWAPCHAIN_COLOR].is_image_acquired()) {
print_line("OpenXR: No swapchain could be acquired to render to!");
}
2021-12-14 02:44:12 +01:00
}
// must have:
// - should_render set to true
2021-12-14 02:44:12 +01:00
// - a valid view pose for projection_views[eye].pose to submit layer
// - an image to render
if (!render_state.should_render || !render_state.view_pose_valid || !render_state.main_swapchains[OPENXR_SWAPCHAIN_COLOR].is_image_acquired()) {
2021-12-14 02:44:12 +01:00
// submit 0 layers when we shouldn't render
XrFrameEndInfo frame_end_info = {
XR_TYPE_FRAME_END_INFO, // type
nullptr, // next
render_state.predicted_display_time, // displayTime
2023-02-02 15:14:16 +01:00
environment_blend_mode, // environmentBlendMode
2021-12-14 02:44:12 +01:00
0, // layerCount
nullptr // layers
};
result = xrEndFrame(session, &frame_end_info);
if (XR_FAILED(result)) {
print_line("OpenXR: rendering skipped and failed to end frame! [", get_error_string(result), "]");
2021-12-14 02:44:12 +01:00
return;
}
end_debug_label_region(); // Session frame #
2021-12-14 02:44:12 +01:00
// neither eye is rendered
return;
}
// release our swapchain image if we acquired it
for (int i = 0; i < OPENXR_SWAPCHAIN_MAX; i++) {
if (render_state.main_swapchains[i].is_image_acquired()) {
render_state.main_swapchains[i].release();
}
2021-12-14 02:44:12 +01:00
}
for (uint32_t eye = 0; eye < render_state.view_count; eye++) {
render_state.projection_views[eye].fov = render_state.views[eye].fov;
render_state.projection_views[eye].pose = render_state.views[eye].pose;
2021-12-14 02:44:12 +01:00
}
Vector<OrderedCompositionLayer> ordered_layers_list;
bool projection_layer_is_first = true;
2021-12-14 02:44:12 +01:00
// Add composition layers from providers
for (OpenXRCompositionLayerProvider *provider : composition_layer_providers) {
for (int i = 0; i < provider->get_composition_layer_count(); i++) {
OrderedCompositionLayer layer = {
provider->get_composition_layer(i),
provider->get_composition_layer_order(i),
};
if (layer.composition_layer) {
ordered_layers_list.push_back(layer);
if (layer.sort_order == 0) {
WARN_PRINT_ONCE_ED("Composition layer returned sort order 0, it may be overwritten by projection layer.");
} else if (layer.sort_order < 0) {
projection_layer_is_first = false;
}
}
2021-12-14 02:44:12 +01:00
}
}
XrCompositionLayerFlags layer_flags = XR_COMPOSITION_LAYER_CORRECT_CHROMATIC_ABERRATION_BIT;
if (!projection_layer_is_first || environment_blend_mode != XR_ENVIRONMENT_BLEND_MODE_OPAQUE) {
layer_flags |= XR_COMPOSITION_LAYER_BLEND_TEXTURE_SOURCE_ALPHA_BIT;
}
2021-12-14 02:44:12 +01:00
XrCompositionLayerProjection projection_layer = {
XR_TYPE_COMPOSITION_LAYER_PROJECTION, // type
nullptr, // next
layer_flags, // layerFlags
render_state.play_space, // space
render_state.view_count, // viewCount
render_state.projection_views, // views
2021-12-14 02:44:12 +01:00
};
ordered_layers_list.push_back({ (const XrCompositionLayerBaseHeader *)&projection_layer, 0 });
// Sort our layers.
ordered_layers_list.sort_custom<OrderedCompositionLayer>();
// Now make a list we can pass on to OpenXR.
Vector<const XrCompositionLayerBaseHeader *> layers_list;
for (OrderedCompositionLayer &ordered_layer : ordered_layers_list) {
layers_list.push_back(ordered_layer.composition_layer);
}
2021-12-14 02:44:12 +01:00
XrFrameEndInfo frame_end_info = {
XR_TYPE_FRAME_END_INFO, // type
nullptr, // next
render_state.predicted_display_time, // displayTime
2023-02-02 15:14:16 +01:00
environment_blend_mode, // environmentBlendMode
2021-12-14 02:44:12 +01:00
static_cast<uint32_t>(layers_list.size()), // layerCount
layers_list.ptr() // layers
};
result = xrEndFrame(session, &frame_end_info);
if (XR_FAILED(result)) {
print_line("OpenXR: failed to end frame! [", get_error_string(result), "]");
return;
}
end_debug_label_region(); // Session frame #
2021-12-14 02:44:12 +01:00
}
float OpenXRAPI::get_display_refresh_rate() const {
OpenXRDisplayRefreshRateExtension *drrext = OpenXRDisplayRefreshRateExtension::get_singleton();
if (drrext) {
return drrext->get_refresh_rate();
}
return 0.0;
}
void OpenXRAPI::set_display_refresh_rate(float p_refresh_rate) {
OpenXRDisplayRefreshRateExtension *drrext = OpenXRDisplayRefreshRateExtension::get_singleton();
if (drrext != nullptr) {
drrext->set_refresh_rate(p_refresh_rate);
}
}
Array OpenXRAPI::get_available_display_refresh_rates() const {
OpenXRDisplayRefreshRateExtension *drrext = OpenXRDisplayRefreshRateExtension::get_singleton();
if (drrext != nullptr) {
return drrext->get_available_refresh_rates();
}
return Array();
}
double OpenXRAPI::get_render_target_size_multiplier() const {
return render_target_size_multiplier;
}
void OpenXRAPI::set_render_target_size_multiplier(double multiplier) {
render_target_size_multiplier = multiplier;
set_render_state_multiplier(multiplier);
}
bool OpenXRAPI::is_foveation_supported() const {
OpenXRFBFoveationExtension *fov_ext = OpenXRFBFoveationExtension::get_singleton();
return fov_ext != nullptr && fov_ext->is_enabled();
}
int OpenXRAPI::get_foveation_level() const {
OpenXRFBFoveationExtension *fov_ext = OpenXRFBFoveationExtension::get_singleton();
if (fov_ext != nullptr && fov_ext->is_enabled()) {
switch (fov_ext->get_foveation_level()) {
case XR_FOVEATION_LEVEL_NONE_FB:
return 0;
case XR_FOVEATION_LEVEL_LOW_FB:
return 1;
case XR_FOVEATION_LEVEL_MEDIUM_FB:
return 2;
case XR_FOVEATION_LEVEL_HIGH_FB:
return 3;
default:
return 0;
}
}
return 0;
}
void OpenXRAPI::set_foveation_level(int p_foveation_level) {
ERR_FAIL_UNSIGNED_INDEX(p_foveation_level, 4);
OpenXRFBFoveationExtension *fov_ext = OpenXRFBFoveationExtension::get_singleton();
if (fov_ext != nullptr && fov_ext->is_enabled()) {
XrFoveationLevelFB levels[] = { XR_FOVEATION_LEVEL_NONE_FB, XR_FOVEATION_LEVEL_LOW_FB, XR_FOVEATION_LEVEL_MEDIUM_FB, XR_FOVEATION_LEVEL_HIGH_FB };
fov_ext->set_foveation_level(levels[p_foveation_level]);
}
}
bool OpenXRAPI::get_foveation_dynamic() const {
OpenXRFBFoveationExtension *fov_ext = OpenXRFBFoveationExtension::get_singleton();
if (fov_ext != nullptr && fov_ext->is_enabled()) {
return fov_ext->get_foveation_dynamic() == XR_FOVEATION_DYNAMIC_LEVEL_ENABLED_FB;
}
return false;
}
void OpenXRAPI::set_foveation_dynamic(bool p_foveation_dynamic) {
OpenXRFBFoveationExtension *fov_ext = OpenXRFBFoveationExtension::get_singleton();
if (fov_ext != nullptr && fov_ext->is_enabled()) {
fov_ext->set_foveation_dynamic(p_foveation_dynamic ? XR_FOVEATION_DYNAMIC_LEVEL_ENABLED_FB : XR_FOVEATION_DYNAMIC_DISABLED_FB);
}
}
Size2 OpenXRAPI::get_play_space_bounds() const {
Size2 ret;
ERR_FAIL_COND_V(session == XR_NULL_HANDLE, Size2());
XrExtent2Df extents;
XrResult result = xrGetReferenceSpaceBoundsRect(session, reference_space, &extents);
if (XR_FAILED(result)) {
print_line("OpenXR: failed to get play space bounds! [", get_error_string(result), "]");
return ret;
}
ret.width = extents.width;
ret.height = extents.height;
return ret;
}
2021-12-14 02:44:12 +01:00
OpenXRAPI::OpenXRAPI() {
// OpenXRAPI is only constructed if OpenXR is enabled.
singleton = this;
2021-12-14 02:44:12 +01:00
if (Engine::get_singleton()->is_editor_hint()) {
// Enabled OpenXR in the editor? Adjust our settings for the editor
} else {
// Load settings from project settings
int form_factor_setting = GLOBAL_GET("xr/openxr/form_factor");
switch (form_factor_setting) {
2021-12-14 02:44:12 +01:00
case 0: {
form_factor = XR_FORM_FACTOR_HEAD_MOUNTED_DISPLAY;
} break;
case 1: {
form_factor = XR_FORM_FACTOR_HANDHELD_DISPLAY;
} break;
default:
break;
}
int view_configuration_setting = GLOBAL_GET("xr/openxr/view_configuration");
switch (view_configuration_setting) {
2021-12-14 02:44:12 +01:00
case 0: {
view_configuration = XR_VIEW_CONFIGURATION_TYPE_PRIMARY_MONO;
} break;
case 1: {
view_configuration = XR_VIEW_CONFIGURATION_TYPE_PRIMARY_STEREO;
} break;
/* we don't support quad and observer configurations (yet)
case 2: {
view_configuration = XR_VIEW_CONFIGURATION_TYPE_PRIMARY_QUAD_VARJO;
} break;
case 3: {
view_configuration = XR_VIEW_CONFIGURATION_TYPE_SECONDARY_MONO_FIRST_PERSON_OBSERVER_MSFT;
} break;
*/
default:
break;
}
int reference_space_setting = GLOBAL_GET("xr/openxr/reference_space");
switch (reference_space_setting) {
2021-12-14 02:44:12 +01:00
case 0: {
requested_reference_space = XR_REFERENCE_SPACE_TYPE_LOCAL;
2021-12-14 02:44:12 +01:00
} break;
case 1: {
requested_reference_space = XR_REFERENCE_SPACE_TYPE_STAGE;
} break;
case 2: {
requested_reference_space = XR_REFERENCE_SPACE_TYPE_LOCAL_FLOOR_EXT;
2021-12-14 02:44:12 +01:00
} break;
default:
break;
}
int environment_blend_mode_setting = GLOBAL_GET("xr/openxr/environment_blend_mode");
switch (environment_blend_mode_setting) {
case 0: {
environment_blend_mode = XR_ENVIRONMENT_BLEND_MODE_OPAQUE;
} break;
case 1: {
environment_blend_mode = XR_ENVIRONMENT_BLEND_MODE_ADDITIVE;
} break;
case 2: {
environment_blend_mode = XR_ENVIRONMENT_BLEND_MODE_ALPHA_BLEND;
} break;
default:
break;
}
submit_depth_buffer = GLOBAL_GET("xr/openxr/submit_depth_buffer");
2021-12-14 02:44:12 +01:00
}
}
OpenXRAPI::~OpenXRAPI() {
// cleanup our composition layer providers
for (OpenXRCompositionLayerProvider *provider : composition_layer_providers) {
memdelete(provider);
}
composition_layer_providers.clear();
if (supported_extensions != nullptr) {
memfree(supported_extensions);
supported_extensions = nullptr;
}
if (layer_properties != nullptr) {
memfree(layer_properties);
layer_properties = nullptr;
}
#ifdef ANDROID_ENABLED
if (openxr_loader_library_handle) {
OS::get_singleton()->close_dynamic_library(openxr_loader_library_handle);
openxr_loader_library_handle = nullptr;
}
#endif
singleton = nullptr;
2021-12-14 02:44:12 +01:00
}
Transform3D OpenXRAPI::transform_from_pose(const XrPosef &p_pose) {
Quaternion q(p_pose.orientation.x, p_pose.orientation.y, p_pose.orientation.z, p_pose.orientation.w);
Basis basis(q);
Vector3 origin(p_pose.position.x, p_pose.position.y, p_pose.position.z);
return Transform3D(basis, origin);
}
template <typename T>
XRPose::TrackingConfidence _transform_from_location(const T &p_location, Transform3D &r_transform) {
Basis basis;
Vector3 origin;
XRPose::TrackingConfidence confidence = XRPose::XR_TRACKING_CONFIDENCE_NONE;
const XrPosef &pose = p_location.pose;
2021-12-14 02:44:12 +01:00
// Check orientation
if (p_location.locationFlags & XR_SPACE_LOCATION_ORIENTATION_VALID_BIT) {
Quaternion q(pose.orientation.x, pose.orientation.y, pose.orientation.z, pose.orientation.w);
r_transform.basis = Basis(q);
if (p_location.locationFlags & XR_SPACE_LOCATION_ORIENTATION_TRACKED_BIT) {
// Fully valid orientation, so either 3DOF or 6DOF tracking with high confidence so default to HIGH_TRACKING
confidence = XRPose::XR_TRACKING_CONFIDENCE_HIGH;
} else {
// Orientation is being tracked but we're using old/predicted data, so low tracking confidence
confidence = XRPose::XR_TRACKING_CONFIDENCE_LOW;
}
} else {
r_transform.basis = Basis();
}
// Check location
if (p_location.locationFlags & XR_SPACE_LOCATION_POSITION_VALID_BIT) {
r_transform.origin = Vector3(pose.position.x, pose.position.y, pose.position.z);
if (!(p_location.locationFlags & XR_SPACE_LOCATION_ORIENTATION_TRACKED_BIT)) {
// Location is being tracked but we're using old/predicted data, so low tracking confidence
confidence = XRPose::XR_TRACKING_CONFIDENCE_LOW;
} else if (confidence == XRPose::XR_TRACKING_CONFIDENCE_NONE) {
// Position tracking without orientation tracking?
confidence = XRPose::XR_TRACKING_CONFIDENCE_HIGH;
}
} else {
// No tracking or 3DOF I guess..
r_transform.origin = Vector3();
}
return confidence;
}
XRPose::TrackingConfidence OpenXRAPI::transform_from_location(const XrSpaceLocation &p_location, Transform3D &r_transform) {
return _transform_from_location(p_location, r_transform);
}
XRPose::TrackingConfidence OpenXRAPI::transform_from_location(const XrHandJointLocationEXT &p_location, Transform3D &r_transform) {
return _transform_from_location(p_location, r_transform);
}
2022-06-14 16:33:03 +02:00
void OpenXRAPI::parse_velocities(const XrSpaceVelocity &p_velocity, Vector3 &r_linear_velocity, Vector3 &r_angular_velocity) {
2021-12-14 02:44:12 +01:00
if (p_velocity.velocityFlags & XR_SPACE_VELOCITY_LINEAR_VALID_BIT) {
XrVector3f linear_velocity = p_velocity.linearVelocity;
r_linear_velocity = Vector3(linear_velocity.x, linear_velocity.y, linear_velocity.z);
} else {
r_linear_velocity = Vector3();
}
if (p_velocity.velocityFlags & XR_SPACE_VELOCITY_ANGULAR_VALID_BIT) {
XrVector3f angular_velocity = p_velocity.angularVelocity;
r_angular_velocity = Vector3(angular_velocity.x, angular_velocity.y, angular_velocity.z);
} else {
r_angular_velocity = Vector3();
}
}
2022-09-26 09:04:12 +02:00
bool OpenXRAPI::xr_result(XrResult result, const char *format, Array args) const {
if (XR_SUCCEEDED(result))
return true;
char resultString[XR_MAX_RESULT_STRING_SIZE];
xrResultToString(instance, result, resultString);
print_error(String("OpenXR ") + String(format).format(args) + String(" [") + String(resultString) + String("]"));
return false;
}
RID OpenXRAPI::get_tracker_rid(XrPath p_path) {
List<RID> current;
tracker_owner.get_owned_list(&current);
2024-03-26 02:57:26 +01:00
for (const RID &E : current) {
Tracker *tracker = tracker_owner.get_or_null(E);
if (tracker && tracker->toplevel_path == p_path) {
2024-03-26 02:57:26 +01:00
return E;
}
}
return RID();
}
RID OpenXRAPI::find_tracker(const String &p_name) {
List<RID> current;
tracker_owner.get_owned_list(&current);
for (const RID &E : current) {
Tracker *tracker = tracker_owner.get_or_null(E);
if (tracker && tracker->name == p_name) {
return E;
}
}
2021-12-14 02:44:12 +01:00
return RID();
}
2021-12-14 02:44:12 +01:00
RID OpenXRAPI::tracker_create(const String p_name) {
ERR_FAIL_COND_V(instance == XR_NULL_HANDLE, RID());
2021-12-14 02:44:12 +01:00
Tracker new_tracker;
new_tracker.name = p_name;
new_tracker.toplevel_path = XR_NULL_PATH;
new_tracker.active_profile_rid = RID();
2021-12-14 02:44:12 +01:00
XrResult result = xrStringToPath(instance, p_name.utf8().get_data(), &new_tracker.toplevel_path);
2021-12-14 02:44:12 +01:00
if (XR_FAILED(result)) {
print_line("OpenXR: failed to get path for ", p_name, "! [", get_error_string(result), "]");
return RID();
}
return tracker_owner.make_rid(new_tracker);
2021-12-14 02:44:12 +01:00
}
String OpenXRAPI::tracker_get_name(RID p_tracker) {
if (p_tracker.is_null()) {
return String("None");
}
Tracker *tracker = tracker_owner.get_or_null(p_tracker);
ERR_FAIL_NULL_V(tracker, String());
return tracker->name;
}
void OpenXRAPI::tracker_check_profile(RID p_tracker, XrSession p_session) {
if (p_session == XR_NULL_HANDLE) {
p_session = session;
}
Tracker *tracker = tracker_owner.get_or_null(p_tracker);
ERR_FAIL_NULL(tracker);
if (tracker->toplevel_path == XR_NULL_PATH) {
// no path, how was this even created?
return;
}
XrInteractionProfileState profile_state = {
XR_TYPE_INTERACTION_PROFILE_STATE, // type
nullptr, // next
XR_NULL_PATH // interactionProfile
};
XrResult result = xrGetCurrentInteractionProfile(p_session, tracker->toplevel_path, &profile_state);
if (XR_FAILED(result)) {
print_line("OpenXR: Failed to get interaction profile for", itos(tracker->toplevel_path), "[", get_error_string(result), "]");
return;
}
XrPath new_profile = profile_state.interactionProfile;
XrPath was_profile = get_interaction_profile_path(tracker->active_profile_rid);
if (was_profile != new_profile) {
tracker->active_profile_rid = get_interaction_profile_rid(new_profile);
if (xr_interface) {
xr_interface->tracker_profile_changed(p_tracker, tracker->active_profile_rid);
}
}
}
void OpenXRAPI::tracker_free(RID p_tracker) {
Tracker *tracker = tracker_owner.get_or_null(p_tracker);
ERR_FAIL_NULL(tracker);
2021-12-14 02:44:12 +01:00
// there is nothing to free here
tracker_owner.free(p_tracker);
2021-12-14 02:44:12 +01:00
}
RID OpenXRAPI::action_set_create(const String p_name, const String p_localized_name, const int p_priority) {
ERR_FAIL_COND_V(instance == XR_NULL_HANDLE, RID());
ActionSet action_set;
action_set.name = p_name;
2021-12-14 02:44:12 +01:00
action_set.is_attached = false;
// create our action set...
XrActionSetCreateInfo action_set_info = {
XR_TYPE_ACTION_SET_CREATE_INFO, // type
nullptr, // next
"", // actionSetName
"", // localizedActionSetName
uint32_t(p_priority) // priority
};
copy_string_to_char_buffer(p_name, action_set_info.actionSetName, XR_MAX_ACTION_SET_NAME_SIZE);
copy_string_to_char_buffer(p_localized_name, action_set_info.localizedActionSetName, XR_MAX_LOCALIZED_ACTION_SET_NAME_SIZE);
XrResult result = xrCreateActionSet(instance, &action_set_info, &action_set.handle);
if (XR_FAILED(result)) {
print_line("OpenXR: failed to create action set ", p_name, "! [", get_error_string(result), "]");
return RID();
}
set_object_name(XR_OBJECT_TYPE_ACTION_SET, uint64_t(action_set.handle), p_name);
2021-12-14 02:44:12 +01:00
return action_set_owner.make_rid(action_set);
}
String OpenXRAPI::action_set_get_name(RID p_action_set) {
if (p_action_set.is_null()) {
return String("None");
}
ActionSet *action_set = action_set_owner.get_or_null(p_action_set);
ERR_FAIL_NULL_V(action_set, String());
return action_set->name;
}
bool OpenXRAPI::attach_action_sets(const Vector<RID> &p_action_sets) {
ERR_FAIL_COND_V(session == XR_NULL_HANDLE, false);
2021-12-14 02:44:12 +01:00
Vector<XrActionSet> action_handles;
action_handles.resize(p_action_sets.size());
for (int i = 0; i < p_action_sets.size(); i++) {
ActionSet *action_set = action_set_owner.get_or_null(p_action_sets[i]);
ERR_FAIL_NULL_V(action_set, false);
2021-12-14 02:44:12 +01:00
if (action_set->is_attached) {
return false;
}
action_handles.set(i, action_set->handle);
}
2021-12-14 02:44:12 +01:00
// So according to the docs, once we attach our action set to our session it becomes read only..
// https://www.khronos.org/registry/OpenXR/specs/1.0/man/html/xrAttachSessionActionSets.html
XrSessionActionSetsAttachInfo attach_info = {
XR_TYPE_SESSION_ACTION_SETS_ATTACH_INFO, // type
nullptr, // next
(uint32_t)p_action_sets.size(), // countActionSets,
action_handles.ptr() // actionSets
2021-12-14 02:44:12 +01:00
};
XrResult result = xrAttachSessionActionSets(session, &attach_info);
if (XR_FAILED(result)) {
print_line("OpenXR: failed to attach action sets! [", get_error_string(result), "]");
2021-12-14 02:44:12 +01:00
return false;
}
for (int i = 0; i < p_action_sets.size(); i++) {
ActionSet *action_set = action_set_owner.get_or_null(p_action_sets[i]);
ERR_FAIL_NULL_V(action_set, false);
action_set->is_attached = true;
}
2021-12-14 02:44:12 +01:00
/* For debugging:
print_verbose("Attached set " + action_set->name);
List<RID> action_rids;
action_owner.get_owned_list(&action_rids);
for (int i = 0; i < action_rids.size(); i++) {
Action * action = action_owner.get_or_null(action_rids[i]);
if (action && action->action_set_rid == p_action_set) {
print_verbose(" - Action " + action->name + ": " + OpenXRUtil::get_action_type_name(action->action_type));
for (int j = 0; j < action->trackers.size(); j++) {
Tracker * tracker = tracker_owner.get_or_null(action->trackers[j].tracker_rid);
if (tracker) {
print_verbose(" - " + tracker->name);
}
}
}
}
*/
2021-12-14 02:44:12 +01:00
return true;
}
void OpenXRAPI::action_set_free(RID p_action_set) {
ActionSet *action_set = action_set_owner.get_or_null(p_action_set);
ERR_FAIL_NULL(action_set);
if (action_set->handle != XR_NULL_HANDLE) {
xrDestroyActionSet(action_set->handle);
}
action_set_owner.free(p_action_set);
}
RID OpenXRAPI::get_action_rid(XrAction p_action) {
List<RID> current;
action_owner.get_owned_list(&current);
2024-03-26 02:57:26 +01:00
for (const RID &E : current) {
Action *action = action_owner.get_or_null(E);
if (action && action->handle == p_action) {
2024-03-26 02:57:26 +01:00
return E;
}
}
return RID();
}
RID OpenXRAPI::find_action(const String &p_name) {
List<RID> current;
action_owner.get_owned_list(&current);
for (const RID &E : current) {
Action *action = action_owner.get_or_null(E);
if (action && action->name == p_name) {
return E;
}
}
return RID();
}
RID OpenXRAPI::action_create(RID p_action_set, const String p_name, const String p_localized_name, OpenXRAction::ActionType p_action_type, const Vector<RID> &p_trackers) {
2021-12-14 02:44:12 +01:00
ERR_FAIL_COND_V(instance == XR_NULL_HANDLE, RID());
Action action;
action.name = p_name;
2021-12-14 02:44:12 +01:00
ActionSet *action_set = action_set_owner.get_or_null(p_action_set);
ERR_FAIL_NULL_V(action_set, RID());
ERR_FAIL_COND_V(action_set->handle == XR_NULL_HANDLE, RID());
action.action_set_rid = p_action_set;
2021-12-14 02:44:12 +01:00
switch (p_action_type) {
case OpenXRAction::OPENXR_ACTION_BOOL:
action.action_type = XR_ACTION_TYPE_BOOLEAN_INPUT;
break;
case OpenXRAction::OPENXR_ACTION_FLOAT:
action.action_type = XR_ACTION_TYPE_FLOAT_INPUT;
break;
case OpenXRAction::OPENXR_ACTION_VECTOR2:
action.action_type = XR_ACTION_TYPE_VECTOR2F_INPUT;
break;
case OpenXRAction::OPENXR_ACTION_POSE:
action.action_type = XR_ACTION_TYPE_POSE_INPUT;
break;
case OpenXRAction::OPENXR_ACTION_HAPTIC:
action.action_type = XR_ACTION_TYPE_VIBRATION_OUTPUT;
break;
default:
ERR_FAIL_V(RID());
break;
}
Vector<XrPath> toplevel_paths;
for (int i = 0; i < p_trackers.size(); i++) {
Tracker *tracker = tracker_owner.get_or_null(p_trackers[i]);
if (tracker != nullptr && tracker->toplevel_path != XR_NULL_PATH) {
ActionTracker action_tracker = {
p_trackers[i], // tracker
2021-12-14 02:44:12 +01:00
XR_NULL_HANDLE, // space
false // was_location_valid
};
action.trackers.push_back(action_tracker);
2021-12-14 02:44:12 +01:00
toplevel_paths.push_back(tracker->toplevel_path);
2021-12-14 02:44:12 +01:00
}
}
XrActionCreateInfo action_info = {
XR_TYPE_ACTION_CREATE_INFO, // type
nullptr, // next
"", // actionName
action.action_type, // actionType
uint32_t(toplevel_paths.size()), // countSubactionPaths
toplevel_paths.ptr(), // subactionPaths
"" // localizedActionName
};
copy_string_to_char_buffer(p_name, action_info.actionName, XR_MAX_ACTION_NAME_SIZE);
copy_string_to_char_buffer(p_localized_name, action_info.localizedActionName, XR_MAX_LOCALIZED_ACTION_NAME_SIZE);
XrResult result = xrCreateAction(action_set->handle, &action_info, &action.handle);
if (XR_FAILED(result)) {
print_line("OpenXR: failed to create action ", p_name, "! [", get_error_string(result), "]");
return RID();
}
set_object_name(XR_OBJECT_TYPE_ACTION, uint64_t(action.handle), p_name);
2021-12-14 02:44:12 +01:00
return action_owner.make_rid(action);
}
String OpenXRAPI::action_get_name(RID p_action) {
if (p_action.is_null()) {
return String("None");
}
Action *action = action_owner.get_or_null(p_action);
ERR_FAIL_NULL_V(action, String());
return action->name;
}
2021-12-14 02:44:12 +01:00
void OpenXRAPI::action_free(RID p_action) {
Action *action = action_owner.get_or_null(p_action);
ERR_FAIL_NULL(action);
if (action->handle != XR_NULL_HANDLE) {
xrDestroyAction(action->handle);
}
action_owner.free(p_action);
}
RID OpenXRAPI::get_interaction_profile_rid(XrPath p_path) {
List<RID> current;
interaction_profile_owner.get_owned_list(&current);
2024-03-26 02:57:26 +01:00
for (const RID &E : current) {
InteractionProfile *ip = interaction_profile_owner.get_or_null(E);
if (ip && ip->path == p_path) {
2024-03-26 02:57:26 +01:00
return E;
}
}
return RID();
}
XrPath OpenXRAPI::get_interaction_profile_path(RID p_interaction_profile) {
if (p_interaction_profile.is_null()) {
return XR_NULL_PATH;
}
InteractionProfile *ip = interaction_profile_owner.get_or_null(p_interaction_profile);
ERR_FAIL_NULL_V(ip, XR_NULL_PATH);
2021-12-14 02:44:12 +01:00
return ip->path;
}
2021-12-14 02:44:12 +01:00
RID OpenXRAPI::interaction_profile_create(const String p_name) {
if (!is_interaction_profile_supported(p_name)) {
2022-09-23 05:20:12 +02:00
// The extension enabling this path must not be active, we will silently skip this interaction profile
return RID();
}
InteractionProfile new_interaction_profile;
XrResult result = xrStringToPath(instance, p_name.utf8().get_data(), &new_interaction_profile.path);
2021-12-14 02:44:12 +01:00
if (XR_FAILED(result)) {
print_line("OpenXR: failed to get path for ", p_name, "! [", get_error_string(result), "]");
return RID();
2021-12-14 02:44:12 +01:00
}
RID existing_ip = get_interaction_profile_rid(new_interaction_profile.path);
if (existing_ip.is_valid()) {
return existing_ip;
}
2021-12-14 02:44:12 +01:00
new_interaction_profile.name = p_name;
return interaction_profile_owner.make_rid(new_interaction_profile);
}
2021-12-14 02:44:12 +01:00
String OpenXRAPI::interaction_profile_get_name(RID p_interaction_profile) {
if (p_interaction_profile.is_null()) {
return String("None");
}
2021-12-14 02:44:12 +01:00
InteractionProfile *ip = interaction_profile_owner.get_or_null(p_interaction_profile);
ERR_FAIL_NULL_V(ip, String());
return ip->name;
}
void OpenXRAPI::interaction_profile_clear_bindings(RID p_interaction_profile) {
InteractionProfile *ip = interaction_profile_owner.get_or_null(p_interaction_profile);
ERR_FAIL_NULL(ip);
ip->bindings.clear();
}
bool OpenXRAPI::interaction_profile_add_binding(RID p_interaction_profile, RID p_action, const String p_path) {
InteractionProfile *ip = interaction_profile_owner.get_or_null(p_interaction_profile);
ERR_FAIL_NULL_V(ip, false);
if (!interaction_profile_supports_io_path(ip->name, p_path)) {
return false;
}
XrActionSuggestedBinding binding;
Action *action = action_owner.get_or_null(p_action);
ERR_FAIL_COND_V(action == nullptr || action->handle == XR_NULL_HANDLE, false);
2021-12-14 02:44:12 +01:00
binding.action = action->handle;
XrResult result = xrStringToPath(instance, p_path.utf8().get_data(), &binding.binding);
if (XR_FAILED(result)) {
print_line("OpenXR: failed to get path for ", p_path, "! [", get_error_string(result), "]");
return false;
2021-12-14 02:44:12 +01:00
}
ip->bindings.push_back(binding);
return true;
}
bool OpenXRAPI::interaction_profile_suggest_bindings(RID p_interaction_profile) {
ERR_FAIL_COND_V(instance == XR_NULL_HANDLE, false);
InteractionProfile *ip = interaction_profile_owner.get_or_null(p_interaction_profile);
ERR_FAIL_NULL_V(ip, false);
2021-12-14 02:44:12 +01:00
const XrInteractionProfileSuggestedBinding suggested_bindings = {
XR_TYPE_INTERACTION_PROFILE_SUGGESTED_BINDING, // type
nullptr, // next
ip->path, // interactionProfile
uint32_t(ip->bindings.size()), // countSuggestedBindings
ip->bindings.ptr() // suggestedBindings
2021-12-14 02:44:12 +01:00
};
XrResult result = xrSuggestInteractionProfileBindings(instance, &suggested_bindings);
if (result == XR_ERROR_PATH_UNSUPPORTED) {
// this is fine, not all runtimes support all devices.
print_verbose("OpenXR Interaction profile " + ip->name + " is not supported on this runtime");
} else if (XR_FAILED(result)) {
print_line("OpenXR: failed to suggest bindings for ", ip->name, "! [", get_error_string(result), "]");
2021-12-14 02:44:12 +01:00
// reporting is enough...
}
/* For debugging:
print_verbose("Suggested bindings for " + ip->name);
for (int i = 0; i < ip->bindings.size(); i++) {
uint32_t strlen;
char path[XR_MAX_PATH_LENGTH];
String action_name = action_get_name(get_action_rid(ip->bindings[i].action));
XrResult result = xrPathToString(instance, ip->bindings[i].binding, XR_MAX_PATH_LENGTH, &strlen, path);
if (XR_FAILED(result)) {
print_line("OpenXR: failed to retrieve bindings for ", action_name, "! [", get_error_string(result), "]");
}
print_verbose(" - " + action_name + " => " + String(path));
}
*/
2021-12-14 02:44:12 +01:00
return true;
}
void OpenXRAPI::interaction_profile_free(RID p_interaction_profile) {
InteractionProfile *ip = interaction_profile_owner.get_or_null(p_interaction_profile);
ERR_FAIL_NULL(ip);
ip->bindings.clear();
interaction_profile_owner.free(p_interaction_profile);
}
2021-12-14 02:44:12 +01:00
bool OpenXRAPI::sync_action_sets(const Vector<RID> p_active_sets) {
ERR_FAIL_COND_V(session == XR_NULL_HANDLE, false);
if (!running) {
return false;
}
Vector<XrActiveActionSet> active_sets;
for (int i = 0; i < p_active_sets.size(); i++) {
ActionSet *action_set = action_set_owner.get_or_null(p_active_sets[i]);
if (action_set && action_set->handle != XR_NULL_HANDLE) {
XrActiveActionSet aset;
aset.actionSet = action_set->handle;
aset.subactionPath = XR_NULL_PATH;
active_sets.push_back(aset);
}
}
ERR_FAIL_COND_V(active_sets.is_empty(), false);
2021-12-14 02:44:12 +01:00
XrActionsSyncInfo sync_info = {
XR_TYPE_ACTIONS_SYNC_INFO, // type
nullptr, // next
uint32_t(active_sets.size()), // countActiveActionSets
active_sets.ptr() // activeActionSets
};
XrResult result = xrSyncActions(session, &sync_info);
if (XR_FAILED(result)) {
print_line("OpenXR: failed to sync active action sets! [", get_error_string(result), "]");
return false;
}
return true;
}
bool OpenXRAPI::get_action_bool(RID p_action, RID p_tracker) {
2021-12-14 02:44:12 +01:00
ERR_FAIL_COND_V(session == XR_NULL_HANDLE, false);
Action *action = action_owner.get_or_null(p_action);
ERR_FAIL_NULL_V(action, false);
Tracker *tracker = tracker_owner.get_or_null(p_tracker);
ERR_FAIL_NULL_V(tracker, false);
2021-12-14 02:44:12 +01:00
if (!running) {
return false;
}
ERR_FAIL_COND_V(action->action_type != XR_ACTION_TYPE_BOOLEAN_INPUT, false);
XrActionStateGetInfo get_info = {
XR_TYPE_ACTION_STATE_GET_INFO, // type
nullptr, // next
action->handle, // action
tracker->toplevel_path // subactionPath
2021-12-14 02:44:12 +01:00
};
XrActionStateBoolean result_state;
result_state.type = XR_TYPE_ACTION_STATE_BOOLEAN,
result_state.next = nullptr;
XrResult result = xrGetActionStateBoolean(session, &get_info, &result_state);
if (XR_FAILED(result)) {
print_line("OpenXR: couldn't get action boolean! [", get_error_string(result), "]");
return false;
}
return result_state.isActive && result_state.currentState;
}
float OpenXRAPI::get_action_float(RID p_action, RID p_tracker) {
2021-12-14 02:44:12 +01:00
ERR_FAIL_COND_V(session == XR_NULL_HANDLE, 0.0);
Action *action = action_owner.get_or_null(p_action);
ERR_FAIL_NULL_V(action, 0.0);
Tracker *tracker = tracker_owner.get_or_null(p_tracker);
ERR_FAIL_NULL_V(tracker, 0.0);
2021-12-14 02:44:12 +01:00
if (!running) {
return 0.0;
}
ERR_FAIL_COND_V(action->action_type != XR_ACTION_TYPE_FLOAT_INPUT, 0.0);
XrActionStateGetInfo get_info = {
XR_TYPE_ACTION_STATE_GET_INFO, // type
nullptr, // next
action->handle, // action
tracker->toplevel_path // subactionPath
2021-12-14 02:44:12 +01:00
};
XrActionStateFloat result_state;
result_state.type = XR_TYPE_ACTION_STATE_FLOAT,
result_state.next = nullptr;
XrResult result = xrGetActionStateFloat(session, &get_info, &result_state);
if (XR_FAILED(result)) {
print_line("OpenXR: couldn't get action float! [", get_error_string(result), "]");
return 0.0;
}
return result_state.isActive ? result_state.currentState : 0.0;
}
Vector2 OpenXRAPI::get_action_vector2(RID p_action, RID p_tracker) {
2021-12-14 02:44:12 +01:00
ERR_FAIL_COND_V(session == XR_NULL_HANDLE, Vector2());
Action *action = action_owner.get_or_null(p_action);
ERR_FAIL_NULL_V(action, Vector2());
Tracker *tracker = tracker_owner.get_or_null(p_tracker);
ERR_FAIL_NULL_V(tracker, Vector2());
2021-12-14 02:44:12 +01:00
if (!running) {
return Vector2();
}
ERR_FAIL_COND_V(action->action_type != XR_ACTION_TYPE_VECTOR2F_INPUT, Vector2());
XrActionStateGetInfo get_info = {
XR_TYPE_ACTION_STATE_GET_INFO, // type
nullptr, // next
action->handle, // action
tracker->toplevel_path // subactionPath
2021-12-14 02:44:12 +01:00
};
XrActionStateVector2f result_state;
result_state.type = XR_TYPE_ACTION_STATE_VECTOR2F,
result_state.next = nullptr;
XrResult result = xrGetActionStateVector2f(session, &get_info, &result_state);
if (XR_FAILED(result)) {
print_line("OpenXR: couldn't get action vector2! [", get_error_string(result), "]");
return Vector2();
}
return result_state.isActive ? Vector2(result_state.currentState.x, result_state.currentState.y) : Vector2();
}
2022-06-14 16:33:03 +02:00
XRPose::TrackingConfidence OpenXRAPI::get_action_pose(RID p_action, RID p_tracker, Transform3D &r_transform, Vector3 &r_linear_velocity, Vector3 &r_angular_velocity) {
2021-12-14 02:44:12 +01:00
ERR_FAIL_COND_V(session == XR_NULL_HANDLE, XRPose::XR_TRACKING_CONFIDENCE_NONE);
Action *action = action_owner.get_or_null(p_action);
ERR_FAIL_NULL_V(action, XRPose::XR_TRACKING_CONFIDENCE_NONE);
Tracker *tracker = tracker_owner.get_or_null(p_tracker);
ERR_FAIL_NULL_V(tracker, XRPose::XR_TRACKING_CONFIDENCE_NONE);
2021-12-14 02:44:12 +01:00
if (!running) {
return XRPose::XR_TRACKING_CONFIDENCE_NONE;
}
ERR_FAIL_COND_V(action->action_type != XR_ACTION_TYPE_POSE_INPUT, XRPose::XR_TRACKING_CONFIDENCE_NONE);
// print_verbose("Checking " + action->name + " => " + tracker->name + " (" + itos(tracker->toplevel_path) + ")");
2021-12-14 02:44:12 +01:00
uint64_t index = 0xFFFFFFFF;
uint64_t size = uint64_t(action->trackers.size());
2021-12-14 02:44:12 +01:00
for (uint64_t i = 0; i < size && index == 0xFFFFFFFF; i++) {
if (action->trackers[i].tracker_rid == p_tracker) {
2021-12-14 02:44:12 +01:00
index = i;
}
}
if (index == 0xFFFFFFFF) {
// couldn't find it?
return XRPose::XR_TRACKING_CONFIDENCE_NONE;
}
XrTime display_time = get_predicted_display_time();
if (display_time == 0) {
return XRPose::XR_TRACKING_CONFIDENCE_NONE;
}
if (action->trackers[index].space == XR_NULL_HANDLE) {
2021-12-14 02:44:12 +01:00
// if this is a pose we need to define spaces
XrActionSpaceCreateInfo action_space_info = {
XR_TYPE_ACTION_SPACE_CREATE_INFO, // type
nullptr, // next
action->handle, // action
tracker->toplevel_path, // subactionPath
2021-12-14 02:44:12 +01:00
{
{ 0.0, 0.0, 0.0, 1.0 }, // orientation
{ 0.0, 0.0, 0.0 } // position
} // poseInActionSpace
};
XrSpace space;
XrResult result = xrCreateActionSpace(session, &action_space_info, &space);
if (XR_FAILED(result)) {
print_line("OpenXR: couldn't create action space! [", get_error_string(result), "]");
return XRPose::XR_TRACKING_CONFIDENCE_NONE;
}
action->trackers.ptrw()[index].space = space;
2021-12-14 02:44:12 +01:00
}
XrSpaceVelocity velocity = {
XR_TYPE_SPACE_VELOCITY, // type
nullptr, // next
0, // velocityFlags
{ 0.0, 0.0, 0.0 }, // linearVelocity
{ 0.0, 0.0, 0.0 } // angularVelocity
};
XrSpaceLocation location = {
XR_TYPE_SPACE_LOCATION, // type
&velocity, // next
0, // locationFlags
{
{ 0.0, 0.0, 0.0, 0.0 }, // orientation
{ 0.0, 0.0, 0.0 } // position
} // pose
};
XrResult result = xrLocateSpace(action->trackers[index].space, play_space, display_time, &location);
2021-12-14 02:44:12 +01:00
if (XR_FAILED(result)) {
print_line("OpenXR: failed to locate space! [", get_error_string(result), "]");
return XRPose::XR_TRACKING_CONFIDENCE_NONE;
}
XRPose::TrackingConfidence confidence = transform_from_location(location, r_transform);
parse_velocities(velocity, r_linear_velocity, r_angular_velocity);
return confidence;
}
bool OpenXRAPI::trigger_haptic_pulse(RID p_action, RID p_tracker, float p_frequency, float p_amplitude, XrDuration p_duration_ns) {
2021-12-14 02:44:12 +01:00
ERR_FAIL_COND_V(session == XR_NULL_HANDLE, false);
Action *action = action_owner.get_or_null(p_action);
ERR_FAIL_NULL_V(action, false);
Tracker *tracker = tracker_owner.get_or_null(p_tracker);
ERR_FAIL_NULL_V(tracker, false);
2021-12-14 02:44:12 +01:00
if (!running) {
return false;
}
ERR_FAIL_COND_V(action->action_type != XR_ACTION_TYPE_VIBRATION_OUTPUT, false);
XrHapticActionInfo action_info = {
XR_TYPE_HAPTIC_ACTION_INFO, // type
nullptr, // next
action->handle, // action
tracker->toplevel_path // subactionPath
2021-12-14 02:44:12 +01:00
};
XrHapticVibration vibration = {
XR_TYPE_HAPTIC_VIBRATION, // type
nullptr, // next
p_duration_ns, // duration
p_frequency, // frequency
p_amplitude, // amplitude
};
XrResult result = xrApplyHapticFeedback(session, &action_info, (const XrHapticBaseHeader *)&vibration);
if (XR_FAILED(result)) {
print_line("OpenXR: failed to apply haptic feedback! [", get_error_string(result), "]");
return false;
}
return true;
}
2022-09-26 09:04:12 +02:00
void OpenXRAPI::register_composition_layer_provider(OpenXRCompositionLayerProvider *provider) {
composition_layer_providers.append(provider);
}
void OpenXRAPI::unregister_composition_layer_provider(OpenXRCompositionLayerProvider *provider) {
composition_layer_providers.erase(provider);
}
2023-02-02 15:14:16 +01:00
const XrEnvironmentBlendMode *OpenXRAPI::get_supported_environment_blend_modes(uint32_t &count) {
count = num_supported_environment_blend_modes;
return supported_environment_blend_modes;
}
bool OpenXRAPI::is_environment_blend_mode_supported(XrEnvironmentBlendMode p_blend_mode) const {
ERR_FAIL_NULL_V(supported_environment_blend_modes, false);
2023-02-02 15:14:16 +01:00
for (uint32_t i = 0; i < num_supported_environment_blend_modes; i++) {
if (supported_environment_blend_modes[i] == p_blend_mode) {
2023-02-02 15:14:16 +01:00
return true;
}
}
return false;
}
bool OpenXRAPI::set_environment_blend_mode(XrEnvironmentBlendMode p_blend_mode) {
if (emulate_environment_blend_mode_alpha_blend && p_blend_mode == XR_ENVIRONMENT_BLEND_MODE_ALPHA_BLEND) {
requested_environment_blend_mode = XR_ENVIRONMENT_BLEND_MODE_ALPHA_BLEND;
environment_blend_mode = XR_ENVIRONMENT_BLEND_MODE_OPAQUE;
return true;
}
// We allow setting this when not initialized and will check if it is supported when initializing.
// After OpenXR is initialized we verify we're setting a supported blend mode.
else if (!is_initialized() || is_environment_blend_mode_supported(p_blend_mode)) {
requested_environment_blend_mode = p_blend_mode;
environment_blend_mode = p_blend_mode;
return true;
}
2023-02-02 15:14:16 +01:00
return false;
}
void OpenXRAPI::set_emulate_environment_blend_mode_alpha_blend(bool p_enabled) {
emulate_environment_blend_mode_alpha_blend = p_enabled;
}
OpenXRAPI::OpenXRAlphaBlendModeSupport OpenXRAPI::is_environment_blend_mode_alpha_blend_supported() {
if (is_environment_blend_mode_supported(XR_ENVIRONMENT_BLEND_MODE_ALPHA_BLEND)) {
return OPENXR_ALPHA_BLEND_MODE_SUPPORT_REAL;
} else if (emulate_environment_blend_mode_alpha_blend) {
return OPENXR_ALPHA_BLEND_MODE_SUPPORT_EMULATING;
}
return OPENXR_ALPHA_BLEND_MODE_SUPPORT_NONE;
}