OLD | NEW |
| (Empty) |
1 // Copyright 2016 The Chromium Authors. All rights reserved. | |
2 // Use of this source code is governed by a BSD-style license that can be | |
3 // found in the LICENSE file. | |
4 | |
5 #include "device/vr/android/gvr/gvr_device.h" | |
6 | |
7 #include <math.h> | |
8 #include <algorithm> | |
9 | |
10 #include "base/time/time.h" | |
11 #include "base/trace_event/trace_event.h" | |
12 #include "third_party/gvr-android-sdk/src/ndk-beta/include/vr/gvr/capi/include/g
vr.h" | |
13 #include "third_party/gvr-android-sdk/src/ndk-beta/include/vr/gvr/capi/include/g
vr_types.h" | |
14 #include "ui/gfx/transform.h" | |
15 #include "ui/gfx/transform_util.h" | |
16 | |
17 namespace device { | |
18 | |
19 namespace { | |
20 | |
21 static const uint64_t kPredictionTimeWithoutVsyncNanos = 50000000; | |
22 | |
23 } // namespace | |
24 | |
25 GvrDevice::GvrDevice(VRDeviceProvider* provider, gvr::GvrApi* gvr_api) | |
26 : VRDevice(provider), gvr_api_(gvr_api) {} | |
27 | |
28 GvrDevice::~GvrDevice() {} | |
29 | |
30 VRDisplayPtr GvrDevice::GetVRDevice() { | |
31 TRACE_EVENT0("input", "GvrDevice::GetVRDevice"); | |
32 | |
33 VRDisplayPtr device = VRDisplay::New(); | |
34 | |
35 device->capabilities = VRDisplayCapabilities::New(); | |
36 device->capabilities->hasOrientation = true; | |
37 device->capabilities->hasPosition = false; | |
38 device->capabilities->hasExternalDisplay = false; | |
39 device->capabilities->canPresent = true; | |
40 | |
41 device->leftEye = VREyeParameters::New(); | |
42 device->rightEye = VREyeParameters::New(); | |
43 VREyeParametersPtr& left_eye = device->leftEye; | |
44 VREyeParametersPtr& right_eye = device->rightEye; | |
45 | |
46 device->displayName = gvr_api_->GetViewerModel(); | |
47 | |
48 gvr::BufferViewportList gvr_buffer_viewports = | |
49 gvr_api_->CreateEmptyBufferViewportList(); | |
50 gvr_buffer_viewports.SetToRecommendedBufferViewports(); | |
51 | |
52 gvr::BufferViewport eye_viewport = gvr_api_->CreateBufferViewport(); | |
53 gvr_buffer_viewports.GetBufferViewport(GVR_LEFT_EYE, &eye_viewport); | |
54 gvr::Rectf eye_fov = eye_viewport.GetSourceFov(); | |
55 left_eye->fieldOfView = VRFieldOfView::New(); | |
56 left_eye->fieldOfView->upDegrees = eye_fov.top; | |
57 left_eye->fieldOfView->downDegrees = eye_fov.bottom; | |
58 left_eye->fieldOfView->leftDegrees = eye_fov.left; | |
59 left_eye->fieldOfView->rightDegrees = eye_fov.right; | |
60 | |
61 eye_viewport = gvr_api_->CreateBufferViewport(); | |
62 gvr_buffer_viewports.GetBufferViewport(GVR_RIGHT_EYE, &eye_viewport); | |
63 eye_fov = eye_viewport.GetSourceFov(); | |
64 right_eye->fieldOfView = VRFieldOfView::New(); | |
65 right_eye->fieldOfView->upDegrees = eye_fov.top; | |
66 right_eye->fieldOfView->downDegrees = eye_fov.bottom; | |
67 right_eye->fieldOfView->leftDegrees = eye_fov.left; | |
68 right_eye->fieldOfView->rightDegrees = eye_fov.right; | |
69 | |
70 gvr::Mat4f left_eye_mat = gvr_api_->GetEyeFromHeadMatrix(GVR_LEFT_EYE); | |
71 gvr::Mat4f right_eye_mat = gvr_api_->GetEyeFromHeadMatrix(GVR_RIGHT_EYE); | |
72 | |
73 left_eye->offset = mojo::Array<float>::New(3); | |
74 left_eye->offset[0] = -left_eye_mat.m[0][3]; | |
75 left_eye->offset[1] = -left_eye_mat.m[1][3]; | |
76 left_eye->offset[2] = -left_eye_mat.m[2][3]; | |
77 | |
78 right_eye->offset = mojo::Array<float>::New(3); | |
79 right_eye->offset[0] = -right_eye_mat.m[0][3]; | |
80 right_eye->offset[1] = -right_eye_mat.m[1][3]; | |
81 right_eye->offset[2] = -right_eye_mat.m[2][3]; | |
82 | |
83 gvr::Sizei render_target_size = gvr_api_->GetRecommendedRenderTargetSize(); | |
84 | |
85 left_eye->renderWidth = render_target_size.width / 2; | |
86 left_eye->renderHeight = render_target_size.height; | |
87 | |
88 right_eye->renderWidth = render_target_size.width / 2; | |
89 right_eye->renderHeight = render_target_size.height; | |
90 | |
91 return device; | |
92 } | |
93 | |
94 VRPosePtr GvrDevice::GetPose() { | |
95 TRACE_EVENT0("input", "GvrDevice::GetSensorState"); | |
96 | |
97 VRPosePtr pose = VRPose::New(); | |
98 | |
99 pose->timestamp = base::Time::Now().ToJsTime(); | |
100 | |
101 gvr::ClockTimePoint target_time = gvr::GvrApi::GetTimePointNow(); | |
102 target_time.monotonic_system_time_nanos += kPredictionTimeWithoutVsyncNanos; | |
103 | |
104 gvr::Mat4f head_mat = gvr_api_->GetHeadPoseInStartSpace(target_time); | |
105 | |
106 gfx::Transform inv_transform( | |
107 head_mat.m[0][0], head_mat.m[0][1], head_mat.m[0][2], head_mat.m[0][3], | |
108 head_mat.m[1][0], head_mat.m[1][1], head_mat.m[1][2], head_mat.m[1][3], | |
109 head_mat.m[2][0], head_mat.m[2][1], head_mat.m[2][2], head_mat.m[2][3], | |
110 head_mat.m[3][0], head_mat.m[3][1], head_mat.m[3][2], head_mat.m[3][3]); | |
111 | |
112 gfx::Transform transform; | |
113 if (inv_transform.GetInverse(&transform)) { | |
114 gfx::DecomposedTransform decomposed_transform; | |
115 gfx::DecomposeTransform(&decomposed_transform, transform); | |
116 | |
117 pose->orientation = mojo::Array<float>::New(4); | |
118 pose->orientation[0] = decomposed_transform.quaternion[0]; | |
119 pose->orientation[1] = decomposed_transform.quaternion[1]; | |
120 pose->orientation[2] = decomposed_transform.quaternion[2]; | |
121 pose->orientation[3] = decomposed_transform.quaternion[3]; | |
122 | |
123 pose->position = mojo::Array<float>::New(3); | |
124 pose->position[0] = decomposed_transform.translate[0]; | |
125 pose->position[1] = decomposed_transform.translate[1]; | |
126 pose->position[2] = decomposed_transform.translate[2]; | |
127 } | |
128 | |
129 return pose; | |
130 } | |
131 | |
132 void GvrDevice::ResetPose() { | |
133 gvr_api_->ResetTracking(); | |
134 } | |
135 | |
136 } // namespace device | |
OLD | NEW |