753 lines
24 KiB
Text
753 lines
24 KiB
Text
/*************************************************************************/
|
|
/* gl_view.mm */
|
|
/*************************************************************************/
|
|
/* This file is part of: */
|
|
/* GODOT ENGINE */
|
|
/* https://godotengine.org */
|
|
/*************************************************************************/
|
|
/* Copyright (c) 2007-2018 Juan Linietsky, Ariel Manzur. */
|
|
/* Copyright (c) 2014-2018 Godot Engine contributors (cf. AUTHORS.md) */
|
|
/* */
|
|
/* Permission is hereby granted, free of charge, to any person obtaining */
|
|
/* a copy of this software and associated documentation files (the */
|
|
/* "Software"), to deal in the Software without restriction, including */
|
|
/* without limitation the rights to use, copy, modify, merge, publish, */
|
|
/* distribute, sublicense, and/or sell copies of the Software, and to */
|
|
/* permit persons to whom the Software is furnished to do so, subject to */
|
|
/* the following conditions: */
|
|
/* */
|
|
/* The above copyright notice and this permission notice shall be */
|
|
/* included in all copies or substantial portions of the Software. */
|
|
/* */
|
|
/* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, */
|
|
/* EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF */
|
|
/* MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT.*/
|
|
/* IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY */
|
|
/* CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, */
|
|
/* TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE */
|
|
/* SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE. */
|
|
/*************************************************************************/
|
|
|
|
#import "gl_view.h"
|
|
|
|
#include "core/os/keyboard.h"
|
|
#include "core/project_settings.h"
|
|
#include "os_iphone.h"
|
|
#include "servers/audio_server.h"
|
|
|
|
#import <OpenGLES/EAGLDrawable.h>
|
|
#import <QuartzCore/QuartzCore.h>
|
|
|
|
/*
|
|
@interface GLView (private)
|
|
|
|
- (id)initGLES;
|
|
- (BOOL)createFramebuffer;
|
|
- (void)destroyFramebuffer;
|
|
@end
|
|
*/
|
|
|
|
bool gles3_available = true;
|
|
int gl_view_base_fb;
|
|
static String keyboard_text;
|
|
static GLView *_instance = NULL;
|
|
|
|
static bool video_found_error = false;
|
|
static bool video_playing = false;
|
|
static CMTime video_current_time;
|
|
|
|
void _show_keyboard(String);
|
|
void _hide_keyboard();
|
|
bool _play_video(String, float, String, String);
|
|
bool _is_video_playing();
|
|
void _pause_video();
|
|
void _focus_out_video();
|
|
void _unpause_video();
|
|
void _stop_video();
|
|
CGFloat _points_to_pixels(CGFloat);
|
|
|
|
void _show_keyboard(String p_existing) {
|
|
keyboard_text = p_existing;
|
|
printf("instance on show is %p\n", _instance);
|
|
[_instance open_keyboard];
|
|
};
|
|
|
|
void _hide_keyboard() {
|
|
printf("instance on hide is %p\n", _instance);
|
|
[_instance hide_keyboard];
|
|
keyboard_text = "";
|
|
};
|
|
|
|
Rect2 _get_ios_window_safe_area(float p_window_width, float p_window_height) {
|
|
UIEdgeInsets insets = UIEdgeInsetsMake(0, 0, 0, 0);
|
|
if (_instance != nil && [_instance respondsToSelector:@selector(safeAreaInsets)]) {
|
|
insets = [_instance safeAreaInsets];
|
|
}
|
|
ERR_FAIL_COND_V(insets.left < 0 || insets.top < 0 || insets.right < 0 || insets.bottom < 0,
|
|
Rect2(0, 0, p_window_width, p_window_height));
|
|
UIEdgeInsets window_insets = UIEdgeInsetsMake(_points_to_pixels(insets.top), _points_to_pixels(insets.left), _points_to_pixels(insets.bottom), _points_to_pixels(insets.right));
|
|
return Rect2(window_insets.left, window_insets.top, p_window_width - window_insets.right - window_insets.left, p_window_height - window_insets.bottom - window_insets.top);
|
|
}
|
|
|
|
bool _play_video(String p_path, float p_volume, String p_audio_track, String p_subtitle_track) {
|
|
p_path = ProjectSettings::get_singleton()->globalize_path(p_path);
|
|
|
|
NSString *file_path = [[[NSString alloc] initWithUTF8String:p_path.utf8().get_data()] autorelease];
|
|
|
|
_instance.avAsset = [AVAsset assetWithURL:[NSURL fileURLWithPath:file_path]];
|
|
|
|
_instance.avPlayerItem = [[AVPlayerItem alloc] initWithAsset:_instance.avAsset];
|
|
[_instance.avPlayerItem addObserver:_instance forKeyPath:@"status" options:0 context:nil];
|
|
|
|
_instance.avPlayer = [[AVPlayer alloc] initWithPlayerItem:_instance.avPlayerItem];
|
|
_instance.avPlayerLayer = [AVPlayerLayer playerLayerWithPlayer:_instance.avPlayer];
|
|
|
|
[_instance.avPlayer addObserver:_instance forKeyPath:@"status" options:0 context:nil];
|
|
[[NSNotificationCenter defaultCenter]
|
|
addObserver:_instance
|
|
selector:@selector(playerItemDidReachEnd:)
|
|
name:AVPlayerItemDidPlayToEndTimeNotification
|
|
object:[_instance.avPlayer currentItem]];
|
|
|
|
[_instance.avPlayer addObserver:_instance forKeyPath:@"rate" options:NSKeyValueObservingOptionNew context:0];
|
|
|
|
[_instance.avPlayerLayer setFrame:_instance.bounds];
|
|
[_instance.layer addSublayer:_instance.avPlayerLayer];
|
|
[_instance.avPlayer play];
|
|
|
|
AVMediaSelectionGroup *audioGroup = [_instance.avAsset mediaSelectionGroupForMediaCharacteristic:AVMediaCharacteristicAudible];
|
|
|
|
NSMutableArray *allAudioParams = [NSMutableArray array];
|
|
for (id track in audioGroup.options) {
|
|
NSString *language = [[track locale] localeIdentifier];
|
|
NSLog(@"subtitle lang: %@", language);
|
|
|
|
if ([language isEqualToString:[NSString stringWithUTF8String:p_audio_track.utf8()]]) {
|
|
AVMutableAudioMixInputParameters *audioInputParams = [AVMutableAudioMixInputParameters audioMixInputParameters];
|
|
[audioInputParams setVolume:p_volume atTime:kCMTimeZero];
|
|
[audioInputParams setTrackID:[track trackID]];
|
|
[allAudioParams addObject:audioInputParams];
|
|
|
|
AVMutableAudioMix *audioMix = [AVMutableAudioMix audioMix];
|
|
[audioMix setInputParameters:allAudioParams];
|
|
|
|
[_instance.avPlayer.currentItem selectMediaOption:track inMediaSelectionGroup:audioGroup];
|
|
[_instance.avPlayer.currentItem setAudioMix:audioMix];
|
|
|
|
break;
|
|
}
|
|
}
|
|
|
|
AVMediaSelectionGroup *subtitlesGroup = [_instance.avAsset mediaSelectionGroupForMediaCharacteristic:AVMediaCharacteristicLegible];
|
|
NSArray *useableTracks = [AVMediaSelectionGroup mediaSelectionOptionsFromArray:subtitlesGroup.options withoutMediaCharacteristics:[NSArray arrayWithObject:AVMediaCharacteristicContainsOnlyForcedSubtitles]];
|
|
|
|
for (id track in useableTracks) {
|
|
NSString *language = [[track locale] localeIdentifier];
|
|
NSLog(@"subtitle lang: %@", language);
|
|
|
|
if ([language isEqualToString:[NSString stringWithUTF8String:p_subtitle_track.utf8()]]) {
|
|
[_instance.avPlayer.currentItem selectMediaOption:track inMediaSelectionGroup:subtitlesGroup];
|
|
break;
|
|
}
|
|
}
|
|
|
|
video_playing = true;
|
|
|
|
return true;
|
|
}
|
|
|
|
bool _is_video_playing() {
|
|
if (_instance.avPlayer.error) {
|
|
printf("Error during playback\n");
|
|
}
|
|
return (_instance.avPlayer.rate > 0 && !_instance.avPlayer.error);
|
|
}
|
|
|
|
void _pause_video() {
|
|
video_current_time = _instance.avPlayer.currentTime;
|
|
[_instance.avPlayer pause];
|
|
video_playing = false;
|
|
}
|
|
|
|
void _focus_out_video() {
|
|
printf("focus out pausing video\n");
|
|
[_instance.avPlayer pause];
|
|
};
|
|
|
|
void _unpause_video() {
|
|
|
|
[_instance.avPlayer play];
|
|
video_playing = true;
|
|
};
|
|
|
|
void _stop_video() {
|
|
[_instance.avPlayer pause];
|
|
[_instance.avPlayerLayer removeFromSuperlayer];
|
|
_instance.avPlayer = nil;
|
|
video_playing = false;
|
|
}
|
|
|
|
CGFloat _points_to_pixels(CGFloat points) {
|
|
float pixelPerInch;
|
|
if (UI_USER_INTERFACE_IDIOM() == UIUserInterfaceIdiomPad) {
|
|
pixelPerInch = 132;
|
|
} else if (UI_USER_INTERFACE_IDIOM() == UIUserInterfaceIdiomPhone) {
|
|
pixelPerInch = 163;
|
|
} else {
|
|
pixelPerInch = 160;
|
|
}
|
|
CGFloat pointsPerInch = 72.0;
|
|
return (points / pointsPerInch * pixelPerInch);
|
|
}
|
|
|
|
@implementation GLView
|
|
|
|
@synthesize animationInterval;
|
|
|
|
static const int max_touches = 8;
|
|
static UITouch *touches[max_touches];
|
|
|
|
static void init_touches() {
|
|
|
|
for (int i = 0; i < max_touches; i++) {
|
|
touches[i] = NULL;
|
|
};
|
|
};
|
|
|
|
static int get_touch_id(UITouch *p_touch) {
|
|
|
|
int first = -1;
|
|
for (int i = 0; i < max_touches; i++) {
|
|
if (first == -1 && touches[i] == NULL) {
|
|
first = i;
|
|
continue;
|
|
};
|
|
if (touches[i] == p_touch)
|
|
return i;
|
|
};
|
|
|
|
if (first != -1) {
|
|
touches[first] = p_touch;
|
|
return first;
|
|
};
|
|
|
|
return -1;
|
|
};
|
|
|
|
static int remove_touch(UITouch *p_touch) {
|
|
|
|
int remaining = 0;
|
|
for (int i = 0; i < max_touches; i++) {
|
|
|
|
if (touches[i] == NULL)
|
|
continue;
|
|
if (touches[i] == p_touch)
|
|
touches[i] = NULL;
|
|
else
|
|
++remaining;
|
|
};
|
|
return remaining;
|
|
};
|
|
|
|
static void clear_touches() {
|
|
|
|
for (int i = 0; i < max_touches; i++) {
|
|
|
|
touches[i] = NULL;
|
|
};
|
|
};
|
|
|
|
// Implement this to override the default layer class (which is [CALayer class]).
|
|
// We do this so that our view will be backed by a layer that is capable of OpenGL ES rendering.
|
|
+ (Class)layerClass {
|
|
return [CAEAGLLayer class];
|
|
}
|
|
|
|
//The GL view is stored in the nib file. When it's unarchived it's sent -initWithCoder:
|
|
- (id)initWithCoder:(NSCoder *)coder {
|
|
active = FALSE;
|
|
if ((self = [super initWithCoder:coder])) {
|
|
self = [self initGLES];
|
|
}
|
|
return self;
|
|
}
|
|
|
|
- (id)initGLES {
|
|
// Get our backing layer
|
|
CAEAGLLayer *eaglLayer = (CAEAGLLayer *)self.layer;
|
|
|
|
// Configure it so that it is opaque, does not retain the contents of the backbuffer when displayed, and uses RGBA8888 color.
|
|
eaglLayer.opaque = YES;
|
|
eaglLayer.drawableProperties = [NSDictionary
|
|
dictionaryWithObjectsAndKeys:[NSNumber numberWithBool:FALSE],
|
|
kEAGLDrawablePropertyRetainedBacking,
|
|
kEAGLColorFormatRGBA8,
|
|
kEAGLDrawablePropertyColorFormat,
|
|
nil];
|
|
|
|
// Create our EAGLContext, and if successful make it current and create our framebuffer.
|
|
context = [[EAGLContext alloc] initWithAPI:kEAGLRenderingAPIOpenGLES3];
|
|
|
|
if (!context || ![EAGLContext setCurrentContext:context] || ![self createFramebuffer]) {
|
|
context = [[EAGLContext alloc] initWithAPI:kEAGLRenderingAPIOpenGLES2];
|
|
gles3_available = false;
|
|
if (!context || ![EAGLContext setCurrentContext:context] || ![self createFramebuffer]) {
|
|
[self release];
|
|
return nil;
|
|
}
|
|
}
|
|
|
|
// Default the animation interval to 1/60th of a second.
|
|
animationInterval = 1.0 / 60.0;
|
|
return self;
|
|
}
|
|
|
|
- (id<GLViewDelegate>)delegate {
|
|
return delegate;
|
|
}
|
|
|
|
// Update the delegate, and if it needs a -setupView: call, set our internal flag so that it will be called.
|
|
- (void)setDelegate:(id<GLViewDelegate>)d {
|
|
delegate = d;
|
|
delegateSetup = ![delegate respondsToSelector:@selector(setupView:)];
|
|
}
|
|
|
|
@synthesize useCADisplayLink;
|
|
|
|
// If our view is resized, we'll be asked to layout subviews.
|
|
// This is the perfect opportunity to also update the framebuffer so that it is
|
|
// the same size as our display area.
|
|
|
|
- (void)layoutSubviews {
|
|
//printf("HERE\n");
|
|
[EAGLContext setCurrentContext:context];
|
|
[self destroyFramebuffer];
|
|
[self createFramebuffer];
|
|
[self drawView];
|
|
[self drawView];
|
|
}
|
|
|
|
- (BOOL)createFramebuffer {
|
|
// Generate IDs for a framebuffer object and a color renderbuffer
|
|
UIScreen *mainscr = [UIScreen mainScreen];
|
|
printf("******** screen size %i, %i\n", (int)mainscr.currentMode.size.width, (int)mainscr.currentMode.size.height);
|
|
self.contentScaleFactor = mainscr.nativeScale;
|
|
|
|
glGenFramebuffersOES(1, &viewFramebuffer);
|
|
glGenRenderbuffersOES(1, &viewRenderbuffer);
|
|
|
|
glBindFramebufferOES(GL_FRAMEBUFFER_OES, viewFramebuffer);
|
|
glBindRenderbufferOES(GL_RENDERBUFFER_OES, viewRenderbuffer);
|
|
// This call associates the storage for the current render buffer with the EAGLDrawable (our CAEAGLLayer)
|
|
// allowing us to draw into a buffer that will later be rendered to screen wherever the layer is (which corresponds with our view).
|
|
[context renderbufferStorage:GL_RENDERBUFFER_OES fromDrawable:(id<EAGLDrawable>)self.layer];
|
|
glFramebufferRenderbufferOES(GL_FRAMEBUFFER_OES, GL_COLOR_ATTACHMENT0_OES, GL_RENDERBUFFER_OES, viewRenderbuffer);
|
|
|
|
glGetRenderbufferParameterivOES(GL_RENDERBUFFER_OES, GL_RENDERBUFFER_WIDTH_OES, &backingWidth);
|
|
glGetRenderbufferParameterivOES(GL_RENDERBUFFER_OES, GL_RENDERBUFFER_HEIGHT_OES, &backingHeight);
|
|
|
|
// For this sample, we also need a depth buffer, so we'll create and attach one via another renderbuffer.
|
|
glGenRenderbuffersOES(1, &depthRenderbuffer);
|
|
glBindRenderbufferOES(GL_RENDERBUFFER_OES, depthRenderbuffer);
|
|
glRenderbufferStorageOES(GL_RENDERBUFFER_OES, GL_DEPTH_COMPONENT16_OES, backingWidth, backingHeight);
|
|
glFramebufferRenderbufferOES(GL_FRAMEBUFFER_OES, GL_DEPTH_ATTACHMENT_OES, GL_RENDERBUFFER_OES, depthRenderbuffer);
|
|
|
|
if (glCheckFramebufferStatusOES(GL_FRAMEBUFFER_OES) != GL_FRAMEBUFFER_COMPLETE_OES) {
|
|
NSLog(@"failed to make complete framebuffer object %x", glCheckFramebufferStatusOES(GL_FRAMEBUFFER_OES));
|
|
return NO;
|
|
}
|
|
|
|
if (OS::get_singleton()) {
|
|
OS::VideoMode vm;
|
|
vm.fullscreen = true;
|
|
vm.width = backingWidth;
|
|
vm.height = backingHeight;
|
|
vm.resizable = false;
|
|
OS::get_singleton()->set_video_mode(vm);
|
|
OSIPhone::get_singleton()->set_base_framebuffer(viewFramebuffer);
|
|
};
|
|
gl_view_base_fb = viewFramebuffer;
|
|
|
|
return YES;
|
|
}
|
|
|
|
// Clean up any buffers we have allocated.
|
|
- (void)destroyFramebuffer {
|
|
glDeleteFramebuffersOES(1, &viewFramebuffer);
|
|
viewFramebuffer = 0;
|
|
glDeleteRenderbuffersOES(1, &viewRenderbuffer);
|
|
viewRenderbuffer = 0;
|
|
|
|
if (depthRenderbuffer) {
|
|
glDeleteRenderbuffersOES(1, &depthRenderbuffer);
|
|
depthRenderbuffer = 0;
|
|
}
|
|
}
|
|
|
|
- (void)startAnimation {
|
|
if (active)
|
|
return;
|
|
active = TRUE;
|
|
printf("start animation!\n");
|
|
if (useCADisplayLink) {
|
|
|
|
// Approximate frame rate
|
|
// assumes device refreshes at 60 fps
|
|
int frameInterval = (int)floor(animationInterval * 60.0f);
|
|
|
|
displayLink = [CADisplayLink displayLinkWithTarget:self selector:@selector(drawView)];
|
|
[displayLink setFrameInterval:frameInterval];
|
|
|
|
// Setup DisplayLink in main thread
|
|
[displayLink addToRunLoop:[NSRunLoop currentRunLoop] forMode:NSRunLoopCommonModes];
|
|
} else {
|
|
animationTimer = [NSTimer scheduledTimerWithTimeInterval:animationInterval target:self selector:@selector(drawView) userInfo:nil repeats:YES];
|
|
}
|
|
|
|
if (video_playing) {
|
|
_unpause_video();
|
|
}
|
|
}
|
|
|
|
- (void)stopAnimation {
|
|
if (!active)
|
|
return;
|
|
active = FALSE;
|
|
printf("******** stop animation!\n");
|
|
|
|
if (useCADisplayLink) {
|
|
[displayLink invalidate];
|
|
displayLink = nil;
|
|
} else {
|
|
[animationTimer invalidate];
|
|
animationTimer = nil;
|
|
}
|
|
|
|
clear_touches();
|
|
|
|
if (video_playing) {
|
|
// save position
|
|
}
|
|
}
|
|
|
|
- (void)setAnimationInterval:(NSTimeInterval)interval {
|
|
animationInterval = interval;
|
|
if ((useCADisplayLink && displayLink) || (!useCADisplayLink && animationTimer)) {
|
|
[self stopAnimation];
|
|
[self startAnimation];
|
|
}
|
|
}
|
|
|
|
// Updates the OpenGL view when the timer fires
|
|
- (void)drawView {
|
|
if (useCADisplayLink) {
|
|
// Pause the CADisplayLink to avoid recursion
|
|
[displayLink setPaused:YES];
|
|
|
|
// Process all input events
|
|
while (CFRunLoopRunInMode(kCFRunLoopDefaultMode, 0, TRUE) == kCFRunLoopRunHandledSource)
|
|
;
|
|
|
|
// We are good to go, resume the CADisplayLink
|
|
[displayLink setPaused:NO];
|
|
}
|
|
|
|
if (!active) {
|
|
printf("draw view not active!\n");
|
|
return;
|
|
};
|
|
|
|
// Make sure that you are drawing to the current context
|
|
[EAGLContext setCurrentContext:context];
|
|
|
|
// If our drawing delegate needs to have the view setup, then call -setupView: and flag that it won't need to be called again.
|
|
if (!delegateSetup) {
|
|
[delegate setupView:self];
|
|
delegateSetup = YES;
|
|
}
|
|
|
|
glBindFramebufferOES(GL_FRAMEBUFFER_OES, viewFramebuffer);
|
|
|
|
[delegate drawView:self];
|
|
|
|
glBindRenderbufferOES(GL_RENDERBUFFER_OES, viewRenderbuffer);
|
|
[context presentRenderbuffer:GL_RENDERBUFFER_OES];
|
|
|
|
#ifdef DEBUG_ENABLED
|
|
GLenum err = glGetError();
|
|
if (err)
|
|
NSLog(@"%x error", err);
|
|
#endif
|
|
}
|
|
|
|
- (void)touchesBegan:(NSSet *)touches withEvent:(UIEvent *)event {
|
|
NSArray *tlist = [[event allTouches] allObjects];
|
|
for (unsigned int i = 0; i < [tlist count]; i++) {
|
|
|
|
if ([touches containsObject:[tlist objectAtIndex:i]]) {
|
|
|
|
UITouch *touch = [tlist objectAtIndex:i];
|
|
if (touch.phase != UITouchPhaseBegan)
|
|
continue;
|
|
int tid = get_touch_id(touch);
|
|
ERR_FAIL_COND(tid == -1);
|
|
CGPoint touchPoint = [touch locationInView:self];
|
|
OSIPhone::get_singleton()->touch_press(tid, touchPoint.x * self.contentScaleFactor, touchPoint.y * self.contentScaleFactor, true, touch.tapCount > 1);
|
|
};
|
|
};
|
|
}
|
|
|
|
- (void)touchesMoved:(NSSet *)touches withEvent:(UIEvent *)event {
|
|
|
|
NSArray *tlist = [[event allTouches] allObjects];
|
|
for (unsigned int i = 0; i < [tlist count]; i++) {
|
|
|
|
if ([touches containsObject:[tlist objectAtIndex:i]]) {
|
|
|
|
UITouch *touch = [tlist objectAtIndex:i];
|
|
if (touch.phase != UITouchPhaseMoved)
|
|
continue;
|
|
int tid = get_touch_id(touch);
|
|
ERR_FAIL_COND(tid == -1);
|
|
CGPoint touchPoint = [touch locationInView:self];
|
|
CGPoint prev_point = [touch previousLocationInView:self];
|
|
OSIPhone::get_singleton()->touch_drag(tid, prev_point.x * self.contentScaleFactor, prev_point.y * self.contentScaleFactor, touchPoint.x * self.contentScaleFactor, touchPoint.y * self.contentScaleFactor);
|
|
};
|
|
};
|
|
}
|
|
|
|
- (void)touchesEnded:(NSSet *)touches withEvent:(UIEvent *)event {
|
|
NSArray *tlist = [[event allTouches] allObjects];
|
|
for (unsigned int i = 0; i < [tlist count]; i++) {
|
|
|
|
if ([touches containsObject:[tlist objectAtIndex:i]]) {
|
|
|
|
UITouch *touch = [tlist objectAtIndex:i];
|
|
if (touch.phase != UITouchPhaseEnded)
|
|
continue;
|
|
int tid = get_touch_id(touch);
|
|
ERR_FAIL_COND(tid == -1);
|
|
remove_touch(touch);
|
|
CGPoint touchPoint = [touch locationInView:self];
|
|
OSIPhone::get_singleton()->touch_press(tid, touchPoint.x * self.contentScaleFactor, touchPoint.y * self.contentScaleFactor, false, false);
|
|
};
|
|
};
|
|
}
|
|
|
|
- (void)touchesCancelled:(NSSet *)touches withEvent:(UIEvent *)event {
|
|
|
|
OSIPhone::get_singleton()->touches_cancelled();
|
|
clear_touches();
|
|
};
|
|
|
|
- (BOOL)canBecomeFirstResponder {
|
|
return YES;
|
|
};
|
|
|
|
- (void)open_keyboard {
|
|
//keyboard_text = p_existing;
|
|
[self becomeFirstResponder];
|
|
};
|
|
|
|
- (void)hide_keyboard {
|
|
//keyboard_text = p_existing;
|
|
[self resignFirstResponder];
|
|
};
|
|
|
|
- (void)keyboardOnScreen:(NSNotification *)notification {
|
|
NSDictionary *info = notification.userInfo;
|
|
NSValue *value = info[UIKeyboardFrameEndUserInfoKey];
|
|
|
|
CGRect rawFrame = [value CGRectValue];
|
|
CGRect keyboardFrame = [self convertRect:rawFrame fromView:nil];
|
|
|
|
OSIPhone::get_singleton()->set_virtual_keyboard_height(_points_to_pixels(keyboardFrame.size.height));
|
|
}
|
|
|
|
- (void)keyboardHidden:(NSNotification *)notification {
|
|
OSIPhone::get_singleton()->set_virtual_keyboard_height(0);
|
|
}
|
|
|
|
- (void)deleteBackward {
|
|
if (keyboard_text.length())
|
|
keyboard_text.erase(keyboard_text.length() - 1, 1);
|
|
OSIPhone::get_singleton()->key(KEY_BACKSPACE, true);
|
|
};
|
|
|
|
- (BOOL)hasText {
|
|
return keyboard_text.length() ? YES : NO;
|
|
};
|
|
|
|
- (void)insertText:(NSString *)p_text {
|
|
String character;
|
|
character.parse_utf8([p_text UTF8String]);
|
|
keyboard_text = keyboard_text + character;
|
|
OSIPhone::get_singleton()->key(character[0] == 10 ? KEY_ENTER : character[0], true);
|
|
printf("inserting text with character %i\n", character[0]);
|
|
};
|
|
|
|
- (void)audioRouteChangeListenerCallback:(NSNotification *)notification {
|
|
printf("*********** route changed!\n");
|
|
NSDictionary *interuptionDict = notification.userInfo;
|
|
|
|
NSInteger routeChangeReason = [[interuptionDict valueForKey:AVAudioSessionRouteChangeReasonKey] integerValue];
|
|
|
|
switch (routeChangeReason) {
|
|
|
|
case AVAudioSessionRouteChangeReasonNewDeviceAvailable: {
|
|
NSLog(@"AVAudioSessionRouteChangeReasonNewDeviceAvailable");
|
|
NSLog(@"Headphone/Line plugged in");
|
|
}; break;
|
|
|
|
case AVAudioSessionRouteChangeReasonOldDeviceUnavailable: {
|
|
NSLog(@"AVAudioSessionRouteChangeReasonOldDeviceUnavailable");
|
|
NSLog(@"Headphone/Line was pulled. Resuming video play....");
|
|
if (_is_video_playing()) {
|
|
|
|
dispatch_after(dispatch_time(DISPATCH_TIME_NOW, 0.5f * NSEC_PER_SEC), dispatch_get_main_queue(), ^{
|
|
[_instance.avPlayer play]; // NOTE: change this line according your current player implementation
|
|
NSLog(@"resumed play");
|
|
});
|
|
};
|
|
}; break;
|
|
|
|
case AVAudioSessionRouteChangeReasonCategoryChange: {
|
|
// called at start - also when other audio wants to play
|
|
NSLog(@"AVAudioSessionRouteChangeReasonCategoryChange");
|
|
}; break;
|
|
}
|
|
}
|
|
|
|
// When created via code however, we get initWithFrame
|
|
- (id)initWithFrame:(CGRect)frame {
|
|
self = [super initWithFrame:frame];
|
|
_instance = self;
|
|
printf("after init super %p\n", self);
|
|
if (self != nil) {
|
|
self = [self initGLES];
|
|
printf("after init gles %p\n", self);
|
|
}
|
|
init_touches();
|
|
self.multipleTouchEnabled = YES;
|
|
self.autocorrectionType = UITextAutocorrectionTypeNo;
|
|
|
|
printf("******** adding observer for sound routing changes\n");
|
|
[[NSNotificationCenter defaultCenter]
|
|
addObserver:self
|
|
selector:@selector(audioRouteChangeListenerCallback:)
|
|
name:AVAudioSessionRouteChangeNotification
|
|
object:nil];
|
|
|
|
printf("******** adding observer for keyboard show/hide\n");
|
|
[[NSNotificationCenter defaultCenter]
|
|
addObserver:self
|
|
selector:@selector(keyboardOnScreen:)
|
|
name:UIKeyboardDidShowNotification
|
|
object:nil];
|
|
[[NSNotificationCenter defaultCenter]
|
|
addObserver:self
|
|
selector:@selector(keyboardHidden:)
|
|
name:UIKeyboardDidHideNotification
|
|
object:nil];
|
|
|
|
//self.autoresizesSubviews = YES;
|
|
//[self setAutoresizingMask:UIViewAutoresizingFlexibleWidth | UIViewAutoresizingFlexibleWidth];
|
|
|
|
return self;
|
|
}
|
|
|
|
//- (BOOL)automaticallyForwardAppearanceAndRotationMethodsToChildViewControllers {
|
|
// return YES;
|
|
//}
|
|
|
|
//- (BOOL)shouldAutorotateToInterfaceOrientation:(UIInterfaceOrientation)interfaceOrientation{
|
|
// return YES;
|
|
//}
|
|
|
|
// Stop animating and release resources when they are no longer needed.
|
|
- (void)dealloc {
|
|
[self stopAnimation];
|
|
|
|
if ([EAGLContext currentContext] == context) {
|
|
[EAGLContext setCurrentContext:nil];
|
|
}
|
|
|
|
[context release];
|
|
context = nil;
|
|
|
|
[super dealloc];
|
|
}
|
|
|
|
- (void)observeValueForKeyPath:(NSString *)keyPath ofObject:(id)object change:(NSDictionary *)change context:(void *)context {
|
|
|
|
if (object == _instance.avPlayerItem && [keyPath isEqualToString:@"status"]) {
|
|
if (_instance.avPlayerItem.status == AVPlayerStatusFailed || _instance.avPlayer.status == AVPlayerStatusFailed) {
|
|
_stop_video();
|
|
video_found_error = true;
|
|
}
|
|
|
|
if (_instance.avPlayer.status == AVPlayerStatusReadyToPlay &&
|
|
_instance.avPlayerItem.status == AVPlayerItemStatusReadyToPlay &&
|
|
CMTIME_COMPARE_INLINE(video_current_time, ==, kCMTimeZero)) {
|
|
|
|
//NSLog(@"time: %@", video_current_time);
|
|
|
|
[_instance.avPlayer seekToTime:video_current_time];
|
|
video_current_time = kCMTimeZero;
|
|
}
|
|
}
|
|
|
|
if (object == _instance.avPlayer && [keyPath isEqualToString:@"rate"]) {
|
|
NSLog(@"Player playback rate changed: %.5f", _instance.avPlayer.rate);
|
|
if (_is_video_playing() && _instance.avPlayer.rate == 0.0 && !_instance.avPlayer.error) {
|
|
dispatch_after(dispatch_time(DISPATCH_TIME_NOW, 0.5f * NSEC_PER_SEC), dispatch_get_main_queue(), ^{
|
|
[_instance.avPlayer play]; // NOTE: change this line according your current player implementation
|
|
NSLog(@"resumed play");
|
|
});
|
|
|
|
NSLog(@" . . . PAUSED (or just started)");
|
|
}
|
|
}
|
|
}
|
|
|
|
- (void)playerItemDidReachEnd:(NSNotification *)notification {
|
|
_stop_video();
|
|
}
|
|
|
|
/*
|
|
- (void)moviePlayBackDidFinish:(NSNotification*)notification {
|
|
|
|
|
|
NSNumber* reason = [[notification userInfo] objectForKey:MPMoviePlayerPlaybackDidFinishReasonUserInfoKey];
|
|
switch ([reason intValue]) {
|
|
case MPMovieFinishReasonPlaybackEnded:
|
|
//NSLog(@"Playback Ended");
|
|
break;
|
|
case MPMovieFinishReasonPlaybackError:
|
|
//NSLog(@"Playback Error");
|
|
video_found_error = true;
|
|
break;
|
|
case MPMovieFinishReasonUserExited:
|
|
//NSLog(@"User Exited");
|
|
video_found_error = true;
|
|
break;
|
|
default:
|
|
//NSLog(@"Unsupported reason!");
|
|
break;
|
|
}
|
|
|
|
MPMoviePlayerController *player = [notification object];
|
|
|
|
[[NSNotificationCenter defaultCenter]
|
|
removeObserver:self
|
|
name:MPMoviePlayerPlaybackDidFinishNotification
|
|
object:player];
|
|
|
|
[_instance.moviePlayerController stop];
|
|
[_instance.moviePlayerController.view removeFromSuperview];
|
|
|
|
video_playing = false;
|
|
}
|
|
*/
|
|
|
|
@end
|