mirror of
https://github.com/CTCaer/RetroArch.git
synced 2024-12-22 02:38:11 +00:00
632 lines
18 KiB
Objective-C
632 lines
18 KiB
Objective-C
/* RetroArch - A frontend for libretro.
|
|
* Copyright (C) 2013-2014 - Jason Fetters
|
|
* Copyright (C) 2011-2015 - Daniel De Matteis
|
|
*
|
|
* RetroArch is free software: you can redistribute it and/or modify it under the terms
|
|
* of the GNU General Public License as published by the Free Software Found-
|
|
* ation, either version 3 of the License, or (at your option) any later version.
|
|
*
|
|
* RetroArch is distributed in the hope that it will be useful, but WITHOUT ANY WARRANTY;
|
|
* without even the implied warranty of MERCHANTABILITY or FITNESS FOR A PARTICULAR
|
|
* PURPOSE. See the GNU General Public License for more details.
|
|
*
|
|
* You should have received a copy of the GNU General Public License along with RetroArch.
|
|
* If not, see <http://www.gnu.org/licenses/>.
|
|
*/
|
|
|
|
#import <AvailabilityMacros.h>
|
|
#include <sys/stat.h>
|
|
#include "cocoa_common.h"
|
|
#ifdef HAVE_COCOA
|
|
#include "../ui_cocoa.h"
|
|
#endif
|
|
#include "../../../verbosity.h"
|
|
|
|
/* Define compatibility symbols and categories. */
|
|
|
|
#ifdef HAVE_AVFOUNDATION
|
|
#include <AVFoundation/AVCaptureSession.h>
|
|
#include <AVFoundation/AVCaptureDevice.h>
|
|
#include <AVFoundation/AVCaptureOutput.h>
|
|
#include <AVFoundation/AVCaptureInput.h>
|
|
#include <AVFoundation/AVMediaFormat.h>
|
|
#ifdef HAVE_OPENGLES
|
|
#include <CoreVideo/CVOpenGLESTextureCache.h>
|
|
#else
|
|
#include <CoreVideo/CVOpenGLTexture.h>
|
|
#endif
|
|
#endif
|
|
|
|
#include "../../../location/location_driver.h"
|
|
#include "../../../camera/camera_driver.h"
|
|
|
|
static CocoaView* g_instance;
|
|
|
|
#if defined(HAVE_COCOA)
|
|
void *nsview_get_ptr(void)
|
|
{
|
|
return g_instance;
|
|
}
|
|
#endif
|
|
|
|
/* forward declarations */
|
|
void cocoagl_gfx_ctx_update(void);
|
|
void *glkitview_init(void);
|
|
|
|
@implementation CocoaView
|
|
+ (CocoaView*)get
|
|
{
|
|
if (!g_instance)
|
|
g_instance = [CocoaView new];
|
|
|
|
return g_instance;
|
|
}
|
|
|
|
- (id)init
|
|
{
|
|
self = [super init];
|
|
|
|
#if defined(HAVE_COCOA)
|
|
[self setAutoresizingMask:NSViewWidthSizable | NSViewHeightSizable];
|
|
ui_window_cocoa_t cocoa_view;
|
|
cocoa_view.data = (CocoaView*)self;
|
|
|
|
[self registerForDraggedTypes:[NSArray arrayWithObjects:NSColorPboardType, NSFilenamesPboardType, nil]];
|
|
#elif defined(HAVE_COCOATOUCH)
|
|
self.view = (__bridge GLKView*)glkitview_init();
|
|
|
|
[[NSNotificationCenter defaultCenter] addObserver:self selector:@selector(showPauseIndicator) name:UIApplicationWillEnterForegroundNotification object:nil];
|
|
#endif
|
|
|
|
return self;
|
|
}
|
|
|
|
#if defined(HAVE_COCOA)
|
|
- (void)setFrame:(NSRect)frameRect
|
|
{
|
|
[super setFrame:frameRect];
|
|
|
|
cocoagl_gfx_ctx_update();
|
|
}
|
|
|
|
/* Stop the annoying sound when pressing a key. */
|
|
- (BOOL)acceptsFirstResponder
|
|
{
|
|
return YES;
|
|
}
|
|
|
|
- (BOOL)isFlipped
|
|
{
|
|
return YES;
|
|
}
|
|
|
|
- (void)keyDown:(NSEvent*)theEvent
|
|
{
|
|
}
|
|
|
|
- (NSDragOperation)draggingEntered:(id <NSDraggingInfo>)sender
|
|
{
|
|
NSDragOperation sourceDragMask = [sender draggingSourceOperationMask];
|
|
NSPasteboard *pboard = [sender draggingPasteboard];
|
|
|
|
if ( [[pboard types] containsObject:NSFilenamesPboardType] )
|
|
{
|
|
if (sourceDragMask & NSDragOperationCopy)
|
|
return NSDragOperationCopy;
|
|
}
|
|
|
|
return NSDragOperationNone;
|
|
}
|
|
|
|
- (BOOL)performDragOperation:(id<NSDraggingInfo>)sender
|
|
{
|
|
NSPasteboard *pboard = [sender draggingPasteboard];
|
|
|
|
if ( [[pboard types] containsObject:NSURLPboardType])
|
|
{
|
|
NSURL *fileURL = [NSURL URLFromPasteboard:pboard];
|
|
NSString *s = [fileURL path];
|
|
if (s != nil)
|
|
{
|
|
RARCH_LOG("Drop name is: %s\n", [s UTF8String]);
|
|
}
|
|
}
|
|
return YES;
|
|
}
|
|
|
|
- (void)draggingExited:(id <NSDraggingInfo>)sender
|
|
{
|
|
[self setNeedsDisplay: YES];
|
|
}
|
|
|
|
#elif defined(HAVE_COCOATOUCH)
|
|
- (void)viewDidAppear:(BOOL)animated
|
|
{
|
|
/* Pause Menus. */
|
|
[self showPauseIndicator];
|
|
}
|
|
|
|
- (void)showPauseIndicator
|
|
{
|
|
g_pause_indicator_view.alpha = 1.0f;
|
|
[NSObject cancelPreviousPerformRequestsWithTarget:g_instance];
|
|
[g_instance performSelector:@selector(hidePauseButton) withObject:g_instance afterDelay:3.0f];
|
|
}
|
|
|
|
- (void)viewWillLayoutSubviews
|
|
{
|
|
float width = 0.0f, height = 0.0f, tenpctw, tenpcth;
|
|
RAScreen *screen = (__bridge RAScreen*)get_chosen_screen();
|
|
UIInterfaceOrientation orientation = self.interfaceOrientation;
|
|
CGRect screenSize = [screen bounds];
|
|
SEL selector = NSSelectorFromString(BOXSTRING("coordinateSpace"));
|
|
|
|
if ([screen respondsToSelector:selector])
|
|
{
|
|
screenSize = [[screen coordinateSpace] bounds];
|
|
width = CGRectGetWidth(screenSize);
|
|
height = CGRectGetHeight(screenSize);
|
|
}
|
|
else
|
|
{
|
|
width = ((int)orientation < 3) ? CGRectGetWidth(screenSize) : CGRectGetHeight(screenSize);
|
|
height = ((int)orientation < 3) ? CGRectGetHeight(screenSize) : CGRectGetWidth(screenSize);
|
|
}
|
|
|
|
tenpctw = width / 10.0f;
|
|
tenpcth = height / 10.0f;
|
|
|
|
g_pause_indicator_view.frame = CGRectMake(tenpctw * 4.0f, 0.0f, tenpctw * 2.0f, tenpcth);
|
|
[g_pause_indicator_view viewWithTag:1].frame = CGRectMake(0, 0, tenpctw * 2.0f, tenpcth);
|
|
}
|
|
|
|
#define ALMOST_INVISIBLE (.021f)
|
|
|
|
- (void)hidePauseButton
|
|
{
|
|
[UIView animateWithDuration:0.2
|
|
animations:^{ g_pause_indicator_view.alpha = ALMOST_INVISIBLE; }
|
|
completion:^(BOOL finished) { }
|
|
];
|
|
}
|
|
|
|
/* NOTE: This version runs on iOS6+. */
|
|
- (NSUInteger)supportedInterfaceOrientations
|
|
{
|
|
return (NSUInteger)apple_frontend_settings.orientation_flags;
|
|
}
|
|
|
|
/* NOTE: This version runs on iOS2-iOS5, but not iOS6+. */
|
|
- (BOOL)shouldAutorotateToInterfaceOrientation:(UIInterfaceOrientation)interfaceOrientation
|
|
{
|
|
switch (interfaceOrientation)
|
|
{
|
|
case UIInterfaceOrientationPortrait:
|
|
return (apple_frontend_settings.orientation_flags & UIInterfaceOrientationMaskPortrait);
|
|
case UIInterfaceOrientationPortraitUpsideDown:
|
|
return (apple_frontend_settings.orientation_flags & UIInterfaceOrientationMaskPortraitUpsideDown);
|
|
case UIInterfaceOrientationLandscapeLeft:
|
|
return (apple_frontend_settings.orientation_flags & UIInterfaceOrientationMaskLandscapeLeft);
|
|
case UIInterfaceOrientationLandscapeRight:
|
|
return (apple_frontend_settings.orientation_flags & UIInterfaceOrientationMaskLandscapeRight);
|
|
|
|
default:
|
|
return (apple_frontend_settings.orientation_flags & UIInterfaceOrientationMaskAll);
|
|
}
|
|
|
|
return YES;
|
|
}
|
|
#endif
|
|
|
|
#ifdef HAVE_AVFOUNDATION
|
|
#include "../../gfx/common/gl_common.h"
|
|
|
|
#ifndef GL_BGRA
|
|
#define GL_BGRA 0x80E1
|
|
#endif
|
|
|
|
#ifdef HAVE_OPENGLES
|
|
#define RCVOpenGLTextureCacheCreateTextureFromImage CVOpenGLESTextureCacheCreateTextureFromImage
|
|
#define RCVOpenGLTextureGetName CVOpenGLESTextureGetName
|
|
#define RCVOpenGLTextureCacheFlush CVOpenGLESTextureCacheFlush
|
|
#define RCVOpenGLTextureCacheCreate CVOpenGLESTextureCacheCreate
|
|
#define RCVOpenGLTextureRef CVOpenGLESTextureRef
|
|
#define RCVOpenGLTextureCacheRef CVOpenGLESTextureCacheRef
|
|
#if COREVIDEO_USE_EAGLCONTEXT_CLASS_IN_API
|
|
#define RCVOpenGLGetCurrentContext() (CVEAGLContext)(g_context)
|
|
#else
|
|
#define RCVOpenGLGetCurrentContext() (__bridge void *)(g_context)
|
|
#endif
|
|
#else
|
|
#define RCVOpenGLTextureCacheCreateTextureFromImage CVOpenGLTextureCacheCreateTextureFromImage
|
|
#define RCVOpenGLTextureGetName CVOpenGLTextureGetName
|
|
#define RCVOpenGLTextureCacheFlush CVOpenGLTextureCacheFlush
|
|
#define RCVOpenGLTextureCacheCreate CVOpenGLTextureCacheCreate
|
|
#define RCVOpenGLTextureRef CVOpenGLTextureRef
|
|
#define RCVOpenGLTextureCacheRef CVOpenGLTextureCacheRef
|
|
#define RCVOpenGLGetCurrentContext() CGLGetCurrentContext(), CGLGetPixelFormat(CGLGetCurrentContext())
|
|
#endif
|
|
|
|
static AVCaptureSession *_session;
|
|
static NSString *_sessionPreset;
|
|
RCVOpenGLTextureCacheRef textureCache;
|
|
GLuint outputTexture;
|
|
static bool newFrame = false;
|
|
|
|
static void event_process_camera_frame(void *pbuf_ptr)
|
|
{
|
|
CVReturn ret;
|
|
RCVOpenGLTextureRef renderTexture;
|
|
CVPixelBufferRef pixelBuffer = (CVPixelBufferRef)pbuf_ptr;
|
|
size_t width = CVPixelBufferGetWidth(pixelBuffer);
|
|
size_t height = CVPixelBufferGetHeight(pixelBuffer);
|
|
|
|
CVPixelBufferLockBaseAddress(pixelBuffer, 0);
|
|
|
|
(void)width;
|
|
(void)height;
|
|
|
|
/*TODO - rewrite all this.
|
|
*
|
|
* create a texture from our render target.
|
|
* textureCache will be what you previously
|
|
* made with RCVOpenGLTextureCacheCreate.
|
|
*/
|
|
#ifdef HAVE_OPENGLES
|
|
ret = RCVOpenGLTextureCacheCreateTextureFromImage(kCFAllocatorDefault,
|
|
textureCache, pixelBuffer, NULL, GL_TEXTURE_2D,
|
|
GL_RGBA, (GLsizei)width, (GLsizei)height,
|
|
GL_BGRA, GL_UNSIGNED_BYTE, 0, &renderTexture);
|
|
#else
|
|
ret = RCVOpenGLTextureCacheCreateTextureFromImage(kCFAllocatorDefault,
|
|
textureCache, pixelBuffer, 0, &renderTexture);
|
|
#endif
|
|
|
|
if (!renderTexture || ret)
|
|
{
|
|
RARCH_ERR("[apple_camera]: RCVOpenGLTextureCacheCreateTextureFromImage failed.\n");
|
|
return;
|
|
}
|
|
|
|
outputTexture = RCVOpenGLTextureGetName(renderTexture);
|
|
glBindTexture(GL_TEXTURE_2D, outputTexture);
|
|
glTexParameteri(GL_TEXTURE_2D, GL_TEXTURE_MIN_FILTER, GL_LINEAR);
|
|
glTexParameteri(GL_TEXTURE_2D, GL_TEXTURE_MAG_FILTER, GL_LINEAR);
|
|
glTexParameteri(GL_TEXTURE_2D, GL_TEXTURE_WRAP_S, GL_CLAMP_TO_EDGE);
|
|
glTexParameteri(GL_TEXTURE_2D, GL_TEXTURE_WRAP_T, GL_CLAMP_TO_EDGE);
|
|
|
|
CVPixelBufferUnlockBaseAddress(pixelBuffer, 0);
|
|
[[NSNotificationCenter defaultCenter] postNotificationName:@"NewCameraTextureReady" object:nil];
|
|
newFrame = true;
|
|
|
|
glBindTexture(GL_TEXTURE_2D, 0);
|
|
|
|
RCVOpenGLTextureCacheFlush(textureCache, 0);
|
|
|
|
CFRelease(renderTexture);
|
|
CFRelease(pixelBuffer);
|
|
|
|
pixelBuffer = 0;
|
|
}
|
|
|
|
- (void)captureOutput:(AVCaptureOutput *)captureOutput
|
|
didOutputSampleBuffer:(CMSampleBufferRef)sampleBuffer
|
|
fromConnection:(AVCaptureConnection *)connection
|
|
{
|
|
/* TODO: Don't post if event queue is full */
|
|
CVPixelBufferRef pixelBuffer = (CVPixelBufferRef)CVPixelBufferRetain(CMSampleBufferGetImageBuffer(sampleBuffer));
|
|
event_process_camera_frame(pixelBuffer);
|
|
}
|
|
|
|
/* TODO - add void param to onCameraInit so we can pass g_context. */
|
|
- (void) onCameraInit
|
|
{
|
|
NSError *error;
|
|
AVCaptureVideoDataOutput * dataOutput;
|
|
AVCaptureDeviceInput *input;
|
|
AVCaptureDevice *videoDevice;
|
|
|
|
CVReturn ret = RCVOpenGLTextureCacheCreate(kCFAllocatorDefault, NULL,
|
|
RCVOpenGLGetCurrentContext(), NULL, &textureCache);
|
|
(void)ret;
|
|
|
|
/* Setup Capture Session. */
|
|
_session = [[AVCaptureSession alloc] init];
|
|
[_session beginConfiguration];
|
|
|
|
/* TODO: dehardcode this based on device capabilities */
|
|
_sessionPreset = AVCaptureSessionPreset640x480;
|
|
|
|
/* Set preset session size. */
|
|
[_session setSessionPreset:_sessionPreset];
|
|
|
|
/* Creata a video device and input from that Device. Add the input to the capture session. */
|
|
videoDevice = (AVCaptureDevice*)[AVCaptureDevice defaultDeviceWithMediaType:AVMediaTypeVideo];
|
|
if (videoDevice == nil)
|
|
assert(0);
|
|
|
|
/* Add the device to the session. */
|
|
input = (AVCaptureDeviceInput*)[AVCaptureDeviceInput deviceInputWithDevice:videoDevice error:&error];
|
|
|
|
if (error)
|
|
{
|
|
RARCH_ERR("video device input %s\n", error.localizedDescription.UTF8String);
|
|
assert(0);
|
|
}
|
|
|
|
[_session addInput:input];
|
|
|
|
/* Create the output for the capture session. */
|
|
dataOutput = (AVCaptureVideoDataOutput*)[[AVCaptureVideoDataOutput alloc] init];
|
|
[dataOutput setAlwaysDiscardsLateVideoFrames:NO]; /* Probably want to set this to NO when recording. */
|
|
|
|
[dataOutput setVideoSettings:[NSDictionary dictionaryWithObject:[NSNumber numberWithInt:kCVPixelFormatType_32BGRA] forKey:(id)kCVPixelBufferPixelFormatTypeKey]];
|
|
|
|
/* Set dispatch to be on the main thread so OpenGL can do things with the data. */
|
|
[dataOutput setSampleBufferDelegate:self queue:dispatch_get_main_queue()];
|
|
|
|
[_session addOutput:dataOutput];
|
|
[_session commitConfiguration];
|
|
}
|
|
|
|
- (void) onCameraStart
|
|
{
|
|
[_session startRunning];
|
|
}
|
|
|
|
- (void) onCameraStop
|
|
{
|
|
[_session stopRunning];
|
|
}
|
|
|
|
- (void) onCameraFree
|
|
{
|
|
RCVOpenGLTextureCacheFlush(textureCache, 0);
|
|
CFRelease(textureCache);
|
|
}
|
|
#endif
|
|
|
|
#ifdef HAVE_CORELOCATION
|
|
#include <CoreLocation/CoreLocation.h>
|
|
|
|
static CLLocationManager *locationManager;
|
|
static bool locationChanged;
|
|
static CLLocationDegrees currentLatitude;
|
|
static CLLocationDegrees currentLongitude;
|
|
static CLLocationAccuracy currentHorizontalAccuracy;
|
|
static CLLocationAccuracy currentVerticalAccuracy;
|
|
|
|
- (bool)onLocationHasChanged
|
|
{
|
|
bool hasChanged = locationChanged;
|
|
|
|
if (hasChanged)
|
|
locationChanged = false;
|
|
|
|
return hasChanged;
|
|
}
|
|
|
|
- (void)locationManager:(CLLocationManager *)manager didUpdateToLocation:(CLLocation *)newLocation fromLocation:(CLLocation *)oldLocation
|
|
{
|
|
locationChanged = true;
|
|
currentLatitude = newLocation.coordinate.latitude;
|
|
currentLongitude = newLocation.coordinate.longitude;
|
|
currentHorizontalAccuracy = newLocation.horizontalAccuracy;
|
|
currentVerticalAccuracy = newLocation.verticalAccuracy;
|
|
RARCH_LOG("didUpdateToLocation - latitude %f, longitude %f\n", (float)currentLatitude, (float)currentLongitude);
|
|
}
|
|
|
|
- (void)locationManager:(CLLocationManager *)manager didUpdateLocations:(NSArray *)locations
|
|
{
|
|
CLLocation *location = (CLLocation*)[locations objectAtIndex:([locations count] - 1)];
|
|
|
|
locationChanged = true;
|
|
currentLatitude = [location coordinate].latitude;
|
|
currentLongitude = [location coordinate].longitude;
|
|
currentHorizontalAccuracy = location.horizontalAccuracy;
|
|
currentVerticalAccuracy = location.verticalAccuracy;
|
|
RARCH_LOG("didUpdateLocations - latitude %f, longitude %f\n", (float)currentLatitude, (float)currentLongitude);
|
|
}
|
|
|
|
- (void)locationManager:(CLLocationManager *)manager didFailWithError:(NSError *)error
|
|
{
|
|
RARCH_LOG("didFailWithError - %s\n", [[error localizedDescription] UTF8String]);
|
|
}
|
|
|
|
- (void)locationManagerDidPauseLocationUpdates:(CLLocationManager *)manager
|
|
{
|
|
RARCH_LOG("didPauseLocationUpdates\n");
|
|
}
|
|
|
|
- (void)locationManagerDidResumeLocationUpdates:(CLLocationManager *)manager
|
|
{
|
|
RARCH_LOG("didResumeLocationUpdates\n");
|
|
}
|
|
|
|
- (void)onLocationInit
|
|
{
|
|
/* Create the location manager
|
|
* if this object does not already have one.
|
|
*/
|
|
|
|
if (locationManager == nil)
|
|
locationManager = [[CLLocationManager alloc] init];
|
|
locationManager.delegate = self;
|
|
locationManager.desiredAccuracy = kCLLocationAccuracyBest;
|
|
locationManager.distanceFilter = kCLDistanceFilterNone;
|
|
[locationManager startUpdatingLocation];
|
|
}
|
|
#endif
|
|
|
|
@end
|
|
|
|
#ifdef HAVE_AVFOUNDATION
|
|
typedef struct apple_camera
|
|
{
|
|
void *empty;
|
|
} applecamera_t;
|
|
|
|
static void *apple_camera_init(const char *device, uint64_t caps, unsigned width, unsigned height)
|
|
{
|
|
applecamera_t *applecamera;
|
|
|
|
if ((caps & (UINT64_C(1) << RETRO_CAMERA_BUFFER_OPENGL_TEXTURE)) == 0)
|
|
{
|
|
RARCH_ERR("applecamera returns OpenGL texture.\n");
|
|
return NULL;
|
|
}
|
|
|
|
applecamera = (applecamera_t*)calloc(1, sizeof(applecamera_t));
|
|
if (!applecamera)
|
|
return NULL;
|
|
|
|
[[CocoaView get] onCameraInit];
|
|
|
|
return applecamera;
|
|
}
|
|
|
|
static void apple_camera_free(void *data)
|
|
{
|
|
applecamera_t *applecamera = (applecamera_t*)data;
|
|
|
|
[[CocoaView get] onCameraFree];
|
|
|
|
if (applecamera)
|
|
free(applecamera);
|
|
applecamera = NULL;
|
|
}
|
|
|
|
static bool apple_camera_start(void *data)
|
|
{
|
|
(void)data;
|
|
|
|
[[CocoaView get] onCameraStart];
|
|
|
|
return true;
|
|
}
|
|
|
|
static void apple_camera_stop(void *data)
|
|
{
|
|
[[CocoaView get] onCameraStop];
|
|
}
|
|
|
|
static bool apple_camera_poll(void *data, retro_camera_frame_raw_framebuffer_t frame_raw_cb,
|
|
retro_camera_frame_opengl_texture_t frame_gl_cb)
|
|
{
|
|
|
|
(void)data;
|
|
(void)frame_raw_cb;
|
|
|
|
if (frame_gl_cb && newFrame)
|
|
{
|
|
/* FIXME: Identity for now.
|
|
* Use proper texture matrix as returned by iOS Camera (if at all?). */
|
|
static const float affine[] = {
|
|
1.0f, 0.0f, 0.0f,
|
|
0.0f, 1.0f, 0.0f,
|
|
0.0f, 0.0f, 1.0f
|
|
};
|
|
|
|
frame_gl_cb(outputTexture, GL_TEXTURE_2D, affine);
|
|
newFrame = false;
|
|
}
|
|
|
|
return true;
|
|
}
|
|
|
|
camera_driver_t camera_avfoundation = {
|
|
apple_camera_init,
|
|
apple_camera_free,
|
|
apple_camera_start,
|
|
apple_camera_stop,
|
|
apple_camera_poll,
|
|
"avfoundation",
|
|
};
|
|
#endif
|
|
|
|
#ifdef HAVE_CORELOCATION
|
|
typedef struct apple_location
|
|
{
|
|
void *empty;
|
|
} applelocation_t;
|
|
|
|
static void *apple_location_init(void)
|
|
{
|
|
applelocation_t *applelocation = (applelocation_t*)calloc(1, sizeof(applelocation_t));
|
|
if (!applelocation)
|
|
return NULL;
|
|
|
|
[[CocoaView get] onLocationInit];
|
|
|
|
return applelocation;
|
|
}
|
|
|
|
static void apple_location_set_interval(void *data, unsigned interval_update_ms, unsigned interval_distance)
|
|
{
|
|
(void)data;
|
|
|
|
locationManager.distanceFilter = interval_distance ? interval_distance : kCLDistanceFilterNone;
|
|
}
|
|
|
|
static void apple_location_free(void *data)
|
|
{
|
|
applelocation_t *applelocation = (applelocation_t*)data;
|
|
|
|
/* TODO - free location manager? */
|
|
|
|
if (applelocation)
|
|
free(applelocation);
|
|
applelocation = NULL;
|
|
}
|
|
|
|
static bool apple_location_start(void *data)
|
|
{
|
|
(void)data;
|
|
|
|
[locationManager startUpdatingLocation];
|
|
return true;
|
|
}
|
|
|
|
static void apple_location_stop(void *data)
|
|
{
|
|
(void)data;
|
|
|
|
[locationManager stopUpdatingLocation];
|
|
}
|
|
|
|
static bool apple_location_get_position(void *data, double *lat, double *lon, double *horiz_accuracy,
|
|
double *vert_accuracy)
|
|
{
|
|
(void)data;
|
|
|
|
bool ret = [[CocoaView get] onLocationHasChanged];
|
|
|
|
if (!ret)
|
|
goto fail;
|
|
|
|
*lat = currentLatitude;
|
|
*lon = currentLongitude;
|
|
*horiz_accuracy = currentHorizontalAccuracy;
|
|
*vert_accuracy = currentVerticalAccuracy;
|
|
return true;
|
|
|
|
fail:
|
|
*lat = 0.0;
|
|
*lon = 0.0;
|
|
*horiz_accuracy = 0.0;
|
|
*vert_accuracy = 0.0;
|
|
return false;
|
|
}
|
|
|
|
location_driver_t location_corelocation = {
|
|
apple_location_init,
|
|
apple_location_free,
|
|
apple_location_start,
|
|
apple_location_stop,
|
|
apple_location_get_position,
|
|
apple_location_set_interval,
|
|
"corelocation",
|
|
};
|
|
#endif
|
|
|