Like the image showing, the element 'Ask ChatGPT' attaches below siri search suggestion , I guess SiriKit or Spotlight can implement it.
But i searched a lot, no one introduct the detail technology.
Do anyone ever implement it?
Please help me. Thanks.
ResearchKit
RSS for tagEnable your iOS app to become a powerful tool for medical research with the ResearchKit open source framework.
Posts under ResearchKit tag
5 Posts
Sort by:
Post
Replies
Boosts
Views
Activity
Hello!
I am a student at a tertiary Epilepsy Center. Our team is building an open source seizure alert app for the apple watch that will undergo a clinical trial. A well conducted clinical trial is crucial for convincing insurance to pay for apple watches, which correlates with tech adoption. This imposes some restrictions on data collection.
Our use case requires that the app run continuously as seizures are not predictable. The documentation on extended runtime sessions recommended the smart alarm mode which lasts for 30 minutes. Individuals would not use an alarm system that only works 30 minutes at a time. Is there any way we could deploy the alarm continuously?
The requirements for the background execution are two fold:
During model development phase the app would need to continuously collect data from accelerometer, gyro-meter, magnetometer, blood oxygen and EKG. I understand these are very demanding tasks so even one sensor would be good to start with. The app would then upload this to a server over the web (perhaps using CloudKit). If continuous data collection is not possible then what is the longest guaranteed latency between samples? What amount of CPU usage would be permissible and how would it impact sampling rates?
During deployment the app would need to run an ML model in the background continuously. We have chosen a light-weight model[1] that has been optimized for mobile devices. It has ~ 1.3 million parameters. What would be the maximum guaranteed latency for running a ML model in the background?
Please let me know if you need any more information. Thank you.
Reference:
Lightweight Transformers for Human Activity Recognition on Mobile Devices, Sannara EK, François Portet, Philippe Lalanda. arXiv:2209.11750
Will Apple Vision Pro support medical imaging diagnostic software such as Horos or Osirix. I am really curious if radiologist will be able to view and manipulate. interpret exams with this headset with ease and use built in microphone for voice recognition/dictation while using hands to manipulate the imaging and also simultaneously being able to view report in field of view. This could really unlock some major potential for interpreting at least CT, Ultrasound and MRI exams. I don't think the resolution will be high enough to interpret diagnostic x-rays and definitely not mammograms due to MQSA regulations and physicist inspections requiring more detail and information. However, I want to be at the forefront of bringing in Vision Pro headsets in the medical imaging space with rel utilization in clinical practices. It may also be beneficial for patients who are curious to see their medical imaging or even as headsets to use while undergoing medical imaging outpatient biopsy procedures for ******/etc to help put them at ease during the procedures. This could really provide patient satisfaction and I think we are scratching the surface on a world of possibilities in healthcare with devices like this. Upon utilixation, I would like to creat presentation series and share information with my Radiology colleagues at national/international meetings.
Hello, I work in ophthalmology at Stanford and I am hoping to develop a tool for monitoring vision health using Apple Vision Pro. However, I am worried that my goal will not be possible due to Apple privacy concerns. I want to develop an app that can gauge eye performance, but to do so I would need precise details about the users eye movements and eye position. Is this feasible? Also, I am eager to learn more about the specifics of the eye-tracking set up, like the accuracy and the sampling rate. If you have any useful information or suggestions, I would really appreciate it, thank you!
Hello all, this could surely be a newbie mistake but I'm unable to adopt this protocol to my class object I've created in a new project.
The Error: "Cannot find type 'OCKSurveyTaskViewControllerDelegate' in scope"
import CareKit
import CareKitUI
import CareKitStore
import ResearchKit
import UIKit
import os.log
final class CareFeedViewController: OCKDailyPageViewController, OCKTaskSurveyViewControllerDelegate {
}
However, I've been able to locate the 'OCKSurveyViewController.swift" file in the CareKit framework from SPM which has the protocol defined there, but I'm still unable to utilize it. Was the protocol only for demonstration purposes and not for use outside of the WWDC21 - CareKit Code Along?