
Geschlossen
Veröffentlicht
Bezahlt bei Lieferung
Project Overview: We are looking for an experienced iOS Engineer to build the core architecture of a Real-Time Computer Vision Application. The app requires a high-performance video pipeline that captures frames, runs local inference (CoreML), and selectively sends data to a Cloud LLM API. This is a pure engineering challenge. We need clean, modular code that handles concurrency, thermal management, and hardware switching without draining the battery instantly. Phase 1 Scope (MVP): You will build a functional "Skeleton App" focusing on performance and stability using the internal camera. Key Technical Requirements: Advanced Camera & Audio Pipeline (AVFoundation): Implement a robust AVCaptureSession to handle video buffers (CMSampleBuffer). CRITICAL: The architecture must support Hot-Plugging External Devices. Requirement: When an external USB device is connected, the app must automatically switch BOTH the Video Feed AND the Audio Input to the external source. Use Case: The device (e.g., smart glasses or webcam) will act as the primary sensor for both vision and sound. Note: Use standard [login to view URL] (type .external). Hybrid AI Logic (The Core Engine): Local: Run a lightweight CoreML model for real-time object detection (offline). Cloud: Integrate a REST/WebSocket API (Gemini/GPT-4o) to analyze frames every few seconds. Smart Switching: Use NWPathMonitor to fallback to "Local-Only" mode if the connection drops. Thermal & Performance Optimization: Implement Adaptive Throttling: Monitor ProcessInfo.processInfo.thermalState. Logic: If the device gets hot (.serious state), automatically reduce the Cloud API request rate while keeping the local model active. Background Execution: Ensure the app’s audio engine (TTS) remains active even when the screen is locked or the app is in the background. Required Tech Stack: Swift 5+, UIKit or SwiftUI. AVFoundation (Deep understanding of capture sessions, inputs, and audio routing). CoreML & Vision Framework. Concurrency (GCD / Swift Actors). Deliverables: A clean Xcode project (GitHub repo). The "Hybrid Engine" working smoothly on a physical device. Code documentation explaining the modular camera/audio setup.
Projekt-ID: 40224123
62 Vorschläge
Remote Projekt
Aktiv vor 17 Tagen
Legen Sie Ihr Budget und Ihren Zeitrahmen fest
Für Ihre Arbeit bezahlt werden
Skizzieren Sie Ihren Vorschlag
Sie können sich kostenlos anmelden und auf Aufträge bieten
62 Freelancer bieten im Durchschnitt $171 USD für diesen Auftrag

I am confident that my skills in Mobile App Development, Swift, Artificial Intelligence, Image Processing, and iOS Development make me a great match for the "iOS Engineer: Real-Time Camera Processing & Hybrid AI Engine" project. I am eager to tackle the pure engineering challenge of building a high-performance video pipeline with robust camera/audio handling, hybrid AI logic, and thermal optimization. I am flexible with the budget and my priority is to deliver within your requirements. Please review my 15 years of experience on my profile. Let's discuss further and I am ready to showcase my commitment by starting work immediately. Looking forward to your response.
$175 USD in 7 Tagen
6,4
6,4

Hi, I am excited about the opportunity to work on your Real-Time Computer Vision Application as an iOS Engineer. With extensive experience in developing high-performance applications using Swift and AVFoundation, I am confident in my ability to build a robust architecture that meets your project's requirements. My expertise in implementing AVCaptureSessions allows me to handle video buffers effectively, ensuring seamless integration of both internal and external camera feeds. I understand the importance of modular and clean code, especially for handling complexities like concurrency and thermal management. My approach includes implementing adaptive throttling to maintain performance and battery efficiency. I am ready to dive into developing the Skeleton App, focusing on ensuring stability and performance while supporting all specified functionalities. We can aim to deliver the foundational version within the next 30 days. Best regards,
$250 USD in 30 Tagen
5,7
5,7

✋ Hi There!!! ✋ The Goal of the project:- Build a high-performance iOS framework that captures real-time camera frames, integrates with a cloud-based LLM, and delivers dynamic, low-latency speech responses. I carefully read your full description and understand you need advanced AVFoundation camera implementation prioritizing the Ultra-Wide lens, efficient frame capture and buffering, seamless integration with the Google Gemini Flash API, dynamic instruction handling, and low-latency speech output on iOS 15+ devices. With 9+ years experience as a full stack developer and mobile engineer, I have deep expertise in Swift, AVFoundation, Vision, and real-time API integrations. • Implement custom AVFoundation session for automatic Ultra-Wide lens selection across iPhone 11–17 • Build efficient frame capture pipeline with proper CMSampleBuffer management for long-term sessions • Integrate Google AI API with dynamic system instructions and AVSpeechSynthesizer for real-time audio I will also handle UI refinement, testing, memory optimization, deployment setup, and full source code delivery at project completion. I have previously developed real-time vision and AI-assisted mobile applications with modular architectures and low-latency processing. Looking forward to chat with you for make a deal Best Regards Elisha Mariam!
$111 USD in 11 Tagen
5,1
5,1

As an experienced iOS engineer with a specialization in Artificial Intelligence (AI), I believe I'm uniquely qualified to take on your Real-Time Computer Vision Application challenge. With proficiency in Swift, AVFoundation, and CoreML & Vision frameworks, I leverage my deep understanding of capture sessions, inputs and audio routing to meet the primary objectives of implementing an AVCaptureSession, handling video buffers and hot-plugging external devices, while also creating a robust and modular system that efficiently processes data without compromising the device's battery life. Lastly, my broad skills including Web Development with emphasis on AWS and Google Cloud Services enable me to seamlessly integrate external APIs, deliver a clean Xcode project (via GitHub) and provide comprehensive documentation for future reference. By choosing me for your project, you will gain not just highly skilled expertise, but also a tech enthusiast who thrives on not just meeting but exceeding client expectations.
$30 USD in 7 Tagen
5,7
5,7

As an accomplished Full Stack Web & Mobile App Developer with a strong focus on iOS Development and over a decade of hands-on experience, I am confident in my ability to deliver the top-tier performance you are seeking for your Real-Time Computer Vision Application. My extensive skills in technologies like Swift, AVFoundation, CoreML, Vision Framework, and more align perfectly with your project's technical requirements. My deep understanding of capture sessions, inputs, and audio routing within AVFoundation will ensure we can successfully implement robust AVCaptureSession to handle video buffers while also supporting Hot-Plugging External Cameras and switching seamlessly between them to maintain a high level of user experience and performance. My proficiency in GCD and Swift Actors will contribute to developing concurrency-aware code that handles thermal management, hardware switching without excessively draining battery life.
$60 USD in 5 Tagen
3,8
3,8

Hello, I’m a Machine Learning Engineer with 8+ years in production CV and AI systems. I’ve built AVFoundation pipelines streaming frames to cloud vision models, handling lens discovery via AVCaptureDevice discovery sessions and optimizing CMSampleBuffer release with dedicated background queues to prevent leaks. I’ve also maintained low power camera sessions using proper AVAudioSession and background modes. I can deliver a stable Gemini integrated framework.
$200 USD in 7 Tagen
3,4
3,4

Hi [ClientFirstName], I bring extensive iOS development experience with Swift 6.0, AVFoundation, Vision, and RESTful API integration to architect a high-performance real-time vision pipeline and LLM bridge. I will design a stable, low-latency pipeline that captures frames from the Ultra-Wide lens across supported devices, buffers efficiently, and streams data to the Gemini API without blocking the main thread, while preserving battery and privacy considerations. Given the scope, how would you prioritize lens discovery and dynamic profile loading in the local configuration to balance Field of View with performance and memory constraints across iPhone 11–17? Best regards,
$180 USD in 2 Tagen
3,0
3,0

Hi there, We’re excited to work on iOS Engineer - Real-time Vision Pipeline & LLM API Integration and bring your vision to life. We specialize in high-performance, user-friendly mobile applications and Progressive Web Applications (PWA) with clean UI/UX, smooth functionality, and scalable design. We’d be happy to showcase some of our recent projects and discuss how we can turn your idea into a successful, high-quality mobile app tailored to your needs. You can also view our full portfolio here: https://www.freelancer.com/u/NexifyStudio Looking forward to collaborating! Best regards, Nexify Studio
$30 USD in 7 Tagen
2,9
2,9

Hello, I will develop a high-performance framework for real-time environmental analysis, integrating the device's camera system with a cloud-based Large Language Model (LLM). With extensive experience in iOS development and camera systems, I have successfully built similar applications that leverage AVFoundation and cloud APIs. My focus is on delivering reliable, efficient solutions. Implementation Approach: - Custom AVFoundation Session: Build a session that automatically detects and prioritizes the Ultra-Wide lens for all compatible devices (iPhone 11-17). - Efficient Frame Capture: Implement a capture system to extract frames every 2-3 seconds, ensuring non-blocking API transmission. - AI Integration: Connect the capture pipeline to the Google Gemini Flash API using REST or the Swift SDK. - Dynamic Instruction Architecture: Create a modular system to pull instructions from a local configuration based on user profiles. - Speech Output Integration: Use AVSpeechSynthesizer for low-latency audio conversion of API responses. Key Questions: - What is your preferred method for automatic lens discovery to ensure optimal Field of View? - How do you want to manage memory usage during continuous frame capture sessions? - Given Apple’s privacy constraints, how should we maintain an active camera session in low-power modes? I’m ready to start immediately and look forward to collaborating on this innovative project.
$140 USD in 7 Tagen
3,1
3,1

Hi there!, I’m Robert, a Senior Full-Stack & AI Engineer with over 10 years of experience architecting and delivering SaaS platforms, automation systems, and intelligent applications, excelling in Swift, camera implementation, and AI integration. I developed a real-time environmental analysis system that included a custom camera framework and integrated with a cloud-based AI API, ensuring efficiency in frame capture and processing. My deep technical background aligns perfectly with your project requirements, as I can implement an advanced AVFoundation session, manage memory effectively, and ensure seamless API interactions. I can complete this project perfectly and deliver scalable, production-ready results. I am committed to clean architecture, structured documentation, CI/CD automation, and OWASP-based security practices. All components will adhere to strict performance validation standards and data privacy regulations. Let’s connect to refine your requirements and begin building a solution that exceeds expectations. What specific challenges do you anticipate with the integration of the Google Gemini Flash API and how should we address them?
$250 USD in 30 Tagen
2,4
2,4

Hello , I checked your project, and it looks interesting. This is something we already work on, so the requirements are clear from the start. We mainly work on Mobile App Development, Swift, Artificial Intelligence, Image Processing, iOS Development, RESTful API, API Integration We focus on making things simple, reliable, and actually useful in real life not overcomplicated stuff. Let’s connect in chat and see if we’re a good fit for this. Best Regards, Ali nawaz
$129 USD in 4 Tagen
2,3
2,3

Hi there! Do you need the framework to support seamless switching between lenses in real time if the user moves between indoor and outdoor lighting conditions? Regardless, this is definitely something that I feel confident delivering on, given my past experience. I would love to discuss your project further! Looking forward hearing from you. Kind Regards, Corné
$100 USD in 14 Tagen
2,5
2,5

Hello, I’m Dinesh Kumar With 14+ years of experience across multiple platforms, I’ve helped build numerous startups through dedication and hard work. I’m committed to delivering high quality work that ensures 100% client satisfaction. Your success is my priority, and I focus on building long term relationships based on trust and excellence. Expertise: Web & App Development – React.js, Node.js, JavaScript, PHP, MySQL, WordPress, Magento, CodeIgniter, Shopify, .NET, Flutter, FoxPro Strong knowledge of frameworks, software design, and development methodologies Proven ability to deliver custom, scalable, and reliable solutions for diverse industries I work with clients globally, providing end to end solutions that meet unique project needs while maintaining the highest quality standards.
$140 USD in 7 Tagen
2,3
2,3

Hello, I appreciate the opportunity to apply for your iOS Developer position focused on building a high-performance framework for real-time environmental analysis. I understand that your project requires a stable bridge between the device's camera system and a cloud-based LLM, and I am excited about the challenge. With extensive experience in Swift development and a strong background in AVFoundation and Vision frameworks, I have successfully implemented advanced camera functionalities and AI integrations in previous projects. My familiarity with RESTful APIs and real-time data handling equips me well to meet your technical requirements. To ensure the project's success, I propose the following approach: - Develop a custom AVFoundation session that dynamically detects and prioritizes the Ultra-Wide lens, ensuring optimal field of view. - Implement an efficient frame capture system that extracts frames at regular intervals while maintaining smooth UI performance. - Integrate Google Gemini Flash API using Swift SDK for seamless AI interaction. - Create a modular instruction architecture that dynamically adjusts based on user profiles, along with low-latency speech output via AVSpeechSynthesizer. I am eager to begin this project and confident in my ability to deliver high-quality results on time. I would love to discuss your project further and answer any questions you may have. Thank you for considering my bid!
$140 USD in 7 Tagen
2,2
2,2

I’m confident that I have the skills necessary to bring your vision to fruition. To address your concerns about automatic lens discovery, my strategy would involve utilizing a camera-available-metadata property that comes with iOS 15+. This metadata reveals which lenses are available for capture at any given time, allowing for dynamic prioritization - always selecting the widest Field of View programmatically. Maintaining memory usage would be critical for continuous, long-term frame capture sessions. As such, I'd optimize CMSampleBuffer release by employing ARC (Automatic Reference Counting) coupled with diligent memory management practices. By using weak references where applicable and eliminating retain cycles, we can maintain memory efficiency throughout. Lastly, in regard to Apple's privacy constraints regarding camera session maintenance while the UI is in a low-power state, I propose leveraging the multitasking capabilities of Apple's devices. Using appropriate Background Modes coupled with clever task management should enable us to keep the camera session alive without compromising battery life or deviating from Apple's guidelines regarding background activity. Let's make your real-time environmental analysis framework a reality together!
$140 USD in 7 Tagen
3,2
3,2

Hi, real-time CV apps break when the camera pipeline, concurrency, and device switching aren’t architected cleanly. If hot-plugging external video/audio or thermal throttling isn’t handled properly from day one, performance collapses fast. I’ll build a modular Swift architecture: a robust AVCaptureSession manager supporting automatic switching to .external devices (video + audio), a hybrid engine running CoreML locally with async Cloud API calls, and a coordinator handling NWPathMonitor fallback and adaptive throttling via ProcessInfo.thermalState. Concurrency will be isolated using Swift actors to keep frame processing non-blocking. Background audio/TTS will remain active with correct session configuration. You’ll receive a clean, documented Xcode project with the hybrid engine running smoothly on a physical device. One question: do you already have the CoreML model ready, or should I stub it for MVP testing? Chand
$50 USD in 2 Tagen
1,8
1,8

Hi, how are you doing? I went through your project description and I can help you in your project. your project requirements perfectly match my expertise. We are a team of expert engineers, we have successfully completed 1000+ Projects for multiple regular clients from OMAN, UK, USA, Australia, Canada, France, Germany, Lebanon and many other countries. We are providing our services in following areas: Neural Network/ Natural Language Processing Machine learning/Data Mining Deep Learning and Computer Vision Image Recognition & Artificial Intelligence AI text analysis model and Reinforcement Learning. Omnet++ and Sumo simulation, Python/ MATLAB Asterisks PBX NS3 simulation Linux We'll make sure that your project is done in a perfect way and do our best until you were satisfied. I am confident I can provide you with top-notch materials that will fit your needs.
$140 USD in 7 Tagen
1,4
1,4

Hello There, I totally understand your requirement. This isn’t a basic camera app — it’s a real-time, low-latency vision pipeline that must be stable across multiple hardware generations while streaming intelligently to an LLM. Performance, memory discipline, and session control are critical here. Screening Answers: Lens discovery: Use DiscoverySession filtering by deviceType (.builtInUltraWideCamera first), position, and format support; dynamically select highest FOV activeFormat. Memory strategy: Avoid frame accumulation, process single-frame snapshots, use CVPixelBufferLockBaseAddress carefully, release CMSampleBuffer immediately after conversion, monitor via Instruments (Leaks + Allocations). Privacy/session handling: Use proper background modes only if justified; otherwise maintain active session with idleTimerDisabled and controlled dimming while respecting App Store policies. Let me know when you’re available to discuss this further I’d be happy to walk you through my approach or showcase examples relevant to this project. Looking forward to hearing from you! Best regards, Mulayam
$140 USD in 7 Tagen
1,2
1,2

Hi there, With over 10 years of experience, I just finish reading your requirement and understand well. The Goal of the project :- I am excited to submit my proposal for the development of your real-time environmental analysis framework. You require a robust iOS framework that automatically prioritizes the Ultra-Wide (0.5x) camera, captures frames at controlled 2–3 second intervals, and efficiently sends them to the Google Gemini Flash API with dynamic, profile-based instructions. 1)Selecting the Widest Camera Automatically The app will first check if the Ultra-Wide (0.5x) camera is available. If it is available, it will automatically use it. If not, it will smoothly switch to the standard wide camera. 2) Managing Memory During Long Capture Sessions The app will only process a frame every 2–3 seconds instead of every single frame. Each frame will be used immediately and then cleared from memory. All processing will happen in the background so the app stays smooth and stable. 3) Keeping the Camera Active in Low-Power / Dimmed State The app will stay active on screen so the camera keeps running. The screen can appear dimmed to save power, but the app will not go into the background. If the app is minimized, the camera will pause and safely resume when reopened. I would be happy to discuss timeline, Looking forward to response Best regards, Anupma
$140 USD in 7 Tagen
1,1
1,1

Hello, I'm a Swift developer with over 10 years of experience in mobile app development. I specialize in building efficient frameworks that integrate seamlessly with APIs. We'll discuss the details in a chat. I can create a reliable real-time environmental analysis system using AVFoundation to prioritize the Ultra-Wide lens automatically. This ensures you get the best Field of View possible. For frame capture, I propose a system that extracts images every 2-3 seconds while maintaining smooth performance by keeping the main thread unblocked. Option A: Use a simple REST API to transfer captured frames to Google Gemini Flash, ensuring efficiency. Option B: Implement the Google AI SDK for a more integrated solution, enhancing overall performance. Which option do you prefer? For speech output, integrating AVSpeechSynthesizer will allow for real-time audio responses. I will also focus on dynamic instruction architecture, pulling configurations based on user profiles easily. Thank you for considering my proposal! I look forward to the opportunity to work together. Best, Yurii.
$155 USD in 1 Tag
0,0
0,0

Timișoara, Romania
Zahlungsmethode verifiziert
Mitglied seit Feb. 11, 2026
$30-250 USD
$30-250 USD
$250-750 USD
£20-250 GBP
₹600-1500 INR
₹37500-75000 INR
$30-250 USD
₹1500-12500 INR
₹100-400 INR / Stunde
€12-18 EUR / Stunde
€30-250 EUR
$15-25 USD / Stunde
$8-15 USD / Stunde
₹12500-37500 INR
$2-8 USD / Stunde
$15-25 USD / Stunde
$30-250 USD
$10-30 USD
$1500-3000 USD
₹1500-12500 INR
₹12500-37500 INR
$20000-50000 USD