Company: 360 World, Inc.
Website: https://360world.com
Location: Remote / Hybrid / On-Site (Silicon Valley)
Employment Options: Full-Time / Part-Time / Contract-to-Hire / Contract
About 360 World, Inc.
360 World is pioneering the Presence Economy™, enabling every human to earn from
anywhere by lending their real-world presence through their smartphone. Xeniq is the app
with intelligent in-app execution engine that makes this possible—turning remote actions into
precise, safe, real-time physical execution.
Role Overview
We are hiring a Senior Mobile Engineer with deep expertise in Flutter, real-time sensor fusion,
AR-style 3D object stabilization, privacy-preserving computer vision, and agentic offline
execution.
This role sits at the heart of our real-time execution engine for millions of remote helpers.
Key Responsibilities
1. Real-Time Command Interpretation
· Build ultra-low-latency pipelines to receive consumer commands.
· Generate synchronized directional animations and voice cues.
· Implement deterministic state machines for safety-critical execution.
2. Sensor Fusion & Intelligent Motion Guidance
Use accelerometer, gyro, and magnetometer data to detect:
· Speed
· Direction & rotation angle
· Jerks, drift, or shaky motion
· Incorrect movement (too slow, too fast, wrong angle)
Trigger auto-corrections via:
· Visual cues
· Professional voice instructions
· Temporary blurring as a warning
Maintain <50ms latency.
3. Agentic AI for Offline Task Execution
Build lightweight on-device agentic systems capable of:
· Understanding intent sequences (“move 3m then stop”)
· Continuing execution when offline
· Predicting incorrect motion
· Auto-correcting without requiring new consumer input
4. Real-Time Object Detection & Privacy Masking
Develop edge AI systems that:
· Detect faces, number plates, and sensitive objects.
· Apply blur before streaming, ensuring true privacy at source.
· Run efficiently across mid-range devices.
5. AR-Style Object Stabilization & Persistent Blurring (Pokémon GO–like)
A major requirement
· Once a face/plate is blurred, the blur must stay locked to that object in 3D space even as:
· The camera moves
· Angles shift
· Lighting changes
Use techniques including:
· Optical flow tracking
· Feature-point detection
· ARKit/ARCore anchoring concepts
· Multi-frame object persistence
Blur should not flicker or jump.
6. Provider — Consumer Feedback Loop
· Send structured telemetry for motion deviation and error states.
· Drive consumer-side safety screens (“billing paused due to difficulty”).
· Integrate deeply with backend telemetry and billing.
7. Cross-Platform Architecture
Build the entire experience in Flutter, with advanced native integrations for:
· Motion sensors
· CV + blurring
· GPU acceleration
· AR stabilization primitives
Required Skills & Experience
Mandatory Technical Requirements
· 5+ years of hands-on Flutter experience (required)
· 5–12 years total mobile engineering experience
· Strong, demonstrable experience in:
o IMU sensor fusion (Kalman filters, quaternions, complementary filters)
o Real-time computer vision (TFLite, MediaPipe, CoreML, ONNX)
o On-device blurring, masking, data redaction
o AR tracking, optical flow, feature locking
o Deterministic state machine design
o High-performance animation + real-time audio sync
o Flutter + native platform channels (Swift/Kotlin)
Strong Plus
· Robotics or drone navigation experience
· SLAM, ARKit, ARCore
· Edge inference optimization
· Telepresence or real-time safety-critical applications
Why Join 360 World & Xeniq
· Shape the Presence Economy™, a new global workforce model.
· Build technology that is equal parts AI, AR, CV, robotics, and human empathy.
· Work directly with ex-Google, Silicon Valley–level leadership.
· Impact millions who can earn from anywhere.
How to Apply
Submit your resume, project samples, or demo links to:
[email protected]
Website: https://360world.com
Apply Now
Apply Now