veda.ng
Essays/The Sensory Internet

The Sensory Internet

The haptic technology market reaches $5-12 billion in 2025. Apple Vision Pro introduced spatial computing. Neuralink implanted its first human BCI. The internet has been text and images for 30 years. The next layer adds touch, spatial presence, and direct neural input.

Vedang Vatsa·June 27, 2025·5 min read
Infographic
The Core Thesis

The internet is a visual medium. Text, images, and video account for nearly all information transmitted. This is a fraction of human sensory capacity. Touch, proprioception, temperature, smell, and spatial awareness carry information that no screen can reproduce. The sensory internet extends digital transmission beyond sight and sound — through haptic interfaces, spatial computing, and brain-computer interfaces — creating a medium that communicates with the full bandwidth of human perception.

The Bandwidth Mismatch

Human sensory processing operates at approximately 11 million bits per second across all channels. Vision accounts for roughly 10 million. Hearing accounts for approximately 100,000. Touch accounts for approximately 1 million. Smell and taste account for the remainder.

The current internet communicates through two channels: vision (text, images, video) and hearing (audio, voice). This captures approximately 92% of sensory bandwidth but misses the embodied, spatial, and tactile dimensions that define physical experience.

$5-12B
Haptic technology market (2025)
Industry estimates
11-13%
Haptic market CAGR through 2030
23 ms
Minimum latency for haptic realism
IEEE research
1M bps
Human touch bandwidth
Neuroscience estimates

The mismatch matters commercially and experientially. A surgeon learning a procedure from a video receives visual information but no haptic feedback — the resistance of tissue, the pressure required for incision, the texture differences between healthy and pathological tissue. A remote worker on a video call receives visual and auditory information but no spatial presence — the embodied sense of being in the same room. An online shopper sees a product image but cannot feel the fabric weight, the surface texture, or the fit.

Haptic Technology

The haptic technology market reached $5-12 billion in 2025 (estimates vary by scope), growing at 11-13% CAGR through 2030.

The technology is advancing through three generations:

Generation 1: Vibration. Eccentric Rotating Mass (ERM) motors — the vibration motor in your phone. Binary feedback: on or off, one intensity. This is what most current devices use for notifications and basic game feedback.

Generation 2: Linear actuators. Linear Resonant Actuators (LRAs) — used in Apple's Taptic Engine and Sony's DualSense controller. These provide variable intensity, frequency, and waveform control, enabling textures, clicks, and directional feedback. The DualSense's adaptive triggers simulate the tension of drawing a bowstring or the resistance of pressing a car brake.

Generation 3: Force feedback and surface haptics. Ultrasonic arrays that project focused pressure in mid-air (companies like Ultraleap), electroactive polymers that change shape and stiffness when voltage is applied, and microfluidic arrays that create localized temperature and pressure sensations. These technologies enable the simulation of touching objects that exist only in digital space.

The latency requirement for haptic realism is approximately 23 milliseconds — the threshold below which the brain perceives touch as instantaneous. Current 5G networks achieve 10-20ms latency under ideal conditions. Earlier network generations could not consistently meet this requirement, which is why haptic interaction over networks has been impractical until now.

Spatial Computing

Apple Vision Pro, released in 2024, defined the "spatial computing" category — a paradigm where digital content is placed in, and interacts with, the physical environment rather than being confined to a rectangular screen.

The interface operates through three input modalities: eye tracking (where you look), hand gestures (what you do with your hands), and voice (what you say). No controllers, no mouse, no keyboard for primary interaction. The device tracks the user's environment through cameras and LiDAR, renders digital objects with real-world lighting and shadows, and allows natural interaction through gaze and gesture.

The screen is a window. Spatial computing replaces the window with the room itself. Digital objects occupy physical space, respond to physical light, and can be manipulated with physical gestures. The interface no longer mediates experience. It augments it.

The competitive response has been immediate. Meta's Quest series targets the consumer and enterprise market at lower price points. Microsoft's HoloLens targets industrial and military applications. Google's XR investments focus on integrating spatial computing with its AI and search infrastructure.

The missing piece in spatial computing is haptics. Current spatial interfaces are visual-gestural: you see digital objects and manipulate them with hand movements, but there is no physical resistance. Picking up a virtual object provides no weight. Pressing a virtual button provides no click. This gap between visual realism and tactile absence is the primary barrier to the "presence" that spatial computing promises.

Brain-Computer Interfaces

The most direct solution to the sensory bandwidth problem is bypassing physical senses entirely.

Neuralink implanted its first human brain-computer interface (BCI) in January 2024. The N1 implant, placed in the motor cortex, enabled a quadriplegic patient to control a computer cursor through thought alone — moving the cursor, clicking, scrolling, and typing at speeds approaching those of able-bodied users.

The current medical applications are focused on restoration: enabling paralyzed patients to interact with digital devices, restoring communication for patients with ALS, and potentially restoring sensory input for patients with spinal cord injuries.

The longer-term trajectory — neural interfaces that provide bidirectional communication between the brain and digital systems — would create a fundamentally new medium. Information would be transmitted not through eyes and ears but directly into sensory cortex. Touch, temperature, spatial awareness, and potentially novel sensory modalities that have no biological equivalent could be delivered directly.

The Sensory Internet Stack
LayerCurrent StateNear-Term (3-5 years)Long-Term (10+ years)
VisualScreens, spatial computingLightweight AR glassesRetinal projection, neural visual
AudioSpeakers, spatial audioBone conduction, directionalNeural audio
HapticVibration motorsForce feedback gloves, mid-airFull-body haptic suits
OlfactoryNoneDigital scent cartridgesNeural olfactory
NeuralMedical BCI (motor cortex)Expanded medical applicationsBidirectional sensory

The Presence Threshold

The commercial and experiential question is whether these technologies — haptics, spatial computing, and neural interfaces — can cross the "presence threshold": the point at which a digital experience is perceptually indistinguishable from a physical one.

Current estimates place this threshold at a combination of: visual resolution exceeding 60 pixels per degree (achieved by Vision Pro), haptic latency below 23ms (achievable on 5G), spatial audio with head-related transfer functions (standard in spatial computing devices), and consistent sub-20ms motion-to-photon latency (the primary contributor to motion sickness in VR).

No current system meets all requirements simultaneously. But each requirement is individually achievable with existing or near-term technology. The integration challenge — combining all sensory modalities with consistent low latency in a device that is comfortable for extended use — is an engineering problem, not a physics problem.

Key Takeaway

The internet transmits through two of five major sensory channels (vision, hearing), capturing ~92% of human sensory bandwidth but missing touch (1M bps), spatial awareness, and other embodied dimensions. The haptic technology market reached $5-12 billion in 2025 (11-13% CAGR). Apple Vision Pro defined spatial computing — eye/hand/voice input replacing screens. Neuralink's first human BCI implant (2024) demonstrated direct neural-to-digital communication. The sensory internet requires converging three technologies: haptic interfaces (sub-23ms latency for realism), spatial computing (60+ pixels per degree, sub-20ms motion-to-photon), and brain-computer interfaces (bidirectional neural communication). The presence threshold — perceptual indistinguishability from physical experience — is individually achievable for each channel. Integration is the remaining engineering challenge.