Today we live in a multi-device
world, and the way we use them spans different platforms and form factors: We read the morning news on our tablets, check email during the morning commute on our phones and at work on our desktop PCs. And at night, we watch movies on our home media consoles.
The Windows platform targets devices ranging from desktop PCs, laptops and smartphones, to large-screen hubs, HoloLens, wearable devices, IoT and Xbox. The device landscape is further diversified with Android and iOS devices, emerging VR/AR solutions, and new IoT products. This heterogeneous environment provides the average user with many choices and device options.
However, the tasks we perform on a daily basis (whether at home with family, or at work with colleagues) are not inherently device-centric, but rather human-centric
. As we increase our device count and rely more on apps to run our lives, it is becoming more complicated to get things done.
is a platform for creating experiences that transcend a single device so they can harmonize across devices – empowering a developer to create human-centric scenarios that move with the user and blur the lines between their devices regardless of form factor or platform. This vision is beginning to take shape in the Windows 10 Anniversary Update (Windows 10, Version 1607) with the Remote Systems API
, enabling developers to extend their app experiences across Windows devices connected proximally or through the cloud.
This blog post covers the functionality of the Remote Systems API by walking through an example app experience built on Project Rome, and encourages developers to break down the barriers between devices to reduce friction and better serve your users’ needs...