Show Menu

User Interface Management Cheat Sheet (DRAFT) by [deleted]

User Interface Management

This is a draft cheat sheet. It is a work in progress and is not finished yet.


The presen­tation and capture of user inform­ation is being isolated from the comput­ational aspects. Could this be the beginning of user interface management systems (UIMSs)?

1. Responsive design

With the rapid adoption of devices with displays of varying sizes, applic­ations must present results on everything from smart watches to large, wall-size screens. The goal of responsive design is to recognize the size and capability of various displays so that inform­ation can be presented in formats compatible with each device. This includes inform­ation presented on devices equipped with only a speaker, using speech synthesis.

2. Flexible capture software

Mobile device applic­ation developers often reuse software that allows users to choose the method for entering data. Users can type, select, click, swipe, touch, or speak to enter inform­ation.

3. Questi­on/­answer dialogues

By using VoiceXML as an intell­igent API, users can speak and listen to a questi­on/­ans­wer­-style conver­sation that produces a string of charac­ters, which are returned to the underlying applic­ation. VoiceXML can also handle conver­sations using text messages.

4. Conver­sat­ion­al-­style user-c­omputer dialogues

Chatbots are trained using machine learning to recognize the intentions of users from spoken words and texts and provide the underlying app with the corres­ponding commands and parame­ters. Software agents based on these techno­logies use short, natural conver­sations through smart speakers.

The UIMS To-Do List

These UIMSs should be able to do the following:

1. Select the presen­tation format and style best suited for the user and enable the most convenient technique for entering inform­ation based on user prefer­ences, usage history, enviro­nmental consid­era­tions, and available devices.
2. Incorp­orate state-­of-­the-art techniques and designs
3. Use standard APIs for transm­itting inform­ation among user interface components and the back-end applic­ations.
4. Use micro-­ser­vices for providing specia­lized functions, including automatic speech recogn­ition (ASR), text to speech (TTS), dialogue manage­ment, vision, orient­ation, and location, while supporting new micro-­ser­vices such as emotion detection and lip reading as they become available.
5. Support the gradual evolution of users from novices to experts as they become more proficient with the UIMS.
6. Provide consis­tency in user interface style across different user interface modes and platforms, so users can apply their knowledge of one user interface to another.

Stumbing Block

A major trend in app develo­pment today could prove a stumbling block to UIMSs: using the APIs of micro-­ser­vices within applic­ations. While this enables a wide variety of new and useful applic­ations, the applic­ations themselves are bound to the micro-­ser­vices, which makes it difficult to update a micro-­service or replace it with one from another vendor. Using micro-­ser­vices in this way will entangle the user interface with the comput­ational logic of applic­ations. This is counter to the philosophy of having a UIMS isolate the user interface from comput­ational logic.