See the attached pdf for article need Select 2 article and 1 page of review for each article: Total 2 pages No specific formatShould include brief summary, discussion, insights, implications to you,

Full Terms & Conditions of access and use can be found at

https://www.tandfonline.com/action/journalInformation?journalCode=hhci20 Human –Computer Interaction

ISSN: (Print) (Online) Journal homepage: https://www.tandfonline.com/loi/hhci20

ML Lifecycle Canvas: Designing Machine Learning-

Empowered UX with Material Lifecycle Thinking

Zhibin Zhou, Lingyun Sun, Yuyang Zhang, Xuanhui Liu & Qing Gong

To cite this article: Zhibin Zhou, Lingyun Sun, Yuyang Zhang, Xuanhui Liu & Qing Gong (2020)

ML Lifecycle Canvas: Designing Machine Learning-Empowered UX with Material Lifecycle

Thinking, Human –Computer Interaction, 35:5-6, 362-386, DOI: 10.1080/07370024.2020.1736075

To link to this article: https://doi.org/10.1080/07370024.2020.1736075

Published online: 30 Apr 2020.Submit your article to this journal Article views: 738View related articles View Crossmark dataCiting articles: 1 View citing articles ML Lifecycle Canvas: Designing Machine Learning-Empowered UX

with Material Lifecycle Thinking

Zhibin Zhou a,b,c , Lingyun Sun a,b,c , Yuyang Zhang b,c, Xuanhui Liu a,c, and Qing Gong a,b,c

aInternational Design Institute, Zhejiang University, Hangzhou, China; bKey Laboratory of Design Intelligence and Digital Creativity of Zhejiang Province, Hangzhou, China; cInternational Design Institute, Zhejiang University, Hang

Zhou, China

ABSTRACTAs a particular type of artificial intelligence technology, machine learning

(ML) is widely used to empower user experience (UX). However, designers, especially the novice designers, struggle to integrate ML into familiar

design activities because of its ever-changing and growable nature. This paper proposes a design method called Material Lifecycle Thinking (MLT)that considers ML as a design material with its own lifecycle. MLT

encourages designers to regard ML, users, and scenarios as three co-crea- tors who cooperate in creating ML-empowered UX. We have developed ML

Lifecycle Canvas (Canvas), a conceptual design tool that incorporates visual representations of the co-creators and ML lifecycle. Canvas guides designersto organize essential information for the application of MLT. By involving

design students in the “research through design ”process, the development of Canvas was iterated through its application to design projects. MLT and

Canvas have been evaluated in design workshops, with completed propo- sals and evaluation results demonstrating that our work is a solid stepforward in bridging the gap between UX and ML.

ARTICLE HISTORYReceived 16 June 2019 Revised 25 February 2020Accepted 25 February 2020

KEYWORDSDesign method; machinelearning; user experience; design tool

1. Introduction

As a particular form of artificial intelligence (AI) technology, machine learning (ML) not only provides

opportunities for the user experience (UX) community , but also raises several desi gn challenges because of

its malleable nature (Gillies, Fiebrink, Garcia, et al., 2016 ; Kuniavsky & Churchill, 2017 ). ML can assist

designers in extracting patterns from complex information (Witten et al., 2016 ), such as understanding

semantics from speech (Deng & Li, 2013 ). Thus, ML can be used to empower the UX of various interactions.

However, the intelligence of ML comes from the learn ing of massive labeled datasets like audio clips

annotated with semantic tags. The learning/developm ent process of ML involves the annotation of data, the

construction and training of an ML model, predicti on, and so on. This constant growth means that the

performance of ML might evolve across the learning p rocess.Thus,wethinkofMLasbeinglikeagrowable

“tree ”instead of “wood ”with fixed attributes. The static “wood ”is easily carved into a desired form. When

faced with a “tree ”, however, the designers have to consider its entire lifecycle and growth process. The

malleable nature of ML hinders designers, especially novice ones, from using it in their regular design

activities (Dove et al., 2017 ;Yang, 2017 ). To cultivate the desired UX, a designer is supposed to behave like

a qualified gardener who would influence the growth ofthetreebyfertilizing ,trimming,andotherwise

caring for it. Designers are supposed to understand the growth process of ML and know how to create the

final UX with proper resources and treatment. For example, the training data could be regarded as the

“fertilizer ”that promotes the growth of ML, the application scenario could provide the required “sunlight

and physical environment ”, and iterative training could be considered as the process of “trimming the

CONTACT Lingyun Sun [email protected] Key Laboratory of Design Intelligence and Digital Creativity of Zhejiang Province, Hangzhou, China

© 2020 Taylor & Francis Group, LLC

HUMAN –COMPUTER INTERACTION 2020, VOL. 35, NOS. 5 –6, 362 –386 https://doi.org/10.1080/07370024.2020.1736075 leaves ”that crafts the performance of ML. The growth of ML is closely related to users and scenarios, so

these three factors should be considered simultaneously following a specific method in the conception

phase.

However, there are currently few design methods that assist novice designers in fully understanding ML

technology, users, and application scenarios to addr ess these development issues in UX design. Firstly, UX

designers lack a technical background and are not well prepared to leverage ML capabilities (Dove et al.,

2017 ), because ML attributes such as accuracy and speed constantly change during the development process.

Secondly, the potential value of users is ignored to improve the performance of ML and bring users better

experience (Zhu et al., 2018 ). When faced with the unfamiliar growable nature of ML, designers are unable

to involve the available user inputs in the development process of ML (De Choudhury & Kiciman, 2018 ;

Yang, Suh et al., 2018 ). Thirdly, designers are confused about the application scenarios of ML. For example,

designers need to consider the restrictions of the envi ronment, laws, and regulations when proposing an ML

solution in the real world (Amershi, Weld et al., 2019 ). As a result of the above challenges, designers rarely

envision how ML might empower UX within current conceptual design activities, which leads to UX issues

remaining unresolved.

Therefore, designing ML-empowered UX demands methods that support designers in consider-

ing the malleable nature of ML, users, and the scenario at the conceptual design stage. Although

typical design concepts like service design thinking (Bitner et al., 2008 ) illustrate how all stake-

holders/co-creators can be taken into account, conceptual design methods for empowering UX with

ML should be tailored to the attributes of ML. This is because the malleable nature of ML makes it

quite complex to consider all co-creators (ML, user, scenario) in a dynamic development process.

Designers are supposed to participate in such development processes with design thinking that they

are familiar with (Rozendaal et al., 2018 ; Yang, 2017 ). However, existing design methods barely

consider the malleable nature of the material, which prevents designers from taking advantage of the

familiar design thinking to conceive design proposals for ML-empowered UX.

Our work aims to incorporate the development process of ML into typical conceptual design activities in

which the designers systematically consider ML, use rs, and the scenario as a whole. As the performance of

ML can evolve during its development process, the more involvement a designer has with this process, the

more likely the design proposal of ML-empowered UX i s to suit the demands of scenarios and users. To this

end, we present the basic concepts and mechanism of ML to designers instead of offering a black-box. We

also adapt the concepts from design thinking (e.g., service design, mate rial driven design) to integrate the

existing research on ML –human interaction, helping designers t o understand the malleable nature of ML

and to consider ML, users, and the scenario in a holistic way.

In this paper, we propose a design method called Material Lifecycle Thinking (MLT) that regards

ML as a constantly growing material with an iterative “lifecycle. ”MLT involves designers in the

development process of ML with the consideration of co-creators who cooperate in creating the UX.

We also present ML Lifecycle Canvas, a conceptual design tool derived from MLT mainly for novice

designers. The evaluation and reflection of MLT are conducted by applying Canvas to design

activities. The Canvas creates a visual schematic incorporating the perspectives of ML, users, and

the scenario, detailing the ML lifecycle from data annotation to ML model update. For this purpose,

we first study the characteristics of ML by identifying associated UX issues and summarizing the

related factors through a literature review (see Figure 1 ). We then present a prototype of Canvas and

Figure 1. Roadmap of our work.

HUMAN –COMPUTER INTERACTION 363 iterate it according to the collected feedback. The final version of Canvas is compared to a typical

conceptual design tool in workshops involving 32 participants. Analysis of completed design

proposals and evaluation results reveals that our system is effective in enhancing the understanding

of ML and tackling ML-related UX issues.

Our contribution consists of three parts. First, the proposed MLT method encourages designers to

reflect on the whole lifecycle of a malleable material like ML and cultivate it into the desired form by

considering the co-creators. Second, we developed Canvas, a conceptual design tool for applying

MLT and inspiring novice designers to conceive ML-empowered UX proposals without extensive

ML skills. By representing ML, users, and the scenario in a visual way, the interplay between them

can be charted and analyzed throughout the entire ML lifecycle. Finally, we share the lessons learned

in developing Canvas, which may offer insights for future research on the integration of other AI

technologies with UX design.

2. Related work

As an important branch of AI, ML is defined as a series of computer programs that learn from

experience and improve performance without using explicit instructions when performing a specific

task (Koza et al., 1996 ; Mitchell, 1999 ). A typical ML system is a data-driven deep neural network

that includes data annotation, neural network building, model training, and model prediction. In

recent years, the broad application of ML has promoted considerable research on democratizing ML

in the field of design.

2.1. Lowering the barrier to ML for novice designers

Existing practices aimed at the democratization of ML have provided programming tools, educa-

tional material, and courses for novice designers, enabling them to understand basic concepts and

practically construct ML prototypes.

Current programming tools assist novice designers in creating ML prototypes with simple code.

These tools can be categorized as MLaaS (machine learning as a service platform) (Ribeiro et al.,

2016 ), such as IBM Watson; open-source toolkits like TensorFlow (Abadi et al., 2016 ) that require

programming skills; and non-programming tools including Yale and Wekinator (Fiebrink & Cook,

2010 ). Additionally, hardware platforms like AIY (Google, 2019b ) provide easy-to-use tools to

assemble physical ML artifacts. Platforms such as Google ’s autoML (Google, 2019a ) or IBM ’s

autoAI (IBM, 2019 ) help engineers and designers to easily construct or train high-quality ML

models according to their needs. These facilitate designers in prototyping through appropriate

simplification of the abstract technical details and processes of ML. However, they do not really

help designers with little technical background to understand the mechanisms of ML and ideate ML-

empowered UX when faced with the unfamiliar growable ML technology.

Existing practice in ML-related design courses provided us with possible course module settings

(McCardle, 2002 ), design processes (Rozendaal et al., 2018 ), and a tangible tool platform (Vlist et al.,

2008 ). Moreover, current educational material (Hebron, 2016 ) and text books for developing

statistics literacy (Kruschke, 2011 ; McElreath, 2018 ) guide designers toward a preliminary under-

standing of ML ’s capabilities and working mechanisms. Although these practices and tools inspire

interesting design cases, there has been little attempt to understand the malleable nature of ML and

consider the interests of the stakeholders/co-creators (i.e., users and scenarios).

The studies mentioned above lower the barrier to accessing ML and simplify the workflow of

manipulating ML technology. However, these practices do not focus on the conceptual design, which

is indispensable for novice designers, and do not incorporate users and scenarios. Thus, there is

a need for ML-related design methods and tools that support conceptual design activities while

considering both users and scenarios.

364 Z. ZHOU ET AL. 2.2. Existing conceptual design tools

Conceptual design is a series of activities that outline and draft the key characteristics of an experience

(Atasoy & Martens, 2011 ; Hanington & Martin, 2012 ). An essential part of the early design process is to

initiate creative reflection and plan subsequent phases (Kim & Ryu, 2014 ). To envision hypothetical

experiences, designers tend to create novel assemblies of known things and conduct reflective con-

versations with design materials, users, and scenarios (Schon & Desanctis, 2005 ).

Typical conceptual design tools (Tschimmel, 2012 ) used in the conception phases cover empathy,

problem definition, idea generation, and experimentation (see Figure 2 ). Some of them are used in

virtually every stage of design.

The above tools work well in terms of understanding users and scenarios when designers have

a tacit understanding of the applied technologies, but they are not tailored to the unique malleable

nature of ML. Hence, they might fail to guide designers to gather crucial ML-related information and

conduct subsequent conceptual design activities. To overcome this deficiency, using the above

foundation for transforming complex information into design insights through a visual approach

(Snyder et al., 2014 ), we modify the existing conceptual design tools in the context of integrating ML

and UX.

2.3. Research on integrating ML and UX

Due to the malleable nature of ML, misalignments occur between the function and interpretation of

ML systems (Baumer, 2017 ), which affects UX. Most designers know little about how a given

technology functions and only use what they do know to extrapolate possibilities (Bucher, 2017 ),

which might not align with the real situation. Thus, recent research integrating ML and UX regards

ML as a design material and identifies its main challenges (Dove et al., 2017 ) and design values

(Yang et al., 2018 ).

Several theoretical attempts have been made to apply design thinking concepts like human-

centered design with the aim of seeking novel design opportunities about ML. Some studies

advocated “power to the people ”(Amershi, Cakmak et al., 2014 ; Hand, 1998 ; Patel et al., 2008 )

and pursued human-centered ML (Gillies, Fiebrink, Tanaka et al., 2016 ) and interactive ML (Yang,

Phase Tool Brief summary

Empathy: gather substantial information to gain an empathic understanding of the user and scenario

Personas A tool to create a fictional character whose attributes can represent a group of users. It can make the abstract idea of the user more personal and reveal insights. (Pruitt & Adlin, 2010)

Problem Definition: to identify the core problems by analyzing the gathered information

User Journey Map

A visual representation of a user’s journey through a service, showing all the touchpoints in the journey. Touchpoints are points of interaction where the user conducts transactions in the service, exchanges informatio n with products, and so on (Stickdorn et al., 2011) .

Idea Generation:envision solutions based on the insights gained in the first two stages

Brainstorm A collaborative tool whereby participants mainly generate ideas based on emotions and intuition ra ther than rati onal thinking. The main objective is to produc e a large number of ideas in a short time

Experimentation: evaluate alternative solutions

Service Blueprint A visualization tool that show s the relationship between all stakeholders, including the users, service offers, and so on. It consists of more stakeholders than just the user, so it can bridge cross-department efforts and identify opportunities (Bitner et al., 2008) .

Figure 2. Typical conceptual design tools.

HUMAN –COMPUTER INTERACTION 365 Suh et al., 2018 ). For instance, Amershi, Cakmak et al. ( 2014 ) and Wood ( 2014 ) allowed users to

freely modify the personalized labels so that ML systems could better understand user-preferences.

The Google UX community (Jess, 2017 ) encourages designers to involve users in the ML process as

a way of gaining design insights. Other recent research has drawn inspiration from the typical design

process (Rozendaal et al., 2018 ; Yang, 2017 ). Yang, Scuito et al. ( 2018 ) demand that designers

identify the problem that ML is intended to solve, validate the technical feasibility, and then iterate

the solution to craft the ML functionality.

Different design strategies have also been proposed to improve the UX of malleable, unpredictable

system. The theoretical strategy applies numerous theories from behavioral and social sciences to

help understand how users behave (Murnane & Counts, 2014 ; Recasens et al., 2013 ). The speculative

strategy involves extrapolating from existing circumstances to imagine possible futures and is less

constrained by technical feasibility (Baumer, Ahn, Bie et al., 2014 ; Dunne & Raby, 2013 ). The

participatory strategy involves people who might be impacted by the system in the design process

(Kuhn & Muller, 1993 ; Muller, 2007 ). Eric Baumer ( 2017 ) developed a human-centered algorithm

design process that incorporates the above strategies.

Moreover, some design guidelines have investigated specific ML issues, such as how to design

interactions that may produce unpredictable results. Amershi et al. ( 2019 ) developed 18 human –AI

interaction design guidelines that are organized into categories based on when the user interacts with

AI. These guidelines require designers to clarify the capabilities of AI, the context, relevant social

norms, and so on. Additionally, Google ( 2019c ) provided suggestions for specific UX issues such as

trust, failure, and data collection.

Although the above guidelines are helpful, they are not well connected to the theoretical

explorations or design strategies that inspire designers to make trade-offs between different guide-

lines based on the real-life situation and to explore other feasible possibilities. To bridge this gap,

designers require a methodological framework that allows them to use practical design guidelines

while following the abovementioned design thinking concepts and strategies.

2.4. Summary

Existing prototyping tools and design education have certainly lowered the barrier to ML. However,

as current conceptual design tools are not tailored to the ML context, there is still no design method

or tool that assist designers in ideating novel and practical approaches using ML. Moreover, recent

research has suggested that there is a gap between detailed guidelines and theoretical recommenda-

tions like “power to the people ”and “retain the typical design process ”. Moreover, there is a lack of

research on encouraging designers to participate in the growing process of ML with design strategies

and consider ML, users, and the scenario as a whole. Therefore, what is required is a method that

can bridge the gap between theoretical foundations and detailed guidance. To construct such

a method, we need to seriously consider the malleable nature of ML and holistically incorporate

the demands of the users and scenarios.

3. Material lifecycle thinking

In this section, we introduce the MLT conceptual design method, which considers malleable tech-

nologies like ML as design materials with iterative lifecycles. MLT also encourages users and

scenarios to collaborate with materials to co-create UX. In this way, designers are able to shape ever-

changing materials through user engagement and scenario arrangement, rather than simply using

the material as a static black box.

366 Z. ZHOU ET AL. 3.1. Why MLT?

We originally intended to apply the “material driven design ”(MDD) (Karana et al., 2015 ) method to

study ML as a literal material. MDD allows designers to firstly characterize the material both

technically and experientially, and then envision the material ’s role when embodied in a product.

The designers then imagine how the users experience the material, rather than using intuition.

Finally, the designers integrate all findings into the final product concept.

MDD and similar design methods work well with static physical materials whose characteristics

might not change significantly over time. However, such methods ignore other resources involved in

the development process and would be constrained when dealing with ever-changing materials. An

ideal method for a malleable material should guide designers to understand the basic mechanisms of

the material ’s lifecycle and the factors that influence its growth.

MLT draws appropriate inspiration from existing design methods, as recent studies have suggested

that the design method should not change significantly in the context of ML (Rozendaal et al., 2018 ;

Yang, 2017 ). Essentially, it is an extension of existing design methods such as MDD, human-centered

design, value-sensitive design, and green design. Most of these methods focus on design materials with

fixed attributes, which can not cope with the challenges caused by the malleable attributes of ML.

In contrast, MLT is tailored to malleable design ma terials and focuses on the development process of

materials. For example, the idea of “lifecycle ”was borrowed from green design, such as lifecycle assessment

(Russell et al., 2005 ), to assess the environmental impacts associated with all stages of a product ’slife,from

raw material processing to maintenance and recycling. In MLT, the concept of “lifecycle ”implies that the

material is “growable ”and describes the development process of a material whose attributes vary over

different evolutionary cycles. Regarding ML, lifecyc le refers to the specific learning process because the

performance of ML can change on each iteration, inevitably affecting the UX.

3.2. What is MLT?

MLT encourages designers to participate in the main development process of the material while

considering the users and scenarios involved in creating UX. Figure 3 illustrates MLT, with the five

Figure 3. Brief introduction to MLT.

HUMAN –COMPUTER INTERACTION 367 main steps presented in a cyclic manner. The steps are as follows: (1) Understanding the technical and

experiential characteristics of the material, which helps to describe the design goal; (2) Identifying the

co-creators and understanding their technical and experiential characteristics. Here, the term “co-

creators ”is used to describe subjects that cooperate with the material to create UX during its lifecycle.

The material itself can also be regarded as a co-creator; (3) Identifying touchpoints where the co-

creators exchange resources with each other and have an impact on the material ’s development; (4)

Envisioning possible UX by considering the trade-offs between co-creators with various characteristics

to gain design insights, embodying the insights through the touchpoints of material for generating

potential UX solutions; (5) Monitoring the material, as its characteristics might evolve because of the

mutual influences among the co-creators and result in different UX. We use the breeding process of

potted plants as an example to explain how MLT deals with the growable material (see Figure 3 ). The

process starts with a malleable material (i.e., ML) and ends with a cycle that continually improves UX.

In the case of ML-empowered UX, we regard the ML, users, and scenarios as co-creators with

different characteristics and interests. We consider ML as a malleable material that requires the

cooperation of users and scenarios to support UX. During each loop of the ML lifecycle, a human

user “teaches ”ML through interaction inputs and their preferences, while the scenario provides

resources to support the ML performance and brings restrictions. Designers gain design insights by

resolving the conflicts resulting from the co-creators ’own demands. This process regards ML and

the scenario as synergistic partners. The objective of MLT is to help designers propose feasible and

suitable conceptual UX solutions, without requiring them to practically implement ML systems.

4. Development of ML lifecycle Canvas

The ML Lifecycle Canvas, a conceptual design tool featuring the holistic visualization of cooperation

among ML, users, and scenarios during the ML lifecycle, is intended to enhance design practices

with the MLT method. The development process includes the “Research Through Design ”

(Hanington & Martin, 2012 ) approach, which is a reflective process consisting of conception,

prototype, evaluation, and iteration to discover insights that will subsequently be incorporated

into the Canvas. As well as conducting a literature review, we invited designers to contribute their

ideas for improving the Canvas in a participatory way.

4.1. Phase 1: identify the design issues and key factors

The first step in applying MLT to grow ML-empowered UX is to understand ML as a malleable

material. We identified the characteristics of ML by reviewing the literature on ML-related design

challenges and influencing factors.

5. Design issues

Using search terms such as “AI, ”“ ML, ”“ UX, ”and “challenges, ”we collected recent (past three

years) publications that mentioned ML-related UX issues from the field of human-computer inter-

action (HCI). We set the scope to the past three years because of time and resource constraints, but

the summary of previous work in state-of-the-art publications also covered earlier work and offered

inspirations. We searched in the HCI community instead of the broader computer science domain

because almost all work in HCI has some relationship to people (though not only related to ML),

and is thus connected to UX design (Yang et al., 2018 ).

From the work of Holmquist ( 2017 ); Yang, Scuito, et al. ( 2018 ); Forlizzi, Zimmerman, Mancuso et al.

(2007 ); Dove et al. ( 2017 ), and our previous work (Zhou, Gong et al., 2019 ), which cover most UX issues

related to ML, we identified potential detailed challenges. Some of these issues were specific to

a particular scenario, e.g., the distrust of automotive driving and unpredictable changes in adaptive

UIs. These issues were extended to the universal scenario, e.g. the distrust to AI systems. Three of the

368 Z. ZHOU ET AL. authors extracted the above-mentioned issues from the collected publications, transformed descriptions

of these issues into post-it notes and pasted them on the board to construct an affinity diagramming

process. They sorted the descriptions into different clusters, and named each cluster with the most

representative description. After an iterative process of reorganizing and merging, we obtained six

clusters, each representing one of the typical issues listed in Figure 4 . Each issue represents several

associated topics. For instance, transparency represents several concerns including explainable AI (Wang

et al., 2019 ), the interpretability of results, and the visualization of decision-making processes.

The factors that may influence these UX issues were determined through a literature review of

HCI publications and social, cultural, and psychological studies. We also reviewed current work on

design guidelines for human –AI interaction (Amershi, Weld et al., 2019 ; Google, 2019c ). After the

same affinity diagramming process, we identified the nine key factors (excluding the user category)

listed in Figure 5 . We divided the factors into the categories of ML and scenario. Regarding the user

category, existing conceptual design toolboxes provide tools such as Personas and Empathy Maps to

collect user demands and preferences.

5.1. Phase 2: conception of Canvas

To apply MLT in conceptual design, we crafted Canvas to present a visual form. Canvas enables

designers to complete the conceptual phase of t he design according to MLT in a cyclic manner.

The conception of Canvas according to the fi ve specific steps of MLT is illustrated in Figure 6 .

The first step of MLT is to understand the attributes of ML. Thus, we conceived Issue cards to

represent the six specific issues caused by ML (see Figure 6a ), which can be considered as the

material attributes. The factors related to ML (see Figure 5) offer tips for understanding the material

and describing the design goal.

Secondly, Canvas uses different rows to represent the co-creators (ML, user, and scenario; see

Figure 6b ). The relationships among the three co-creators are somewhat ambiguous, but we clearly

assigned distinct parts to each co-creator to avoid pre-defining relationships and restricting the

flexibility of Canvas. Canvas does not intend to imply any relationships among the three co-creators,

leaving designers with the freedom to mine potential relationships using the advantage of MLT. We

used different approaches to help designers understand these co-creators. In particular, we custo-

mized the Persona to promote empathy with users. Our Persona format contains the pain points,

needs, and behavioral demographic information related to ML, which aids user studies in an ML

Issues Brief summary

Unpredictability The issue of managing the unpredictable conclusions given by ML systems according to given data no matter how well the ML system is trained.

Transparency The issue of designing the product whose workings nobody can quite explain. And how does this affect qualities like trust and confidence in the system?

Learning The issue of using available user input to not only evolve the performance of ML with unobtrusive interaction but also avoid the bias in data.

Control The issue of designing ML systems that allow the co-control with the user, so that users can truly cooperate with ML system in decision making.

Anthropomorphism

The issue of determining the anthropomo rphic degree that can impact on how people perceive the intelligent system like perceived task suitability, engagement with agents, and perceived attractiveness.

Interactivity

The issue of supporting the interaction between user feedback and newly learned

ML models. Many ML models are built to offer improvements in accuracy and

may not run quickly enough for real-time interaction.

Figure 4. Specific UX issues caused by ML.

HUMAN –COMPUTER INTERACTION 369 context. As for the scenario and ML system, the related factors in Table 3 provide guidance for

determining the questions worth considering.

In stage 3, the different columns of Canvas are used to describe different touchpoints. The

touchpoints at which the users, scenario, and ML system exchange resources are shown in

Figure 6c . We used the stages of the ML process (i.e., data annotation, training, inference) to pre-

define the touchpoints, reducing the cognitive burden on the designers.

The most important step is to envision the potential UX and embody the design insights about users

and scenarios in specific steps of the ML lifecycle (touchpoints). To this end, the information collected

in stages 1 –3 is organized in a visual form in Canvas to enable the identification of design insights (see

Figure 6d ). The three co-creators have their own goals and optimal interests, but sometimes they have to

make compromises and concessions for the final experience. For example, acquiring more data can

certainly help to train a better-performing ML application, but the data size might be limited by the

user ’s willingness and regulations in the scenario. As discussed above, the mutual relationships of the

co-creators vary greatly with the type of ML system, the user, and the scenario. Therefore, Canvas

provides a template guiding designers to think about issues worth considering for such ambiguous

relationships, instead of clearly mapping the factors in Table 3 to specific issues or touchpoints.

Figure 6. Conception of Canvas.

Category Brief summary of the factors

Scenario

Societal impact: Aspirations concerning culture, poli tical system and safety. (Cuevas et

al., 2007)

Ta s k t y p e s : Cognitive type, control type and perceptual aids type, etc.

Physical environmen t:Light, humidity, noise, location etc.

ML system

Automation level: Predictability and stability of automation (Merritt et al., 2013)

Capability: Types of tasks that the ML system can support.

Aesthetics: Aesthetic appearance, which is related to automation level and instructions

(Li & Yeh, 2010).

Co-control Mode: Fixed, adjustable (e.g., set by the us er), or adaptive (e.g., accounts for

user’s state) (Hancock & Chignell, 1988).

Hardware: Internet, power supply (e.g., battery), computing capability (e.g., GPU), etc.

Learning ability: Amount and quality of required data, the difficulty of iterative training,

and the required computing resources etc.

Figure 5. Factors influencing the scenario and ML system.

370 Z. ZHOU ET AL. Finally, the attributes of ML might change after each loop of the lifecycle following the involve-

ment of the users and scenario. The design process supported by Canvas is iterative, and designers

thus need to pay close attention to the evolution and growth of the ML system to revise the design

proposal accordingly.

5.2. Phase 3: Canvas Prototype

The Canvas Prototype consists of Canvas I, Canvas II, Issue cards, and Persona cards. Their usage is

illustrated in Figure 7 . To seek potential improvements, the prototype was used in the conceptual

phase of a design project, with essential technical support provided.

6. Participants

Two instructors and 30 junior (third-year) undergraduate design students majoring in industrial

design (M: 17, F: 13) were involved in a design project with the theme of “ML-empowered product

design ”(see Figure 8 ). The project was part of a compulsory course and students received credit for

the course. We selected the students who had long-term (at least 3 months) internship experience

and had done at least 2 practical design projects to ensure that they could be regarded as novice

designers. These design students had been trained in the same way. They had already mastered

mainstream design tools, but lacked a deep understanding of ML, which helps us to obtain objective

results. The participants were split into six groups, with the two instructors offering guidance to each

group as required. The instructors explained the usage of Canvas and offered technical-related

suggestions, but were not responsible for the design process.

Figure 7. Prototype of Canvas.

Figure 8. Working with the Canvas prototype.

HUMAN –COMPUTER INTERACTION 371 7. Procedure

First, the participants were introduced to UX issues, related factors, and the usage of Canvas.

According to the topic assigned to each group, they then gathered essential information through

user studies and online research. Furthermore, the participants entered the information into Canvas

and discussed the trade-offs in the optional solutions to serve the application scenario and users.

Finally, Canvas was used to exchange and evaluate design ideas.

Building on the design tools discussed in Section 2 and other research on developing creative

tools (Atasoy & Martens, 2016 ; Jacobs et al., 2017 ), we developed four evaluation criteria. Each

participant took part in a 10 –15 minutes semi-structured interview. During the in-person interviews,

we asked the laddering questions (Reynolds, & Gutman, 1998 ) that were adjusted for deeper

investigation according to the participant's answers. . Laddering refers to an in-depth interviewing

technique for selecting the most appropriate level of question or researcher response to respondent

dialogue, based on the premise that we arranged our questions in an order that starts with the least

invasive and proceeds to deeper matters. The following were part of the laddering questions.

Ease-of-use: Do you think the tool is easy to use? Why do you feel that it is (or is not) easy to use?

What elements do you think are important for an easy-to-use tool? Why is this important to you?

Creativity: Do you think the tool helps you to come up with diverse ideas? Why do you feel that it

is (or is not) helpful for creativity? What kind of ideas does the tool help you to create? Why is the

tool helpful for creating the mentioned ideas?

Process: Do you think the tool helps you to integrate prior expertise in the design process? Why

can you (or can you not) integrate prior expertise? Why do you feel that it is (or is not) helpful in the

design process? How does this tool relate to the traditional design process?

Understanding: Do you think the tool helps you improve your understanding of ML? Do you

think the tool encourages you to think about issues and factors related to ML? What do you feel

about that? What are the critical aspects of the tool in helping your understanding of ML?

8. Results

At the end of the project, all six groups produced concept designs varying from modular devices for

connecting distant families (Zhou, Jiang, et al., 2019 ) to a mobile app for assisting children in

drawing comics. A completed Canvas is shown in Figure 9b , including detailed information and

solutions for Cabe (see Figure 9a ), an AI robot that senses changes in emotion and reacts appro-

priately to help users achieve a better learning experience. How the final version of Cabe dealt with

the UX issues has been discussed in our previous work (Zhibin Zhou et al., 2019).

We applied a content analysis process (Howitt & Cramer, 2010 ) to summarize the interview results.

We took notes during the interviews and tagged them with the eight pre-defined codes related to the

Figure 9. Cabe: one of the design outcomes of the six groups.

372 Z. ZHOU ET AL. strengths and weaknesses of four evaluation criteria. We used the audio recordings of interviews to

avoid missing important information. One of the authors who is most familiar with the notes

conducted the coding and other analysis processes. For example, “the cyclic nature of the ML lifecycle

is not obvious in the Canvas (participant 5) ”was assigned the code of “negative comment to under-

standing criterion ”while “it helps me learn more about the ML concepts and its process (participant

16) ”was assigned the code of “positive comment to understanding criterion. ”The coder was respon-

sible for the whole project and designed all the questions for the interview, as well as taking the notes

during the interviews. Thus, we considered him as the most suitable person for the coding task to

ensure that the analyzer of the data understood the context of the entire conversation. From the

analysis, we found that Canvas helped the majority of participants gain an increased understanding of

ML-related factors and UX issues and imagine appropriate solutions. To identify the weaknesses of

Canvas, the main negative comments are summarized in Figure 10 .

8.1. Phase 4: iteration of Canvas

On the basis of the evaluation results, we made the following improvements to the Canvas. These

improvements mainly addressed issues including overly simplified visualization layout, ML process

(touchpoints) involving technical terms, and ambiguous factors.

● We transformed the tabular visualization layout implying a linear ML process into

a iterativeCanvas is split into six fan-like sections, one to visually suggest a recyclable and

realistic ML process.

● We combined Canvas I and Canvas II, isolated the Persona cards from Canvas, and rearranged

its format to evoke empathy because users do not change their demands significantly at

different stages of ML process. We also presented completed Canvas designs as examples and

provided instructions that introduce the usage of Canvas.

Evaluation criteria Brief summary

Ease-of-use

1) Canvas had many ML technology terms and details that hinders participants from understanding touchpoints and the stages of ML process;

2) Participants learn little Creativity about the usage of Canvasthrough the

Creativity

1) Some aspects of ML and scenario ar e ambiguous and complex so it was difficult for participants to ask appropriate questions;

2) Canvas offers large quantities of information combining ML technology with UX, while participants are accustomed to thinking about the design issues without much technical detail.

Process

1) Canvas was kind of inconsistent with similar design tools because it was split into two parts;

2) Some participants complained that th e complexity of factors and ML process prevented them from ffreely using Canvas Understanding in design process.

Understanding

1) ML process in Canvas was still quite abstract, especially when participants wanted to match the ML process to the corre sponding after stage o f product lifecycle;

2) The factors regarding the ML and scen ario were ambiguous so the participants might have different understandings;

3) Some participants ignored the cyclic nature of the ML process because the linear form of Canvas assumes a simple linear ML process.

information transmitted by the tool itself without the explanation of attached cases.

Figure 10. Interview results of the prototype Canvas.

HUMAN –COMPUTER INTERACTION 373 ● We modified the description of touchpoints to make them less like the terminology of ML. We

combined the data annotation and dataset processing stages into data collection and split the

original inference stage into prediction and performance.

● We provided a Question list instead of an ambiguous description of factors, as it was difficult

for designers to understand the given factors.

In particular, the most typical questions of Question list were extracted from the information of

the completed Canvas in phase 3 and existing guidebooks on human-AI interaction (Amershi, Weld

et al., 2019 ; Google, 2019c ).

The Question list contains problems worth considering in the conceptual phase. We provided

a list because the evaluation showed that designers struggled to carry out the design process without

explicit instructions. However, some questions might have implications for the ambiguous relation-

ships among the three co-creators. Additionally, this list does not cover all aspects of ML technology.

It serves as a self-checking list that can be freely modified whenever necessary, and designers are not

required to consider every question. The aim is to inspire designers to think about the most valuable

issues by answering questions.

8.2. Description of the final Canvas

The final version of Canvas is split into six fan-l ike sections, each representing one stage of the

ML lifecycle: data collection, construction, training, prediction, performance, and updating.

Canvas can be filled in collaboratively, with gro ups of people using the Question list to sketch

and model the various aspects of ML, users, and t he scenario. In each section, the green area (see

Figure 11b ) of Canvas is used to record information ab out the scenario, while the pink area (see

Figure11a ) is for ML-related information. Issue cards ( Figure 11f )arelocatedinthegrayarea

(see Figure 11c ). The Persona card shown in Figure 11e allows designers to record users ’

attitudes and feedback regarding UX issues. Figure 11d presents the Question list, with questions

categorized according to the relevant co-creator s and touchpoints. Brief usage instructions are

presented below.

(1) Answer questions related to ML in the list and add relevant information, such as the dataset,

inference speed, and neural network types in the corresponding area. Identify the UX issues

Figure 11. The final version of Canvas.

374 Z. ZHOU ET AL. to be addressed at each stage and put the corresponding Issue cards in the gray area, thus

clarifying the design goal.

(2) Conduct the user study and complete the Persona as required; identify the key features of

potential users, fill in user preferences and core appeals for ML products.

(3) Answer questions related to the scenario and add the corresponding information, thus

constructing a basic understanding of the scenario in terms of the laws, regulations, noise,

temperature and etc. Adjustments can be made to the Issue cards in this process.

(4) Explore potential proposals based on the guidance of the MLT and collected information.

To this end, identify the connections and relationships among the three co-creators and

explore solutions to deal with the issues at different stages. At this point, the conflicts caused

by the different interests of the co-creators should be resolved.

(5) Summarize the final proposal and write it down on the Issue cards for iterative evaluation.

9. Evaluation of ML lifecycle canvas

To test how well the iterated Canvas works in terms of the four criteria presented in the previous

section, we organized two-week design workshops . As there were no conceptual tools specifically

designed for ML, we compared Canvas with Service Blueprint (Blueprint) which has similar

features and goals to those of Canvas. Moreover, we presented the design outcomes of Canvas

to show its effectiveness in dealing with UX is suesrelatedtothemalleablenatureofMLin

conceptual design.

9.1. Participants

Thirty-two participants (M: 11, F: 21) were recruited to participate in the evaluation. All participants

were junior (third-year) undergraduate design students with the same background as the participants

in phase 3. They were randomly divided into six teams containing six or five participants. None of

the participants had been involved in the project in phase 3; the instructors were the same.

9.2. Procedure

We invited the 32 participants to an evaluation session to compare ML Canvas and Blueprint in two

workshops (each lasted for two weeks) on the theme of ML-empowered UX. To ensure the

participants would not be biased by the previous evaluated tool, there was a six-month interval

between the two workshops. The group organization, design topics, and environment of the two

workshops were the same. Thus, we regarded the collected data as two independent groups. Both the

workshops were part of a compulsory course and students received credit for the course.

Before each workshop, the design students were introduced to the malleable nature of ML

technology, including the mechanisms of ML, its attributes of unpredictability, data-centric nature,

the related UX issues and etc. In the first workshop, the participants were given a two-hour

introduction to the MLT method and Canvas (see Figure 12a ), whereas in the second workshop,

they were introduced to service design thinking and Blueprint.

After the introduction, all the participants were asked to complete two assignments. First, they were

asked to use Canvas/Blueprint to illustrate an existing design case, then they were required to improve

this case using Canvas/Blueprint (see Figure 12c ). During the workshop, there were several discussions

where the instructors answered questions on the use of Canvas and Blueprint (see Figure 12b ). After

using Canvas/Blueprint in each workshop, participants evaluated the four aspects of Canvas/Blueprint

on a 5-point Likert scale, with 1 indicating strong disagreement and 5 indicating strong agreement.

Each workshop was followed by the same procedure of in-person semi-structured interviews as those

in Section 4.3.2, each lasting for 10 –15 minutes. The laddering questions regarding the four criteria

HUMAN –COMPUTER INTERACTION 375 were also the same as those in Section 4.3.2. We did not show any expectancies regarding the

performance of the participants in the workshops and the evaluation part.

9.3. Evaluation results

The Shapiro-Wilk test indicated that the data from the four dependent variables (ease of use,

creativity, understanding, process) was not normal distribution. Thus, we conducted a Mann –

Whitney U test to compare Canvas and Blueprint with respect to the above four criteria. The

analysis results showed that Canvas outperforms Blueprint regarding creativity and understanding.

However, there was no significant difference between the two design tools regarding ease of use and

process. The Boxplot is presented in Figure 13 .

Figure 12. Workshops where the participants used the final version of Canvas. a) Introduction to Canvas; b) discussion; c) design with Canvas and Blueprint.

Figure 13. Comparison between Canvas and Blueprint across the four criteria.

376 Z. ZHOU ET AL. In the corresponding in-depth analysis, we analyzed the advantages and limitations of Canvas

compared with Blueprint from the results of the semi-structured interviews, using the same content

analysis process as in Section 4.3.2. The following descriptions explain the reasons for the

differences.

9.3.1. Ease of use

There was no significant difference between Canvas ( Mdn = 3) and Blueprint ( Mdn = 3.5),

U = 544.0, p=.645, r= .06.

Many participants thought that Canvas was well-structured and friendly. The reasons include the

following: (1) The visualization layout of Canvas “assisted us to organize information in a clear

manner ”(P17) because of the “well-arranged blocks corresponding to the different co-creators (P5) ”

and “concise symbols indicating different stages of ML process (P11) ”. They regarded Canvas as

a good tool to record thoughts, present ideas, and review the proposals. (2) Several participants

described Issue cards as “flexible ”because they are “quite easy to make adjustments ”(P5). For

example, someone thought that they could easily remove or add some issues cards (P8) and take

notes on the cards (P4) when discussing with partners to make sure that everyone was on the same

page. (3) They also mentioned that the Question list was friendly because “some questions directly

reminded us of problems that we had not considered ”(P22). By answering questions, participants

were motivated to “collect the related information ”(P17) and “figure out if the concept design is

feasible ”(P19).

Despite all this, the negative evaluations were mainly caused by unfamiliarity with the ML

concepts and usage method of Canvas. The participants thought that (1) the layout was too different

from the existing design tools that designers are familiar with; (2) the visual template of Canvas did

not intuitively indicate the order in which information should be entered, which complicated the

usage method. In contrast, the two-dimensional chart of Blueprint is easier to understand.

9.3.2. Creativity

The score was significantly higher for Canvas ( Mdn =4) than for Blueprint ( Mdn = 4), U = 318.0,

p= .004, r= .36.

Most of the participants thought that Canvas helped them to conceive diverse and practical UX

solutions in an ML context. (1) The majority of participants commended Canvas for concerns about

the malleable nature of ML. It allowed them to have an “open mind with regard to the six issues ”

(P7, P14, P16) and offered them definite objects instead of seeking solutions blindly. (2) The

consideration of co-creators helped them “come up with multiple solutions to a single issue ”(P7)

and “figure out not only ML-related issues, but also issues involving other aspects ”(P20). (3) The

division into six ML stages made the identification of solutions more in-depth. P17 said that he

learned to pay attention to cultivating “a system for the common progress of co-creators. ”Canvas

reminded designers to try to “strike a balance among the three co-creators (P12) ”, thus avoiding

paying too much attention to one co-creator and damaging the interests of others.

However, we also received the following complaints: (1) Several participants felt that the ideas

derived through Canvas were more like incremental innovation rather than breakthrough innova-

tion. They thought the main reason would be that the Canvas aims to provide a feasible design

scheme that can specify the next steps like prototype, manufacture and so on. (2) Canvas lacked

a mechanism to link scattered solutions into a complete scheme, so the design ideas corresponding

to the issues are difficult to integrate into a comprehensive solution. For instance, P5 and P6 pointed

out that their ideas were limited to the given stages or issues. (3) Several participants thought that the

semi-structured Question list somehow restricted their ideas. P2 and P9 suggested adding more

open-ended questions to inspire divergent thinking. (4) P15 and P21 suggested adding a blank Issue

card to describe those UX issues that are not caused by ML. Though Canvas is designed to address

ML-related UX issues, providing a way to visualize other issues would help participants to have

overall control of the design proposal.

HUMAN –COMPUTER INTERACTION 377 9.3.3. Process

There was no significant difference between Canvas ( Mdn =3) and Blueprint ( Mdn = 3), U= 591.0,

p= .256, r= .14.

Several participants reported that they could easily integrate other design tools into the use of

Canvas: (1) Some found that ideating with Canvas was “similar to service design thinking that

requires comprehensive consideration of all stakeholders ”(P14). P16 drew an analogy between

Canvas and Blueprint: “Data collection, performance, and update can be regarded as the frontstage

in service design, while other stages are like the backstage. ”(2) Most participants spontaneously

integrated typical design tools like brainstorming, interviews, and 5W2H into the process of

completing Canvas, which shows the compatibility of Canvas with other design tools. (3) Canvas

offered guidelines for conducting traditional design activities in the context of ML. Some steps of the

traditional design process need to be adjusted by the participants to incorporate ML. For example,

user research should involve new content about the UX issues that are related to ML. P13 mentioned

that “the Issue cards are helpful when creating the outline of user interviews. ”

There were some suggestions on optimizing the process. (1) Several participants believed that the

layout of Canvas should be more like those of other traditional tools to enable better collaboration.

(2) Some hoped that Canvas would provide detailed and formal guidelines on how to integrate

design knowledge into the process of using Canvas.

9.3.4. Understanding

The score was significantly higher for Canvas ( Mdn =4) than for Blueprint ( Mdn = 3.5), U= 249.5,

p< .001, r= .48.

Most participants stated that their understanding of ML had been improved by working with Canvas.

(1) Canvas not only reminded designers of issues that had been ignored (P1, P10), but also “set up

a framework to learn about ML in the way of design thinking ”(P5). (2) Some students said that Canvas

changed their perception on the role of designer in ML-empowered UX design: “I realized that UX

design can extend into the ML process like data collection and prediction ”(P7). Through using Canvas,

they learned to enhance the performance of ML by appropriate UX design, e.g., “helping the ML system

to collect user data in a more pleasant way ”(P18).

However, Issue cards might not cover all ML-related UX issues(P15, P21). Additionally, Canvas

does not clearly show the relationship among users, the ML system, and the scenario and “it is quite

difficult to image the mutual influence between them ”(P21).

9.4. Design outcomes

The completed proposals (Canvas group) are presented in Figure 14. The design outcomes of the six

groups are all conceptual design proposals and not the actual implemented algorithm/systems.

Te a m Brief summary

1. Hello Barbie Chat robot for accompanying children through their preferences

2. Clips Intelligent camera that automati cally records memorable moments of and their pets

3. Makeup Camera App A mobile app that helps users to vi rtually beautify themselves with virtual make-up in real time

4. Language Translator A website for academic translation tha t can translate sentences and words into different languages

5. Echo Dot Smart speaker tailored to the needs of children

6. E-shopping recommendation Shopping recommendation app based on social relationships and browsing history

Figure 14. Description of completed projects.

378 Z. ZHOU ET AL. To illustrate how the participants engaged in the process of growing the ML-empowered UX

iteratively over time, we present part of the design activities of team 4 as follows. The presented

activities are derived from the completed Canvas, the presentation of team 4, and instructors ’records.

9.4.1. Understand the material

Understand the attributes of ML models, including key elements such as the required dataset, the

type of the required ML model, and the inference speed (see Figure 15d ). The ML models for text

translation use a Recurrent Neural Network (RNN) that requires a large amount of computation and

storage, thus cloud computing is usually adopted as a solution. Similarly, several studies are

addressing the interpretability of RNN models and are transformed into practical applications. In

addition, such models require a large amount of high quality, annotated data to improve system

performance.

9.4.2. Understand co-creators

(1) Produce Persona (see Figure 15c ) for researchers and college students who require high-quality

scientific translation. They expect the platform to interpret the translation results, which helps to

evaluate the credibility. Meanwhile, they possess academic translation skills and can provide profes-

sional revisions. (2) Understand the scenario that mainly includes laws and regulations, social rules

and other restrictions in this case (see Figure 15d ). For example, we may need to give relevant tips

during the translation process for sensitive words and terminologies. In addition, academic activities

require high confidentiality of text, and are sensitive to data privacy.

9.4.3. Identify touchpoint

Team 4 filled the above-collected informatio ninto the corresponding areas of the ML process

(touchpoints) and identified how the co-creators could have an impact on the material ’s

growth. In this case, users revise the transl ation result so that the ML can learn from the

data provided by the users. As the ML model is re fined, the accuracy of translation will be

improved so users can benefit from the platform. Also, existing interpretable practices in

Figure 15. Part of the design proposal that engages in the process of growing the ML-empowered UX.

HUMAN –COMPUTER INTERACTION 379 RNN models can meet the user ’s demand for reliable results. However, the data privacy

requirements in this scenario conflict with the ML system that currently only provides online

services.

9.4.4. Embody

The design team tried to eliminate conflicts between co-creators ’dem ands and then filled in the

solutions of issues in the corresponding area. For example, using different colors to present the

different operating status of the system (see Figure 15b ); using example sentence and word

frequencytopresenttheinterpretationoftranslationresults(see Figure 15a ). Finally, they

produced a preliminary conceptual design pr oposal by embedding these design solutions to

the ML process.

9.4.5. Start another loop

The design team continued to monitor how the co-creators influence each other and ensure that UX

issupported by the continuous growth of ML. For example, they generated the mechanism of the

medal symbolizing professionalism (see Figure 15a ) that motivates users to produce data to con-

tinuously improve the ML. The system also differentiates data providers for their professionalism to

quickly select high-quality samples, and so on.

We also present novel ideas for coping with six UX issues from these proposals (see Figure 16 ).

These ideas demonstrate how designers tackled the malleable nature of ML through addressing the

six UX issues. Some of the completed Canvas is shown in Figure 16d .

9.4.6. Unpredictability

Participants explored several ways to alleviate the confusion caused by the uncertainty of malleable

ML applications. (1) Team 5 designed guiding questions for when the smart speaker cannot under-

stand what the user has said. For example, the robot will guide users to repeat themselves or use

other terms to interpret what has been said when the voice quality is poor. (2) Team 1 explored reply

strategies for when their robot is not confident about the answers. For example, the robot will seek

help from parents by saying “Let ’s ask mom together ”to the child. (2) Team 6 used easily under-

stood tags to describe recommendations (see Figure 16b ). This allows users to quickly understand

why an item is recommended and somehow reduces the feeling of unpredictability.

Figure 16. Outcomes of the design workshop.

380 Z. ZHOU ET AL. 9.4.7. Transparency

Various methods were conceived to expose the complexity and malleable mechanism of ML to users,

which might reduce unpleasant situations. (1) Team 3 envisaged that the ML system could visualize

feature point annotation of the face, whereupon users could adjust misplaced feature points to

enhance the performance of ML, thus improving the UX (see Figure 16c ). Moreover, manual

annotation may help the ML optimize its learning dataset. (2) Team 4 showed the current working

status (working/completed) through the color of the results box. When users choose a particular

vocabulary, the interpretation and other information (such as frequency of occurrence, reference

source) will be displayed (see Figure 16e ) to illustrate the credibility of the results.

9.4.8. Learning

The participants applied novel strategies to produce labeled data and promote the growth of the ML

application, although only with the authorization of the users. (1) Team 1 spontaneously stimulated

emotional expressions and obtained data, such as capturing the child ’s happy expression by telling

jokes. (2) Team 4 used rewards (medal symbolizing professionalism) to facilitate users making sug-

gestions on the translation results. Based on the users ’proficiency, the translation platform assigns

different weights to different users, thus attaching more importance to high-quality samples. (3) All

design teams applied different approaches to protect the privacy of users. For example, teams 1 and 2

allowed training, prediction, and so on to be performed locally without connecting to the Internet.

9.4.9. Control

Participants came up with different strategies to balance the control exerted by the ML application

and the user over the products. (1) Team 2 listed the recognized subjects on a mobile app and

enabled users to remove subjects that should no longer be recognized. (2) Team 5 offered different

modes for children of different ages. The complexity of the voice interaction will increase as the child

grows up. (3) Team 6 offered a seamless browsing mode, allowing users to remove some behavior

data from the training dataset so that the ML system will not be misled.

9.4.10. Anthropomorphism

The participants explored various anthropomorphic styles to avoid the “uncanny valley ”while

providing a pleasant impression. (1) Team 1 increased the anthropomorphism of their robot by

using body language and voice characteristics instead of facial expressions, thus avoiding the

“uncanny valley. ”(2) In contrast, team 5 was cautious about anthropomorphism. They customized

certain wake words, but reduced the anthropomorphism. This avoids cognitive problems in children

who should not be over-dependent on such products. (3) Team 2 retained the non-anthropomorphic

appearance of Clips because it clearly makes itself known as a camera. To some extent, this protects

users ’privacy.

9.4.11. Interactivity

Participants tried to make the interaction smooth, and thus provided improved UX and promoted

the users ’contribution to the growth of ML. (1) Team 6 simplified the process of collecting user

performances. For example, the option for the user to specify that they “do not like this recommen-

dation ”was very easy to select. To further enhance reliability, users can easily state why they do not

like a particular recommendation. (2) To reduce the computation time used for prediction, team 5

decreased the complexity of the voice content, e.g., setting modal words with no special meaning to

respond to 1 –3-year-old children and to encourage them to keep talking. (3) Team 1 used an ML

accelerator called a neural computing stick to decrease the computation time, as all the computation

is performed within the local hardware.

HUMAN –COMPUTER INTERACTION 381 10. Discussion and future work

The proposed MLT method focuses on the malleable nature of ML-empowered UX design and

enables designers to participate in the development of the material. During the design process of

Canvas, we obtained several insights and lessons that may inspire further research into the design

methodology and tools unifying AI and UX.

10.1. Contribution to the integration of AI and UX design

In this paper, we have described the characteristics of ML from the perspective of conceptual design

and summarized the design challenges caused by the growable nature of ML. This might provide

a foundation for follow-up UX research that considers AI technology as a growable “tree ”instead of

static “wood. ”

The MLT method encourages designers to reflect on the whole lifecycle of a malleable material

like ML and cultivate it into the desired form. As an extension of existing design methods, MLT

provides specific features for human-AI interaction design. In contrast with Human-centered

Design, MLT pays equal attention to factors besides human users, especially the design material,

and supports co-creation among users, ML, and the scenario. Compared with Value-Sensitive Design

and participatory design methods, MLT focuses on the development process of the design material

and how to intervene in the process to gain the desired UX.

The Canvas provides a conceptual design template for applying MLT in growing ML-empowered

UX and developing design proposals. By discussing issues about ML, users, and the scenario in

a visual way, the interplay between them can be analyzed throughout the entire ML lifecycle. Our

evaluation result illustrates that Canvas performs better than Blueprint in terms of creativity and

understanding. However, we do not observe significant difference between the two design tools

regarding ease of use and process. The interviews reveals that Canvas introduces ML-related

concepts in a way that is consistent with design thinking, which is easy

understandable for designers. Besides, the forms and compositions of Canvas offer meaningful

creative opportunities for designers by revealing the malleable nature of ML.

Additionally, the Canvas framework can be extended to other malleable design materials,

enabling designers to fulfil the five stages of the MLT process. The Issue cards could be used to

describe any kinds of malleable material (see Figure 17a ).TherowsandcolumnsoftheCanvas

format can be used to present the co-creators and touchpoints of any other materials (see

Figure 17b,c ). Designers can insert information into each space representing the different co-

creators at specific touchpoints. Finally, desi gners could summarize the design insights from

collected information (see Figure 17d ). The extended Canvas could easily be transformed to be

applied to a similar technology.

10.2. Future work on improving MLT and Canvas

For now, MLT and Canvas do not violate the traditional design process; rather, they complement

original design methods because they deal with a kind of technology with malleable attributes that

Figure 17. The extended material lifecycle Canvas.

382 Z. ZHOU ET AL. has not been widely applied as a design material before. We chose to retain typical design activities

because designers prefer to learn about an unfamiliar material in a familiar design process. However,

the design thinking process might evolve as designers better understand the inner workings of ML

and prepare for working with ML. For instance, the prototyping process might disappear when AI-

empowered products can iterate themselves to form better versions.

The evaluation of MLT and Canvas was conducted in a university, but no implementation in

a working environment has yet been investigated. Although instructors played the role of technical

teams, the design students lack experiences (e.g., cooperating with engineers, applying other emer-

ging technology in design) that might facilitate considerations of technical feasibility and thus

potentially influence the evaluation result. Besides, the construction of evaluation and recruitment

of participants may inevitably bring some biases including the educational setting of the evaluation,

order effects, and etc. Future work should eliminate the possibility of order effects and other biases.

We would also try to gain feedback in an industrial environment.

The current Canvas platform has several limitations and requires further improvements. (1) The

Question list may constrain the creative design activities, as it cannot cover all possible questions and

inevitably presupposes some relationships among the three co-creators. Future practice might examine

a way of providing sufficient guidance without implying these relationships. (2) Canvas should not be

confined to designers, as it could also serve as a bridge for connecting experts in different disciplines.

For example, Canvas might help engineers learn about design thinking, or offer a way for product

managers to gain a comprehensive view of the lifecycle of ML-empowered products, thus improving

communication efficiency and promoting cooperation. However, experts without a design background

might require more time to learn Canvas before using it in their tasks. (3) Within the scope of our

current research, Canvas is for the conceptual phase and cannot really be used in prototyping, such as

helping designers to implement the ML systems. We look forward to extending Canvas to a toolbox that

is not constrained to the conceptual phase.

In the future, we would like MLT to be used in various industries and be used by experts in

different disciplines. This would provide us with insights that motivate the further unification of UX

and AI and generate new design guidelines or principles in this filed.

11. Conclusion

In this paper, we have introduced MLT as a means of helping designers, especially novice designers, to

explore and appreciate the malleable nature of ML, and encouraging them to consider the ML system,

users, and scenario as co-creators of ML-empowered UX. The ML Lifecycle Canvas, a visual schematic

for applying MLT, was developed in an iterative design process involving designers. Canvas visually

represents three co-creators (ML, user, and scenario) and outlines notable questions within the ML

lifecycle to enable the identification of crucial design possibilities. Through a comparison with Service

Blueprint, we found that our approach guides novice designers to tackle the identified UX issues

throughout the whole ML lifecycle and gain an increased understanding of the co-creators of ML-

empowered UX. Though our work is at an early stage, it takes us one step closer to unifying UX and ML.

We expect that the MLT method and Canvas will be improved through future use in design practice.

Funding

Project supported by the National Natural Science Foundation of China (No. 61672451), the China Postdoctoral

Science Foundation (2018M630658), Zhejiang Provincial Key Research and Development Plan of Zhejiang Province (No. 2019C03137), and the National Science and Technology Innovation 2030 Major Project of the Ministry of Scienceand Technology of China (2018AAA0100703).

HUMAN –COMPUTER INTERACTION 383 Notes on contributor

Lingyun Sun is a Professor of Design at Zhejiang University. He ’scurrently a deputy director of International Design Institute. He has interdisciplinary research experiences, including artificial intelligence, computer graphics, design

cognition, interaction design, ergonomics and drama. He is taking part in the MIT-SUTD Teach-the-Teacher program and teaching courses regularly at SUTD. He is the author or coauthor of over 20 journal or conference papers on

Industrial Design, Computer Aided Sketching and Interaction Design. He holds patents on inventions relating to interaction and sketching technology. He is a member of the Industrial Designers Society of America, a member of thecouncil of Chinese Mechanical Engineering Society Industrial Design Institution, a member of China Computer

Federation.

References

Abadi, M., Barham, P., Chen, J., Chen, Z., Davis, A., Dean, J., Devin M, Ghemawat S, Irving G, Isard, M., & Kudlur M

(2016 ).Tensorflow: A system for large-scale machine learning [Paper presentation], Savannah, GA, USA. OSDI. Amershi, S., Cakmak, M., Knox, W. B., & Kulesza, T. ( 2014 ). Power to the people: The role of humans in interactive machine learning. AI Magazine ,35 (4), 105 –120. https://doi.org/10.1609/aimag.v35i4.2513 .

Amershi, S., Weld, D., Vorvoreanu, M., Fourney, A., Nushi, B., Collisson, P.,Suh J, Iqbal S, Bennett PN, Inkpen K, &Teevan J. ( 2019 ).Guidelines for human-AI interaction [Paper presentation]. Proceedings of the 2019 CHI Conference on Human Factors in Computing Systems, Glasgow, Scotland, Uk.

Atasoy, B., & Martens, J.-B. ( 2011 ).Crafting user experiences by incorporating dramaturgical techniques of storytelling [Paper presentation]. Procedings of the Second Conference on Creativity and Innovation in Design, Eindhoven,

Netherlands.Atasoy, B., & Martens, J.-B. ( 2016 ). STORYPLY: Designing for user experiences using storycraft. In Collaboration in creative design (pp. 181 –210). Springer.

Baumer, E. P. ( 2017 ). Toward human-centered algorithm design. Big Data & Society ,4(2), 2053951717718854. https:// doi.org/10.1177/2053951717718854 .

Baumer, E. P. S., Ahn, J., Bie, M., Bonsignore, E. M., B, A., & Yip, J. ( 2014 ).CHI 2039: Speculative research visions [Paper presentation]. CHI ‘14 Extended Abstracts on Human Factors in Computing Systems, Toronto, Ontario, Canada.

Bitner, M. J., Ostrom, A. L., & Morgan, F. N. ( 2008 ). Service blueprinting: A practical technique for service innovation. California Management Review ,50(3), 66 –94. https://doi.org/10.2307/41166446

Bucher, T. ( 2017 ). The algorithmic imaginary: exploring the ordinary affects of facebook algorithms. Information, Communication & Society ,20(1), 30-44. doi: 10.1080/1369118X.2016.1154086 Cuevas, H. M., Fiore, S. M., Caldwell, B. S., & Laura, S. (2007). Augmenting team cognition in human-automation teams

performing in complex operational environments. Aviation Space & Environmental Medicine ,78(5), B63 –B70. https:// www.ingentaconnect.com/content/asma/asem/2007/00000078/A00105s1/art00010.De Choudhury, M., & Kiciman, E. ( 2018 ). integrating artificial and human intelligence in complex, sensitive problem

domains: experiences from mental health. AI Magazine ,39(3), 69 –80. https://doi.org/10.1609/aimag.v39i3.2815 Deng, L., & Li, X. ( 2013 ). Machine learning paradigms for speech recognition: An overview. IEEE Transactions on

Audio, Speech, and Language Processing ,21(5), 1060 –1089. https://doi.org/10.1109/TASL.2013.2244083 Dove, G., Halskov, K., Forlizzi, J., & Zimmerman, J. ( 2017 ).UX design innovation: challenges for working with machine learning as a design material [Paper presentation]. Proceedings of the 2017 CHI Conference on Human Factors in

Computing Systems, Denver, Colorado, USA.Dunne, A., & Raby, F. ( 2013 ).Speculative everything: Design. Fiction, and social. The MIT press.

Fiebrink, R., & Cook, P. R. ( 2010 ).The Wekinator: A system for real-time, interactive machine learning in music [Paper presentation]. Proceedings of The Eleventh International Society for Music Information Retrieval Conference (ISMIR 2010), Utrecht.

Forlizzi, J., Zimmerman, J., Mancuso, V., & Kwak, S. ( 2007 ).How interface agents affect interaction between humans and computers [Paper presentation]. Proceedings of the 2007 conference on Designing pleasurable products and

interfaces, Helsinki, Finland.Gillies, M., Fiebrink, R., Garcia, A. T., Bevilacqua, F., Heloir, A., Nunnari, F., Mackay W, Amershi S, Lee B, &d’Alessandro N. ( 2016 ).Human-centred machine learning [Paper presentation]. Proceedings of the 2016 CHI

Conference Extended Abstracts on Human Factors in Computing Systems, San Jose, California, USA.Gillies, M., Fiebrink, R., Tanaka, A., Garcia, J., Bevilacqua, F., Heloir, A., Nunnari, F., Mackay W, Amershi S, Lee B, &d’Alessandro N. ( 2016 ).Human-centred machine learning [Paper presentation]. Proceedings of the 2016 CHI

Conference Extended Abstracts on Human Factors in Computing Systems, San Jose, California, USA.Google. ( 2019a ).AI & machine learning products: Cloud AutoML . Retrieved from https://cloud.google.com/automl/

Google. ( 2019b ).Do-it-yourself artificial intelligence .https://aiyprojects.withgoogle.com/

384 Z. ZHOU ET AL. Google. (2019c).People+ AI guidebook: Designing human-centered AI products .https://pair.withgoogle.com/intro/

Hancock, P. A., & Chignell, M. H. (1988). Mental workload dynamics in adaptive interface design. IEEE Transactions

on Systems Man & Cybernetics ,18 (4), 647 –658.https://doi.org/10.1109/21.17382.

Hand, D. J. ( 1998). Data mining: Statistics and more? The American Statistician,52(2), 112 –118. doi: 10.1080/

00031305.1998.10480549.

Hanington, B., & Martin, B. ( 2012). Universal methods of design: 100 ways to research complex problems. Develop

Innovative Ideas, and Design Effective Solutions: Rockport Publishers . https://books.google.com/books/about/

Universal_Methods_of_Design.html?id=uZ8uzWAcdxEC

Hebron, P. ( 2016).Machine learning for designers .O’Reilly Media.

Holmquist, L. E. ( 2017). Intelligence on tap: Artificial intelligence as a new design material. Interactions,24 (4), 28–33.

https://doi.org/10.1145/3085571 .

Howitt, D, & Cramer, D. ( 2010). Introduction to qualitative methods in psychology (vol. In 1). UK: Pearson

Education.

IBM. ( 2019).AutoAI with IBM watson studio .https://www.ibm.com/cloud/watson-studio/autoai

Jacobs, J., Gogia, S., Radom, M., & Brandt, J. R. ( 2017).Supporting expressive procedural art creation through direct

manipulation [Paper presentation]. Proceedings of the 2017 CHI Conference on Human Factors in Computing

Systems, Denver, Colorado, USA.

Jess, H. ( 2017).Human-centered machine learning .https://medium.com/google-design/human-centered-machine-

learning-a770d10562cd

Karana, E., Barati, B., Rognoli, V., Der Laan, V., & Zeeuw, A. (2015 ).Material driven design (MDD): A method to

design for material experiences.

Kim, J., & Ryu, H. ( 2014). A design thinking rationality framework: Framing and solving design problems in early

concept generation. Human–Computer Interaction ,29(5 –6), 516– 553.https://doi.org/10.1080/07370024.2014.

896706

Koza, J. R., Bennett, F. H., Andre, D., & Keane, M. A. ( 1996). Automated design of both the topology and sizing of

analog

electrical circuits using genetic programming. In Artificial intelligence in design’96 (pp. 151 –170). Springer.

Kruschke, J. K. ( 2011).Doing Bayesian data analysis: A tutorial with R and BUGS . Academic Press/Elsevier.

Kuhn, S., & Muller, M. J. ( 1993). Participatory design. Communications of the ACM ,36(6), 24–29. https://doi.org/10.

1145/153571.255960

Kuniavsky, M., & Churchill, E. ( 2017).Designing the user experience of machine learning systems SS-17-04 .

Li, Y.-M., & Yeh, Y.-S. (2010). Increasing trust in mobile commerce through design aesthetics. Computers in Human

Behavior ,26 (4), 673 –684.https://doi.org/10.1016/j.chb.2010.01.004 .

McCardle, J. R. ( 2002). The challenge of integrating AI & smart technology in design education. International Journal

of Technology & Design Education ,12 (1), 59 –76.https://doi.org/10.1023/A:101308940 .

McElreath, R. ( 2018).Statistical rethinking: A Bayesian course with examples in R and Stan . Chapman and Hall/CRC.

Merritt, S. M., Heather, H., Jennifer, L. C., & Deborah, L. (2013). I trust it, but I don ’t know why: Effects of implicit

attitudes toward automation on trust in an automated system. Human Factors,55 (3), 520 –534.https://doi.org/10.

1177/0018720812465081.

Mitchell, T. M. ( 1999). Machine learning and data mining. Communications of the ACM,42(11), 30–36. https://doi.

org/10.1145/319382.319388

Muller, M. J. ( 2007). Participatory design: The third space in HCI. In The human-computer interaction handbook(pp.

1087 –1108). CRC press.

Murnane, E. L., & Counts, S. (2014 ).Unraveling abstinence and relapse: Smoking cessation reflected in social media

[Paper presentation]. Proceedings of the SIGCHI Conference on Human Factors in Computing Systems, Toronto,

Ontario, Canada.

Patel, K., Fogarty, J., Landay, J. A., & Harrison, B. (2008 ).Examining difficulties software developers encounter in the

adoption of statistical machine learning [Paper presentation]. Proceedings of the 23rd national conference on

Artificial intelligence - Volume 3, Chicago, Illinois.

Pruitt, J., & Adlin, T. (2010). The

persona lifecycle: Keeping people in mind throughout product design . Elsevier.

Recasens, M., Danescu-Niculescu-Mizil, C., & Jurafsky, D. (2013 ).Linguistic models for analyzing and detecting biased

language [Paper presentation]. Proceedings of the 51st Annual Meeting of the Association for Computational

Linguistics (Volume 1: Long Papers). https://doi.org/10.1016/S2213-2600(13)70187-5

Reynolds, T. J, & Gutman, J. J. ( 1998). Laddering theory, method, analysis, and interpretation. Journal of Advertising

Research ,28(1), 11-31.

Ribeiro, M., Grolinger, K., & Capretz, M. A. M. ( 2016).MLaaS: machine learning as a service [Paper presentation].

IEEE International Conference on Machine Learning and Applications. Miami, FL, USA, the DOI is 10.1109/

ICMLA.2015.152.

Rozendaal, M. C., Ghajargar, M., Pasman, G., & Wiberg, M. ( 2018). Giving form to smart objects: Exploring

intelligence as an interaction design material. In M. Filimowicz & V. Tzankova (Eds.), New directions in third

wave human-computer interaction: Volume 1 - technologies (pp. 25–42). Springer International Publishing.

HUMAN–COMPUTER INTERACTION 385 Russell, A., Ekvall, T., & Baumann, H. (2005). Life cycle assessment introduction and overview. Journal of Cleaner

Production, 13(13–14), 1207 –1210.c

Schon, D. A., & Desanctis, V. (2005 ). The reflective practitioner: How professionals think in action. Proceedings of the

IEEE , 73(4), 845 –846. https://books.google.com/books?hl=zh-CN&lr=&id=ceJIWay4-

jgC&oi=fnd&pg=PR7&dq=The+reflective+practitioner:+How+professionals+thin\

k+in

+action&ots=q88QS_DWum&sig=LrVDHtR4asDzR7DiTkl_cXJElk4#v=onepage&q=The%2\

0reflective%20practi-

tioner%3A%20How%20professionals%20think%20in%20action&f=false.

Snyder, J., Baumer, E. P., Voida, S., Adams, P., Halpern, M., Choudhury, T., & Gay, G. (2014 ). Making things visible:

Opportunities and tensions in visual approaches for design research and practice. Human–Computer Interaction ,29

(5 –6), 451 –486.c

Stickdorn, M., Schneider, J., Andrews, K., & Lawrence, A. (2011). This is service design thinking: Basics, tools, cases

(Vol. 1). Wiley Hoboken.

Tschimmel, K. ( 2012).Design thinking as an effective toolkit for innovation [Paper presentation]. ISPIM Conference

Proceedings, Barcelona, Spain. https://doi.org/10.1094/PDIS-11-11-0999-PDN

Vlist, B., Westelaken, R., Bartneck, C., Hu, J., Ahn, R., Barakova, E., Delbressine F, & Feijs, L. (2008 ).Teaching

machine learning to design students [Paper presentation]. Proceedings of the 3rd international conference on

Technologies for E-Learning and Digital Entertainment, Nanjing, China.

Wang, D., Yang, Q., Abdul, A., & Lim, B. Y. (2019 ).Designing theory-driven user-centric explainable AI [Paper

presentation]. Proceedings of the 2019 CHI Conference on Human Factors in Computing Systems, Glasgow,

Scotland, Uk.

Witten, I. H., Frank, E., Hall, M. A., & Pal, C. J. ( 2016).Data mining: Practical machine learning tools and techniques .

Morgan Kaufmann.

Wood, M. ( 2014).Hopes to serve up music in a novel way .http://nyti.ms/1fr8LY5

Yang, Q. ( 2017).The role of design in creating machine-learning-enhanced user experience [Paper presentation]. 2017

AAAI Spring Symposium Series on Designing the User Experience of Machine Learning Systems, Stanford, CA,

USA.

Yang, Q., Banovic, N., & Zimmerman, J. (2018 ).Mapping machine learning advances from HCI research to reveal

starting places for design innovation [Paper presentation]. Proceedings of the 2018 CHI Conference on Human

Factors in Computing Systems.

Yang, Q., Scuito, A., Zimmerman, J., Forlizzi, J., & Steinfeld, A. (2018 ).Investigating how experienced ux designers

effectively work with machine learning [Paper presentation]. Proceedings of the 2018 Designing Interactive Systems

Conference, Hong Kong, China.

Yang, Q., Suh, J., Chen, N.-C., & Ramos, G. ( 2018).Grounding interactive machine learning tool design in how

non-experts actually build models [Paper presentation]. Proceedings of the 2018 Designing Interactive Systems

Conference,

Hong Kong, China.

Zhou, Z., Gong, Q., Qi, Z., & Sun, L. ( 2019).ML-process Canvas: A design tool to support the UX design of machine

learning-empowered products [Paper presentation]. Extended Abstracts of the 2019 CHI Conference on Human

Factors in Computing Systems, Glasgow, Scotland, Uk.

Zhou, Z., Jiang, H., Yang, C., Yang, J., Loy, Y. Y. W., & Sun, L. (2019 ).QianLi: A modular system for connecting distant

family members through tacit interaction [Paper presentation]. International Conference on Human-Computer

Interaction, HCII 2019, Orlando, Florida, USA.

Zhu, H., Yu, B., Halfaker, A., & Terveen, L. ( 2018). Value-sensitive algorithm design: Method, case study, and lessons.

Proceedings of the ACM on Human-Computer Interaction ,2 (CSCW), 1–23. https://doi.org/10.1145/3274463

386 Z. ZHOU ET AL.