Growth/Personalized first day/Welcome survey/ko

이 문서는 성장 팀의 "환영 설문"에 대한 설명입니다. 이 프로젝트는 "개인화된 첫 날" 이라는 작업의 일부입니다. 이 문서에는 주요 요소, 디자인, 결정 사항을 나열하고 있습니다. 대부분의 순차적인 업데이트는 성장 팀 업데이트에 게시되며, 대규모 업데이트가 이곳에 기록됩니다.

2018년 11월 20일에 환영 설문이 한국어 위키백과와 체코어 위키백과의 일부 사용자에게 배포되었으며, 2018년 12월에 최초 보고서를 통해 데이터에 대한 분석과 가공이 이루어졌습니다. By March 2019, we had concluded that the design called "Variation A" is best, and that is now shown to all new users in Czech and Korean Wikipedias. 어떤 설문 변형이 베트남어 위키백과에 적합한지에 대해 알아보고 있습니다.

현재 상태

 * 2018-11-20: 한국어 위키백과와 체코어 위키백과에 환영 설문이 배포되었습니다. 사용자의 절반은 변형 A를 보고, 나머지 절반은 통제 그룹입니다.
 * 2018-12-22: 가 게시되었습니다.
 * 2019-01-16: 한 후, 한국어와 체코어 위키백과의 새 사용자의 절반은 변형 A를 보고 나머지 반은 변형 C를 보도록 변경하였습니다.
 * 2019-01-24: 베트남어 위키백과에 환영 설문이 배포되었습니다. 절반의 사용자는 변형 A를 보고, 나머지 절반은 변형 C를 봅니다.
 * 2019-02-27: 베트남어 위키백과의 설정을 변경하여 절반의 사용자가 변형 A를 보고, 나머지는 설문을 받지 않도록 변경했습니다. 최초 실험과 비교한 진행 포기율을 추적하기 위한 조치였습니다.
 * 2019-03-06: 체코어와 한국어 위키백과에서의 실험을 분석한 후, 모든 시나리오와 모든 지표상에서 더 나은 성적을 보인 변형 A를 모든 사용자에게 보여주기로 결정했습니다. 한국어와 체코어 위키백과에서 진행중인 실험이 없습니다.
 * 2019-03-21: 베트남어 위키백과에서의 실험이 설문을 받지 않는 사용자보다 변형 A 설문을 받는 사용자가 더 많이 편집을 포기하는 것을 확인한 후 중단되었습니다. 이 결과를 조사 중입니다.
 * 다음: 베트남어 위키백과의 편집 포기율을 조사합니다.

요약
새 사용자 경험 연구를 통해 새 사용자는 특정한 목적을 가지고 위키에 방문한다는 것을 알 수 있습니다. 이 사용자들이 이러한 목적을 달성하지 못한다면, 돌아오지 않을 것임을 알 수 있습니다. 따라서 환영 설문을 통해 새 사용자에게 그들의 목적에 대해 물어본 후, 그 목적을 달성하도록 도울 수 있도록 지원합니다. 새 사용자가 계정을 만들 때, 왜 계정을 만드는지, 무엇을 이루고자 하는지, 관심 있는 주제라던지, 멘토를 만나고 싶은지 여부를 물어보고자 합니다.

이 프로젝트의 동기에 대해서는 "개인화된 첫 날" 작업 문서를 참조하세요.

환영 설문의 데이터가 어떻게 새 사용자의 초기 단계를 개인화하는 데 사용되는지 알아보려면 "참여 메일"과 "새 사용자 홈페이지" 프로젝트를 참조하세요.

Comparative review
To learn how best to design the welcome survey, our team's designer reviewed the way that other platforms (e.g. Reddit, Coursera, Medium) ask initial questions of newcomers. While the experience we want to give newcomers is definitely different than other platforms (we want to give newcomers an optional, lightweight, non-invasive experience), we also recognize that there are best practices we can learn from other software. The comparisons are shown in this slide deck, and the main takeaways are:


 * Modals were generally more in use or otherwise a full page overlay for forms.
 * No more than 4 questions were asked during onboarding.
 * Questions were either progressively disclosed on a single screen or otherwise shown 1 question per step with clear visual indication of the number of steps (eg. via a stepper UI).
 * Multi-question forms tended to be skip-able entirely, but with mandatory questions if the user wanted to participate.
 * An optional guided tour or further help was often provided afterwards once user was in the app.
 * Friendly voice and tone - welcoming, informal language framed as asking for information to help orient the user ("Tell us about yourself", "Get recommendations", etc.)

Design
Our evolving designs can always be found in these clickable mockups, and with additional contextual information in this Phabricator task. Our goal for the design is for newcomers to answer all the questions they want to, and to not depress the number of users who complete their signup process and get back to what they were doing.

We considered three main design approaches for this project. The new account holder lands at each approach immediately after clicking "Create your account" at Special:CreateAccount. And after completing each one, they are returned to the page from which they originally clicked to create their account. Please note that the question wording and text in these mockups does not reflect current drafts -- the mockups are more for the visuals and workflows. Question wording and text will be posted separately.


 * Variation A: a Special page containing the survey. This was the simplest to implement, and we deployed this one first on November 20, 2018.  Though this doesn't follow best practices that we learned in comparative review, it is the fastest way to get a survey up and to start learning.  It will also work automatically on mobile web and for no-Javascript users.
 * Variation B: this option shows users one question at a time, and also removes the surrounding links from the left and top navigation of the wiki for a more streamlined experience. This is inspired by the interface for the Content Translation tool.  We decided not to pursue this variation at all, since it is more work than Variation C, which we like better.
 * Variation C: this is our ideal design, which is a modal over the context where the user came from.  It helps make it clear that user is close to returning to the task they were doing when they started creating an account. It also presents questions one at a time so the survey is not overwhelming.  This incorporates our learning from the comparative review.  We will have to design and engineer a separate mobile web version, as well as consider what no-Javascript users will receive.  This is being engineered during November and December 2018.

The team built and deployed Variation A in November 2018, and simultaneously built Variation C. We will be deploying Variation C during January 2019.

Questions, responses, and other text
The questions that this feature asks to newcomers, along with their specific wording and the other text in the feature, is really important to getting useful data. The team worked on these things with help from several other people experienced with surveys at WMF. All the questions will be optional, and will be translated into the languages of the wikis on which the welcome survey is deployed. Below are the questions we want to ask, and we are still working on the wording. To see the current text of the feature, check out the current mockups.


 * Why did you create your account today?
 * Goal: if we understand what a newcomer is trying to accomplish, we may be able to show them help materials that help them accomplish it.
 * Response options:
 * To fix a typo or error in a Wikipedia article
 * To add information to a Wikipedia article
 * To create a new Wikipedia article
 * To read Wikipedia
 * Other (please describe)
 * Have you ever edited Wikipedia?
 * Goal: we want to know how many people creating accounts may have already made IP edits. This would also allow us to refrain from giving too much help material to people who know what they're doing already.
 * Response options:
 * Yes, many times
 * Yes, once or twice
 * No, I didn't know I could edit Wikipedia
 * No, for other reasons
 * I don't remember
 * People can edit Wikipedia articles on topics they care about. We've listed a few topics below that are popular for editing.  Select some topics that you may wish to edit.
 * Goal: it may be possible in the future to connect newcomers with experienced editors who share their interests, or just give them recommendations on articles to work on.
 * Response options (see this update for explanation of how these were determined):
 * As checkboxes: Arts, Science, Geography, History, Music, Sports, Literature, Religion, Popular culture
 * Behind typeahead: Entertainment, Food and drink, Biography, Military, Economics, Technology, Film, Philosophy, Business, Politics, Government, Engineering, Crafts and hobbies, Games, Health, Social science, Transportation, Education
 * It will also possible for users to add their own topics.
 * Email address
 * Goal: Only if the user did not supply it on Special:CreateAccount. Email is important for engaging new editors and allowing them to recover their accounts, so we want to ask twice.
 * We are considering starting a program for more experienced editors to help newer users with editing. Are you interested in being contacted to get help with editing?
 * Goal: We want to learn whether newcomers feel that they need human-to-human help.

In addition to the questions themselves, we will be showing users a link to the survey privacy statement (translated into the wiki's language) and to the Tutorial and Help Desk pages on their wikis.

User testing
During the week of October 22, 2018, we used usertesting.com to conduct six tests of our Variation A mockups with internet users unaffiliated with the Wikimedia movement. In these tests, respondents are compensated for trying out the mockups, speaking aloud on what they observe, and answering questions about the experience. As our team's designer described on the Phabricator task, the goals of this testing were:


 * 1) Identify improvements to Understanding of the survey (update copy depending on users' comprehension of the instructions and questions on the form).
 * 2) Identify improvements to the Usability of the survey (check whether users are able to correctly input and submit their intended responses).
 * 3) * Do users know how the visual design and layout works? (Eg., do they understand they should be selecting one answer from a radio button group)
 * 4) * Are users able to navigate through the form?
 * 5) * Any feedback missing that could help users complete the form?
 * 6) Gauge user Reactions to the survey and Expectations of how the information will be used.

Summary of findings


 * Survey was clearly optional, and seen as short, low-effort, and non-intrusive to complete
 * Generally seen as data capture for research (though a few testers thought more specific explanation about how responses may be used might make them more inclined to fill it in)
 * "Mentorship program" was the one aspect not clearly understood, with about half of participants mistakenly assuming they would be providing the help rather than receiving it.
 * A couple of users had misgivings about providing email and feared it may be misused for marketing or given to 3rd parties (also as they expected it to be mandatory for account creation if it was needed for recovery)
 * Users liked the post-submission message with more information about "Getting started with editing"
 * Several users assumed that upon completing the welcome survey, they would be directed to their "dashboard" to get started with editing.

Recommendations


 * Add a full Thanks message after survey completion
 * Rephrase “Mentor” checkbox question and clarify expectations for how users may be contacted if they select the mentor checkbox
 * Revise phrasing of Q3 (Wordsmithing so that the the second sentence “We've listed a few below popular for editing” is more easily understood.)
 * Add a tooltip beside or assistive text under the “Add more topics” field for no-js users explaining they can comma-separate multiple topics
 * Add more information about how the email is used
 * Potentially be extra clear by placing a tooltip with the following message (taken from Help:Account_management): “If you choose to give an email address, other users will be able to contact you by email. This feature is anonymous—the user who emails you will not know your email address. You don't have to give your email address if you don't want to (but doing so is required to reset your password if you forget it)”

Analysis and experiments
There are two kinds of analysis we'll be doing with respect to this feature. To read about those plans in-depth, see this page containing the "experiment plan".

Analysis of the responses to the survey
The main objective of the survey is to collect information about what new editors are trying to accomplish, so we can figure out if it's possible to personalize their experiences based on their responses. We are doing straightforward analysis on responses, broken out by elements like platform (mobile vs. desktop) and context from which the account was created (homepage, reading, editing).

The initial report on survey responses can be found here. Community members should feel free to translate this report into their languages. In-depth reporting will be created by March 2019. Some topline results:


 * The survey has high response rates: 67% in Czech and 62% in Korean.
 * Many people create accounts just to read articles -- a potential opportunity to engage new editors.
 * High numbers of people indicate they are interested in being contacted for help: 36% in Czech and 53% in Korean.

Analysis of the impact of the survey on activation
Although we would not have intended it, it is possible that the welcome survey could depress the number of newcomers who make edits. Community members first brought up this concern, saying that by making the sign-up process longer, and keeping newcomers away from their objectives, we could cause them to leave the site. Therefore, we conducted an A/B test to find out whether being presented with the survey decreases activation rate (the rate at which new users make their first edit).

From November 19 to January 15, half of newcomers in Czech and Korean Wikipedia were given Variation A of the survey, and half were given no survey. Our experiment results are published here, and show that having the survey does not significantly decrease the rate at which newcomers make their first edit. Based on this analysis, we are now testing Variation A against Variation C to see if Variation C has increased response rates. Results from this experiment are expected in February 2019. After this experiment is complete, we will determine whether Variation A or Variation C is the survey we should deploy to all newcomers in our target wikis going forward.

Analysis of the two survey variants
After seeing that Variation A did not cause a decrease in activation rate in Czech and Korean Wikipedias, we moved on to comparing Variation A to Variation C in those wikis. During the month of February 2019, half of newcomers received each variant. After comparing the variants on activation rate, abandonment rate, and response rate, we saw that Variation A performs better or equal to Variation C on all languages and all platforms. We therefore ended the experiments in Czech and Korean Wikipedias and started using Variation A with all newcomers starting on March 6, 2019. In-depth results are forthcoming.