Edit Review Improvements/New filters for edit review/Research/pt-br

Essa página fornece uma visão geral da pesquisa de experiência de usuário interativa realizada para os novos recursos para revisão de edições. Cada etapa da pesquisa está esboçada e seus relatórios correspondentes estão juntos. Nossa meta é dar aos nossos usuários um olhar mais profundo sobre como a pesquisa é feita, como ela informa as decisões e a direção do produto e quais possibilidades já foram consideradas.

Visão geral
Os novos filtros para revisão de edições tornaram-se disponíveis nas mudanças recentes como recurso Beta em maio e lançadas em outubro de 2017. Esses filtros pretendem ajudar editores a encontrar edições de interesse para revisar. Alguns filtros fornecidos são totalmente novos, e alguns são baseados em novas habilidades de aprendizado de máquina.

Queríamos fornecer uma experiência aos usuários intuitiva e ao mesmo tempo poderosa para solucionar alguns antigos problemas. Trabalhar nessa área nos apresentou desafios únicos, onde somente com a pesquisa conseguiríamos saber sobre nossos usuários e como nossas soluções propostas funcionariam para os mesmos.

Sinta-se livre para examinar os detalhes nos sumários da pesquisa e nos relatórios abaixo.

Questões e descobertas principais da pesquisa

 * Questão 1: Os participantes acharam úteis os novos filtros e funções para a revisão de edições nas mudanças recentes?
 * O que aprendemos: Sim, a maioria dos participantes reagiram bem positivamente aos novos filtros e funções.
 * Como isso influenciou nas decisões: Devido aos comentários positivos, testamos filtros e funções adicionais, e fizemos ajustes interativos nos elementos já existentes baseando-se nos mesmos comentários.
 * Questão 2: Até que ponto os participantes descobriram, entenderam e usaram os novos filtros/funções nas mudanças recentes?
 * O que aprendemos: Cada etapa de testes obteve resultados diferentes, mas a cada etapa houveram comentários de avaliação e melhorias a serem feitas para facilitar a descoberta, o entendimento e, como correlação direta, o quanto os participantes pretenderam experimentar o Beta em seus fluxos de trabalho nas wikis.
 * Como isso influenciou nas decisões: Em casa estágio do processo de pesquisa, estivemos constantemente adicionando, removendo e editando ambos os elementos novos e já existentes das mudanças recentes, com base nas recomendações e comentários dos participantes.
 * Questão 3: Como o sucesso dos filtros/funções traduzem-se à lista de páginas vigiadas?
 * O que aprendemos: A coorte participante não expressou muito interesse em trazer essas funcionalidades à lista de páginas vigiadas, e indicou que não acharia útil essa noção.
 * Como isso influenciou nas decisões: Por ora, não estaremos trabalhando nisso, mas faremos pesquisas futuras, dado que a coorte era composta de usuários das mudanças recentes, não da lista de páginas vigiadas. Os novos filtros nela não serão expostos por padrão, permanecendo como recurso Beta para coletarmos mais comentários.

Desafios

 * Nosso objetivo principal é tornar as wikis e suas ferramentas mais intuitivas e úteis para todos, mas, nessas sessões, limitamos nossa lista de participantes para aqueles usuários (relativamente frequentes) das páginas de revisão de edições. Concluímos que a maioria dos nossos participantes trabalhavam com a Wikipédia inglesa, mas tivemos representação de outros idiomas da Wikipédia e da Wikiversidade.
 * Mudanças podem ser difíceis, independente de utilidade, intensão ou resultados. Embora qualquer mudança em particular nem sempre agrade a todos, geralmente visamos tomar o máximo número possível de decisões baseando-se nos comentários de nossos usuários e no que vai ajudar a maioria dos usuários.
 * A usabilidade, num escopo relativamente amplo, é um grande cometimento, especialmente quando elementos de pesquisa são combinados. Nossa meta é manter o visual, sentimento e utilidade ao passo que tomamos decisões, tanto dos detalhes como do geral.
 * Precisamos de mais participantes no geral! Se tiver interesse no espaço de Contribuidores e gostaria de participar das próximas pesquisas de estudo, entre em contato com Daisy Chen, informando suas áreas de interesse, e não esqueceremos de você!

What’s next

 * Although the results have been encouraging, our wikis support many different usecases and we want to hear from you how the new filters work for your activities. Feel free to share your experience on the project talk page.
 * Are you interested in the Contributors space? Would you like to participate in upcoming research studies and provide your feedback? If so, please contact Daisy Chen with your areas of interest and we will keep you in the loop!

Research Questions

 * Concept and Intent
 * Do users understand what the filter categories represent, specifically the Quality, Intent, and Experience filters?
 * Do subjects understand what the different ORES (Objective Revision Evaluation Service) levels mean?
 * Do users understand the difference between filtering and highlighting, and how they might use these singly or in combination?


 * Usability
 * Do users notice and/or understand the default filters?
 * Are users able to find the correct filters and use them in combination effectively?
 * Do users find and learn to effectively use the highlighting to isolate targets?
 * Can they manage and interpret the highlight colors?
 * Do they find the “ignore highlights” button and understand its use?


 * Effectiveness
 * How do users feel the new tools will affect their work and workflows?
 * What are users’ reactions regarding the new interface and its functionalities?
 * Do users feel any differently about the new interface as compared to the current Recent changes page?
 * Do they see that the new interface as being functionally neutral, inferior or superior?

Curious about the more nitty-gritty details on the research set-up? Take a look here.

Findings
Overall, the new Recent Changes filters features were well-received. Users generally find that the prototype Recent Changes are improvements over the current and a majority think they would incorporate these features into their work.

Though features can still use some clarification, refinement, and ongoing iteration/improvement, using them is pretty intuitive overall and even when there are user hesitations, is a learnable process.

Over the course of this first round of usability testing, two prototypes were used with the second prototype building on findings from the first prototype.

Next Steps
Discuss existing issues and corresponding recommendations for improvement. Prototype 2 is recommended to go to beta for further user feedback and iteration.

Research Questions
After introducing the new filters for edit review to positive feedback, the Collaboration team’s next iteration aimed to include support for advanced functions.

In these sessions, we utilized a prototype testing environment and lead users through a series of tasks. Through completing said tasks, we are able to evaluate if research participants were aware of the functionalities, found them intuitive to use, and whether they found utility in them.


 * What elements of the extended filters are most and least useful for users?
 * Do users notice the new capabilities?
 * If they find them, are they easy to understand and use?

New page elements/functionalities
 * Namespaces, tag, users, and categories filters
 * Filter set bookmarking
 * Display options (size and pagination of results)
 * Navigating and updating results

Read more details on the research planning and protocol here.

Findings
Generally, participants had positive responses to the new filters and functionalities. To different degrees/varying use cases, most of these new features have audiences ready to wield these tools.

However, there are some areas in the prototype that can benefit from additional clarifications, UI tweaks, and workflow considerations. The details on user recommendations can be found in the research report.

Next Steps
Discuss existing issues and corresponding recommendations for improvement.

If possible, integrate changes into beta feature to allow time for users to experience them. After period of use, users of the beta should be prompted for information on how much the feature was utilized day-to-day, what specific elements/functionalities they used, and their feedback and recommendations for how the beta has worked for them and where areas for improvement remain.

Research Questions
The new filters for edit review have been out as a beta feature on various wikis. Our goal is to learn from users directly how the experience of using the new filters over a period of time has been for them.

Since the initial rounds of generative and usability testing, the team made some tweaks from past usability testing and feature requests that were then applied to a new prototype, named ‘integrated’ filters, along with a set of filters made specifically for the functionality on watchlist. In this round of testing, we evaluated the usability of the latest iteration, the applicability of its watchlist mirror, and the live updates feature in conjunction with the beta satisfaction testing.


 * Rating users’ satisfaction with the Recent Changes beta:
 * What do users think of the filters? Have their opinions changed over time?
 * Have they put the filters to use? If so, how? What are some of these uses?
 * Are the filters supporting their existing activities and workflows?
 * Are they creating new ‘existing activities/workflows’ as a result of having the additional capabilities?
 * How do they feel about the beta Recent Changes vs current Recent Changes?
 * More specifically, how do the new set of filters compare with the current set?
 * What does the page need to contain for it to reach peak effectiveness for the users?


 * Integrated filters/live updates usability:
 * How are discoverability of the filters and general navigation for users? Overall, has there been improvement?
 * What reactions do users have to the live updates button and pagination options?
 * Do users know about / recall / use the resources listed in the related links section?


 * Watchlist actions usability:
 * What actions do users typically take on their watchlists?
 * What filters would be useful? How do they feel about the watchlist filter set on the prototype? What filters are still missing?


 * Is the purpose of the watchlist-related filters provided in the Watchlist page understood, discoverable, and useful for watchlist users.
 * Do users create the same filter sets in the Watchlist they would for Recent Changes?
 * Are users able to identify which pages require their attention (are not visited) at a glance and why the distinction is made? Is it clear for users how to mark some/all of them as visited?

Read more research project details here.

Findings

 * User Satisfaction with the new filters beta:

Overall, the users feel that the beta feature provides a cleaner interface and a better experience than the current RC page. Though not without caveats; there are user suggestions for improvement, and adoption and use of the Content Quality and User Intent filter categories does not appear widespread. Improvements should be made to clarify the filters’ function, means of usage, and reliability, and other user concerns and suggestions should be evaluated and addressed.

The primary applications of the filters are related to vandalism patrol, and users are generally able to achieve their intended goals using the beta page.


 * Integrated filters/live updates usability:

Discoverability, general navigation, pagination Live updates Related links
 * Testing revealed no issues involving the discoverability of the dropdown menu
 * Though participants successfully complete tasks involving ‘advanced’ filters (users, tags, namespaces, categories), their initial discoverability on the page is still lacking
 * Pagination does not appear to cause much confusion for users even when testing an atypical use case (sorting oldest first), although most do not process the need for that functionality on the page.
 * All 5 participants had overall positive reactions to the introduction of this feature, with some minor concerns and suggestions for improvement.
 * All 5 participants use the options provided in the related links to some degree.
 * Mixed reaction as to whether the section should be hidden/collapsible.


 * Watchlist actions usability:

Among this cohort, the number of articles in their watchlists varied (along with the reasons for adding them) but their use of watchlist is rather consistent; they don’t use the page much if at all and if they do, it is not rigorous and more of a casual scanning for new activity. The recruiting of this cohort, however, was based on their use of Recent Changes page, so they may not be representative of the opinions of heavy watchlist users.

A majority of participants indicate that they do not and would not use filters on their watchlist page, consistent with responses when presented with the hypothetical posed during beta satisfaction testing.

All participants understood the bold/filled bullet represented a page with new activity/not yet visited, and it's an expectation that clicking on any part of the entry (diff, hist, or article link) should mark the entry as ‘read’.

Next Steps
Discuss existing issues and corresponding recommendations for improvement (this will always be a recurring element of our work).

Discuss the utility of further examining the idea of bringing the edit review filters to pages like watchlist, but with more seasoned/heavier users of watchlist pages specifically.