Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

IA1: Testing & iteration #19706

Closed
7 tasks done
Tracked by #153
nadonomy opened this issue Nov 11, 2021 · 4 comments
Closed
7 tasks done
Tracked by #153

IA1: Testing & iteration #19706

nadonomy opened this issue Nov 11, 2021 · 4 comments
Assignees
Labels
T-Enhancement Z-IA Issues relating to information architecture Z-Labs

Comments

@nadonomy
Copy link
Contributor

nadonomy commented Nov 11, 2021

We'd like to test & research the Phase I IA experiments in chorus, to learn, iterate and gain confidence in changes to ship to production. We'll use the following methods:

1. In-app feedback (@kittykat )

  • Monitor in-app feedback
  • Triage, looking for common themes
  • Outreach to set up interviews with users who consented
  • In-app feedback project

2. Moderated tests, internal (@niquewoodhouse )

  • Put out a call to arms to folk on the core team to help with moderated user tests
  • Do interviews, for initial findings and to help refine research methods
  • Application form
  • Responses

3. Community tests (@niquewoodhouse with support from @kittykat )

  • Arrange community testing sessions to gather wider feedback
  • 15th Dec results are here
    We should translate findings into separate issues, or consider tracking small snags in the final build issue.

4. Organic feedback via community (@kittykat )

  • Once we have confidence in the above, broadcast to wider organic channels like Matrix rooms, TWIM, Twitter, etc

5. Unmoderated tests, external (@niquewoodhouse )

  • Generate ad hoc builds with all experiments enabled by default
  • Engage in paid user testing with relevant scripts

6. R&D PostHog (@kittykat )

Note: While this issue is public, most of the links in this issue are private and will not be made public, to preserve folks privacy.


Small iterations, based on testing

Some of the above testing has led us to identify some iterations to the experiments that will improve them for further testing.

@nadonomy nadonomy added T-Enhancement Z-IA Issues relating to information architecture Z-Delight labels Nov 11, 2021
@nadonomy
Copy link
Contributor Author

nadonomy commented Dec 2, 2021

Note to self: Grab conclusions from https://docs.google.com/document/d/1xGfZ0z6yXKAO-1vQTk8TxItZlCKhRvRlhsIc-a5KHyA/edit Done.

@kittykat
Copy link
Contributor

Community testing artefacts are here - permission to access is needed on a case by case basis

@nadonomy
Copy link
Contributor Author

@kittykat thx. I just updated the issue with a summary of work and included this in there.

On (1) In-app feedback, we're currently experimenting with using a GitHub project board so we can maintain context directly in GitHub. If you have any bright ideas around automation/adding labels based on inputs (e.g. if a user consents to be contacted) they would be greatly received!

@kittykat
Copy link
Contributor

We are done with testing and processing results for IA1, further testing will be part of IA2.

Metrics have been split out into their own epic outside of IA, so closing this issue.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
T-Enhancement Z-IA Issues relating to information architecture Z-Labs
Projects
None yet
Development

No branches or pull requests

3 participants