New questions from the Postmortem Article #422
Unanswered
Ultrabenosaurus
asked this question in
Community calls
Replies: 2 comments 4 replies
-
@Ultrabenosaurus Would you mind putting your questions in an enumerated list, to make it easier to respond, and put only keywords in bold, rather than the entire questions, to make them easier to read? 🙏 |
Beta Was this translation helpful? Give feedback.
3 replies
-
@caugner @LeoMcA @Rumyra when will these questions be addressed? |
Beta Was this translation helpful? Give feedback.
1 reply
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
From reading the postmortem article a few things stood out to me, which I feel may justify new questions or should absolutely be referenced when answering existing questions on these topics.
Due to how close we are to the community call I will apologise for being brief in order to give MDN / Mozilla as much time as possible to review this. I will put a quote from the postmortem (not necessarily in the same order as in the article) followed by a question in bold, and then consolidate all questions at the end.
1. Why are MDN using peer pressure as a justification for intentionally introducing LLMs, known to generate only plausible content not true and accurate information, to their "best-in-class" technical documentation?
2. What do MDN propose makes their documentation "best-in-class" if not accuracy and being approachable to readers?
3. Why are MDN claiming that intentionally giving an incorrect response to novice readers of "best-in-class" technical documentation is a good thing?
4. Have MDN considered the impact to their reputation as well as the impact to novice developer's time, project success, and frustration / mental health from dealing with incorrect responses given intentionally by MDN?
See also: a great summary of the likely novice user experience by @kyanha on #9230.
I acknowledge this states "one method we tried was" and thus other methods were also employed, however:
5. Why did MDN trust that another LLM would correctly flag output as "accurate, somewhat inaccurate, or incorrect"?
6. Why did MDN only review the output which the separate LLM deemed "low-quality", and not also the ones it deemed "accurate" to ensure they were, indeed, accurate?
7. Why do MDN / Mozilla believe it is viable to use a model trained on content that is 2-3 years out of date?
8. Will MDN take steps to identify such complex pages and prevent the LLM from attempting to summarise them until such time they have been sufficiently improved by a human?
9. How do MDN expect to collect sufficient reports of problematic LLM output when the target audience are those least likely to know the output is wrong at the time they read it, and given the time and effort involved in the user explaining the problem on GitHub when they're trying to work on their own project and needed a quick technical reference?
10. Can MDN not have a way to either only show LLM output already vetted by knowledgeable humans, or improve the reporting process so users only need click a button and it sends their prompt / page and the LLM output to MDN for review automatically? This would be beneficial for all "experimental" updates in the future to ensure the widest possible collection of reports.
And, finally, credit to @nicuveo for this:
11. Why are MDN so insistent on introducing a third-party content generator that, by design, will always be able to produce inaccurate-but-plausible output if MDN know that technical accuracy is why readers come to MDN?
Questions
Beta Was this translation helpful? Give feedback.
All reactions