-
Notifications
You must be signed in to change notification settings - Fork 0
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Committees: First Steps #20
Comments
I like this idea. It reminds me of a few ideas that have previously been thought about or tried, besides the one linked in the issue:
My suggested first step would probably be talking to the people involved in (1) to understand what we could learn from it. Then:
Another thought I have here for which you might have cached thoughts or there might be prior work, is on how much people will actually agree more about evaluations of organizations, pieces of research, etc. vs. forecasts. I imagine it varies based on the type of thing being forecasted/evaluated. It could actually be interesting to do an experiment on this if one hasn't been done already. An intuition pump that people might not agree more is that I'd imagine observing MIRI's work over the past year wouldn't do much to change disagreements between proponents/detractors if they made forecasts a year ago. |
Thanks! About the AI Forecasting Resolution Council: First, I'm familiar with it; I was one of the people who suggested it. Second, a major reason why it ended was just because Ben G and Jacob didn't stay active in the area. I could easily imagine a version of it with more continued effort behind it going much further. For "what to evaluate", I like the idea of "big lists of similar things". Some ideas:
Also, another idea, instead of an "evaluation committee of experts", would be something more like a survey or "deliberation democracy". Like, we randomly select a bunch of people who went to 2+ EAGs, put them in a panel, and use their responses. |
There are clearly some big cruxes in our community. If you are firm with one side, seeing an evaluation made by the other, or by a mix, might not help much. That said, if you could see evaluations made by those who agree with you on the big cruxes, that could still be great. My hunch is that differences matter less than a lot of people would expect. It was very frequent with Guesstimate models that the answer would be pretty over-determined, so people who initially disagreed with each other would end up being totally fine with the answer. |
I'd be very interested to learn more about your experiences and double crux on this, because my intuition is the opposite. |
(Happy to discuss sometime) |
Eventually, it could be great to have strong "evaluation bodies" that could be forecasted upon.
One first step would just be to set up a few evaluation bodies, and have them make a few group evaluations.
If it's very cheap to add forecasting, it could be interesting to add forecasting as well, to see how well forecasters could predict these evaluations.
See:
https://www.lesswrong.com/posts/cLtdcxu9E4noRSons/part-1-amplifying-generalist-research-via-forecasting-models
The text was updated successfully, but these errors were encountered: