Tl;dr: Try out the new Wikimetrics[1] module and let me know what you think!
Dear Wikimetrics users!
We have been working over the last few weeks to develop an expanded
Wikimetrics learning module. You can find the module on the Program
Evaluation Learning Modules page in our Evaluation portal:
https://meta.wikimedia.org/wiki/Programs:Evaluation_portal/Learning_modules
While the Wikimetrics tool will continue to evolve, the expanded module
includes more detailed information about how program leaders and other
Wikimetrics users can prepare for, and currently use, Wikimetrics to
analyze cohort data.
We would love to hear your input as Wikimetrics users; does this module
assist in avoiding the challenges you may have faced using the tool? Your
constructive feedback is most welcome.
Feel free to contact me directly or comment on learning module talk
page[2], which is actually slated to receive the new Flow[3] feature in the
near future! I am hoping to announce this new training to the wider
community of program leaders in the next couple of weeks.
Thank you in advance for your input!
All the best,
Edward
[1]
https://meta.wikimedia.org/wiki/Programs_talk:Evaluation_portal/Learning_mo…
[2]
https://meta.wikimedia.org/wiki/Programs_talk:Evaluation_portal/Learning_mo…
[3]https://en.wikipedia.org/wiki/Wikipedia:Flow
--
Edward Galvez
Intern, Program Evaluation and Design
Wikimedia Foundation
There's an outage that's preventing wikimetrics from accessing the labsdb
databases..
I'm not sure if this is related, but mediawiki OAuth stopped working for me
on wikimetrics as well. I am guessing it is related because the outage
seems to have weird networking implications.
Updates on both issues when we have them.
Dear Wikimetrics community,
While trying to create my first cohort today, I continually got the error message "Error! Server error while processing your upload." I checked the training, FAQ, etc. and couldn't tell what the problem could be. Is the server down? If not, might anyone have a sense of what could be going wrong?
Thanks in advance for any help,
Lea
Leanora Lange
-----
Processing Archivist
Center for Jewish History
llange(a)cjh.org
Hi Simon,
UserMetrics has been superseded by a new tool called Wikimetrics [1] open for anybody to use. More information about the project can be found here [2].
Best,
Dario
[1] http://metrics.wmflabs.org/
[2] http://www.mediawiki.org/wiki/Analytics/Wikimetrics
> On Jan 3, 2014, at 15:19, <simon.knight(a)wikimedia.org.uk> wrote:
>
> Hi
> I’m a trustee at Wikimedia UK (see signature link), I’m interested in gaining beta-access to the user metrics, partly for our own strategy planning which we are currently reviewing, partly because I have a wider interest in impact measurement, and partly for operational planning on metrics (for which it would be useful if staff had access to systems). Could I have details on how I/we gain access, or when/if we would be granted such access?
>
> Kind regards
> Simon
>
> Simon Knight @sjgknight
> PhD Candidate, Open University
> Director/Trustee, Wikimedia UK
>
We've just deployed https://gerrit.wikimedia.org/r/#/c/102618/ which
implements the identify method as part of our OAuth handshake with
meta.wikimedia.org. This is a much more secure way of using
pseudo-authenticate with OAuth, so everyone can rest easier that their user
names won't be hacked.
As always, let us know if you run into trouble.
Dan
I'm taking wikimetrics down for a bit, I have to reset some passwords that
were accidentally leaked. I don't suspect anything bad happened as we
caught it within a few minutes.
Dan
Hi,
I was going to try and add OAuth support. One problem is that I'd have to
pick a specific wiki and I know how people dislike revolving around enwiki.
Does anyone have any opinion on what wiki project we could use to
authenticate our wikimetrics users against? Commons? Enwiki? A few of
the top wikis (this would be harder)?
Dan
Hi everyone,
A quick note about something that just messed me up. When uploading a
cohort to wikimetrics, you are told you can use either user_name, user_id,
or a mixture in the first column. However, this can really produce
unexpected results if you don't know how it works. I think it needs to
change, but until then, this is how it works and how it can bite you:
Let's say I have a list of users:
1,en
2,en
3,en
When it validates, it will look up user_name == 1, if it doesn't find
anything it will look up user_id == 1. Then user_name == 2, user_id == 2,
user_name == 3, user_id == 3. If what you meant with the above cohort was
the users with ids 1, 2, and 3, then you might be very confused later when
you see user id 234215 in your output results. This might happen if a
user_name is actually 2! So, for now, until I figure out how to fix this,
it will always prefer user_names before user_ids.
Please let me know if this is confusing. Also, the whole problem stems
from needing to accept both user_id and user_name in the *same* upload. If
everyone agrees, I'd much rather just allow people to toggle between one or
the other. This would speed up validation and make it much clearer what is
going on.
Dear Wikimetrics users,
I've just deployed asynchronous cohort upload. This is feature #818:
https://mingle.corp.wikimedia.org/projects/analytics/cards/818 and
basically allows you to upload larger cohorts because validation is
happening behind the scenes. I'll go over how the new functionality works
here, and will rely on one of you to point me to the appropriate on-wiki
place to update documentation.
So basically, visiting /cohorts and clicking "Upload Cohort" works as
before. But once you click "Upload CSV", your form is validated,
processed, and you're taken back to the cohorts page. Your new cohort is
immediately created but is not yet validated. While it validates, you'll
see the validation status and have a few options:
* Remove Cohort. This is destructive and will remove this cohort from your
list. Use this in case you made a mistake, uploaded the wrong file, etc.
* Validate Again. This will run validation again. One possible use for it
is, let's say you upload a cohort with some *very* newly registered users.
And because of replication lag to the labsdb databases, most of them come
up invalid. You can then run validation again.
* Refresh. This just refreshes the status of the validation and will
update the counts that show up below.
You will not have the "Create Report" option until validation is done. And
when you do create a report, only valid users will be considered and used
in the output.
One caveat. Validation is still slow. And the time limit for the
asynchronous task is set to 1 hour. I have some ideas for making this
faster by batching, and I can increase the time limit per task (but that
has other repercussions). For now, just keep in mind that the theoretical
maximum cohort size you should upload is roughly 18,000 users. I would
love some feedback about whether it's ok to increase the time limit or if
people want me to focus on making validation faster.
Dan