Browse Source

Updating app-reviewers details

Signed-off-by: Mary Anthony <mary@blockstack.com>
feat/clarity-updates
Mary Anthony 6 years ago
parent
commit
8fcde42612
  1. 1
      _data/navigation_learn.yml
  2. 116
      _develop/app-reviewers.md
  3. 65
      _develop/mining_intro.md

1
_data/navigation_learn.yml

@ -33,6 +33,7 @@
- title: App Mining - title: App Mining
docs: docs:
- develop/mining_intro - develop/mining_intro
- develop/app-reviewers
- develop/mining_enroll - develop/mining_enroll
- develop/vote-blockstack - develop/vote-blockstack
- community/app-miners-guide - community/app-miners-guide

116
_develop/app-reviewers.md

@ -0,0 +1,116 @@
---
layout: learn
permalink: /:collection/:path.html
---
# Learn more about the reviewers
{:.no_toc}
Blockstack uses third-party reviewers who interact with, review, and provide scores for the apps in the App Mining program. The scores from each reviewer are used to determine an app`s ultimate rank. In this section, you learn more about the reviewers and how they score.
* TOC
{:toc}
## Product Hunt
Product Hunt is the place to find the best new products in tech. They have a massive trove of user data (upvotes, comments, etc), that they use for ranking. Product Hunt ranks apps based on a “community” score.
Their community score is determined only by the number of **credible** upvotes an app received on Product Hunt, relative to other apps that are registered. For example, if an app got more upvotes than any other app in a cohort, their community score would be 100. If a different app got 60% as many upvotes, they’d get a score of 60. Blockstack converts these scores into *z-scores*, more about z-scores below.
## Democracy Earth
Democracy Earth is a platform for borderless peer-to-peer democracy. They’ve
built a platform that anyone can use to gather votes in a trust-less,
decentralized way.
Democracy Earth has built a platform for Stacks token holders to vote on how
apps should be ranked. Each token holder gets a 1000 votes, and they can
distribute those votes however they want. It’s possible to give all of your
votes to a single app, and you can also “downvote” an app with one of your
votes.
After a voting period, each app has a certain amount of upvotes and downvotes.
First, Democracy Earth calculates the percentage of total votes that are
upvotes. If an app receives 90 upvotes and 10 downvotes, the resulting
“likability score” is 90. Secondly, Democracy Earth calculates a “traction
score”, which ranks how many total votes (including downvotes) an app received,
relative to other apps.
## Digital Rights Reviewer
As a "digital rights reviewer", New Internet Labs reviews apps submitted to the app mining program on the basis of the degree to which they respect and protect users` fundamental digital rights. Digital Rights Reviewer ranks based on the following criteria:
Blockstack Auth, and is the Blockstack ID:
- the only auth method
- one of multiple methods (presented equally with all others)
- a secondary auth method
- not used at all
- more detail [here](https://github.com/blockstack/app-mining/blob/master/DigitalRightsAuthScoringCriteria.pdf).
The app`s use of gaia. Each app will be given a rating based on which category it falls into:
- doesn`t use gaia at all
- uses gaia for some things (some data is stored elsewhere or we are unable to determine if some critical user data is sent elsewhere)
- data is only stored in gaia.
**Please note:** Digital rights reviewer will not provide an exact testing environment, but will test various environments changing month to month. Review will be conducted on the current public release of mac, windows, iOS or android with either the default browser on the platform or the current release of chrome or Firefox. If your app fails this test, by definition you`re not eligible for App Mining since your app does not have working Blockstack auth.
## Awario
Awario is a tool that brands world-wide use to better understand the online "conversation" surrounding their brand and to collect meaningful data with which to grow it. If you are interested in using Awario independent of App Mining, please [see Awario’s documentation](https://awario.com/help/) as it answers most questions about how the platform works.
Awario provides Blockstack with an awareness score beginning with the second month an ap is in the program. App Miners do not receive a score from Awario in their first eligible month so that Awario can spend that first month evaluating, honing, and updating an apps' brand data. The purpose of this is to zero in on only relevant data.
### Awareness scoring
At a high-level, Awario focuses on two major aspects of awareness mentions and reach.
<table class="uk-table">
<tr>
<td>mentions</td>
<td>Captured mentions of a brand or app online, on social networks, and on news sites. A mention is registered when the name the brand or app appears publicly, for example, a tweet mentioning the app name.</td>
</tr>
<tr>
<td>reach</td>
<td>The estimated reach of the combined mentions collected for your brand, for example, Hhow far the tweet about said app traveled online.</td>
</tr>
</table>
Mentions are captured and provided to App Miners. For the purposes of App Mining, the focus is on reach which is the less gameable and thus more suitable for App Mining. For example, it would be fairly easy to create many fake individual mentions (for example, a Twitter bot), but it would be unlikely that those fake mentions generate much if any actual reach. Using reach, Awario provides beginning in the second month of an app`s entry, a **Reach Score** and a **Growth Score**.
An app's **Reach Score** is based on total reach of all eligible mentions for the previous calendar month. A score is `log10(total_reach)`. So, if an app reaches 10 people, the score is 1, 100 is 2, 1000 is 3, and so on. This is much better than only using your actual reach, because outliers would totally skew the distribution. No matter what your reach is, you need to improve 10x your reach to increase this by 1. Using `log10` is also similar to how Blockstack handles the `theta` function in the mining algorithm, because the higher a score, the more an app needs to improve to bump its score.
A **Growth Score** is the month-over-month (MoM) growth in an apps total reach (not `log10`). If an app went from a reach of 1000 to 1500, its MoM growth is 0.5 (or 50%).
Like all the other reviewers, the z-score is first calculated for each of these metrics, and then then averaged. Then, the theta function is applied, resulting in a `final` Awario score.
### What counts as reach
The Awario team directly builds the query and search parameters for mention alerts. This leverages their expertise on their own platform. For unique brand or app names, the query is fairly straightforward, the name is loaded into Awario and it begins crawling sites and networks for public instances of it. For common names or names where context is important, such as a brand name like Stealthy, it is important to filter out non-relevant results like someone simply using the verb ‘stealthy’. For these cases, Awario sets up much [more complex queries](https://awario.com/help/boolean-search/boolean-syntax-and-operators/) to trim it down to the ones actually related to the project.
Websites are also excluded from reach totals. Such exclusion is generally pretty accurate and useful in a normal business use-case, it can be gamed because of the way Awario estimates the website reach. The site`s Alexa rank is used in the reach calculation meaning, for example, a mention on Github would register as massive reach, even if that particular mention didn’t really spread that far.
Mentions and associated reach from Blockstack accounts are not counted. This is allows Blockstack PBC to support apps publicly, without worrying that it needs to be evenly distributed across apps, which simply isn’t possible. Also not counted are the handles of all Blockstack PBC employees and all official Blockstack PBC managed accounts such as `@blockstack`. This is so everyone at Blockstack PBC can continue to freely support applications without running the risk of unintentionally biasing the results.
### Mention auditing
As part of the monthly process for generating Awario scores, queries not only optimized for mentions of the eligible apps, Awario also helps to audit the mentions coming in. This audit ensures that mentions that shouldn’t count don`t. Awario is confident in their ability to only collect and count relevant mentions through their search operators and are available to answer and questions or concerns App Miners may have. Last, with Awario’s platform, it is extremely easy to remove individual mentions (and thus the associated reach count) or to blacklist any accounts found to be fraudulent or accounting for false-positive mentions.
## TryMyUI
TryMyUI’s panelists score using a special survey they developed expressly for the App Mining program: the ALF Questionnaire (Adoption Likelihood Factors). Desktop, iOS, and Android versions of apps are tested as they are applicable.
Answers to this questionnaire will be used to calculate an overall score reflecting the following 4 factors:
* Usability
* Usefulness
* Credibility
* Desirability
Each factor corresponds to 4 questionnaire items, for a total of 16 items that comprise the ALFQ. Users mark their answers on a 5-point Likert scale, with 5 meaning **Strongly agree** and 1 meaning **Strongly disagree**. The final result is a score for each of the 4 factors, and a composite ALF score.
<img src="images/alf-score.png" alt="">
For example, consider an application that is both Android and iOS. Each platform version receives 4 tests of each. In total, 8 user tests are created, the highest and lowest scores are dropped. App developers receive the raw TryMyUI scores. The App Mining process calculates Z scores for each category. As a result, the TryMyUI results in the App Mining scores differ from raw scores visible in an app`s TryMyUI account.

65
_develop/mining_intro.md

@ -19,70 +19,12 @@ This section explains App Mining, a program for developers. For Blockstack, App
Blockstack worked with a team of Ph.D. game theorist and economists from Blockstack worked with a team of Ph.D. game theorist and economists from
Princeton and NYU to put together a [ranking Princeton and NYU to put together a [ranking
algorithm](https://blog.blockstack.org/app-mining-game-theory-algorithm-design/) algorithm](https://blog.blockstack.org/app-mining-game-theory-algorithm-design/)
which is fair and resistant to abuse. Blockstack uses three third-party which is fair and resistant to abuse. Blockstack uses the third-party
reviewers, Product Hunt, TryMyUI, and Democracy.earth. These reviewers are reviewer: Product Hunt, Awario, TryMyUI, and Democracy.earth. These reviewers are
independent, and generally rely on their own proprietary data and insights to independent, and generally rely on their own proprietary data and insights to
generate rankings. generate rankings.
### Product Hunt To learn in detail about these reviewers, see the page on [who reviews apps](app-reviewers.html).
Product Hunt is the place to find the best new products in tech. They have a
massive trove of user data (upvotes, comments, etc), that they use for ranking.
Product Hunt comes up with two different scores for each app — a “community”
score and a “team” score.
Their community score is determined only by the number of upvotes an app
received on Product Hunt, relative to other apps that are registered. For
example, if an app got more upvotes than any other app in a cohort, their
community score would be 100. If a different app got 60% as many upvotes, they’d
get a score of 60.
Their team score is determined by internal team members conducting reviews on
different aspects of an app. They judge based on a few criteria, like execution,
uniqueness, and desirability. Each app gets ranked 1-10 on each criterion, and
their final score is the average of each criterion. Finally, this average is
multiplied by 10, so the highest score you can get is 100.
Once each app has a community and team score, Blockstack converts these scores into
_z-scores_, more about z-scores below.
### Democracy Earth
Democracy Earth is a platform for borderless peer-to-peer democracy. They’ve
built a platform that anyone can use to gather votes in a trust-less,
decentralized way.
Democracy Earth has built a platform for Stacks token holders to vote on how
apps should be ranked. Each token holder gets a 1000 votes, and they can
distribute those votes however they want. It’s possible to give all of your
votes to a single app, and you can also “downvote” an app with one of your
votes.
After a voting period, each app has a certain amount of upvotes and downvotes.
First, Democracy Earth calculates the percentage of total votes that are
upvotes. If an app receives 90 upvotes and 10 downvotes, the resulting
“likability score” is 90. Secondly, Democracy Earth calculates a “traction
score”, which ranks how many total votes (including downvotes) an app received,
relative to other apps.
### TryMyUI
TryMyUI’s panelists score using a special survey they developed expressly for the App Mining program: the ALF Questionnaire (Adoption Likelihood Factors). Desktop, iOS, and Android versions of apps are tested as they are applicable.
Answers to this questionnaire will be used to calculate an overall score reflecting the following 4 factors:
* Usability
* Usefulness
* Credibility
* Desirability
Each factor corresponds to 4 questionnaire items, for a total of 16 items that comprise the ALFQ. Users mark their answers on a 5-point Likert scale, with 5 meaning **Strongly agree** and 1 meaning **Strongly disagree**. The final result is a score for each of the 4 factors, and a composite ALF score.
<img src="images/alf-score.png" alt="">
For example, consider an application that is both Android and iOS. Each platform version receives 4 tests of each. In total, 8 user tests are created, the highest and lowest scores are dropped. App developers receive the raw TryMyUI scores. The App Mining process calculates Z scores for each category. As a result, the TryMyUI results in the App Mining scores differ from raw scores visible in an app's TryMyUI account.
## Reaching the final scores ## Reaching the final scores
@ -110,6 +52,7 @@ Once each app has a calculated a z-score in every category, the average of those
4 z-scores results in a final number. A higher number is better than a lower 4 z-scores results in a final number. A higher number is better than a lower
one, and so apps are ranked from highest to lowest. one, and so apps are ranked from highest to lowest.
## Determining how much an app is paid ## Determining how much an app is paid
{% include payout-appmining.md %} {% include payout-appmining.md %}

Loading…
Cancel
Save