Week 27-28: Writing
Hello all.
The past two weeks since returning from spring break, I have been working with the grad student to write the vis paper. It is due next week on Friday. A couple of the undergraduates (along with my CREU peer, Diana) have stayed on to make changes to the RanKit application to make it better for publication.
Briefly, I'll share a couple things I wrote in the past two weeks that have gone into the paper. First, I'll share some equations that describe the amount of information collected from the different build tools for each unit of user effort. Second, I'll give three use cases that describe why these tools might be used.
The past two weeks since returning from spring break, I have been working with the grad student to write the vis paper. It is due next week on Friday. A couple of the undergraduates (along with my CREU peer, Diana) have stayed on to make changes to the RanKit application to make it better for publication.
Briefly, I'll share a couple things I wrote in the past two weeks that have gone into the paper. First, I'll share some equations that describe the amount of information collected from the different build tools for each unit of user effort. Second, I'll give three use cases that describe why these tools might be used.
Equations
The ranking algorithm is based on pairwise comparisons. We will assume that the user's input is meaningful and that there exist no cycles in their rank (e.g. user says Toy Story 1 > Toy Story 2, Toy Story 2 > Toy Story 3, and Toy Story 3 > Toy Story 1). The collection of all possible rankings is a directed acyclic graph. For n objects, this is (n^2 - n)/2 pairs. If there are 100 objects, this translates to 4950 pairs. For purposes of this argument, let's say that an "optimal" preference elicitation is one where n log(n) pairs are generated. (This can be thought of as the point at which the user's preferences start to "fight" the model. More information does not contribute to the model, and user preferences will be less reflected in the output model. Less information, however, and the model is uncertain. Multiple rankings could exist that explain the user's preferences). In a pool of 100 objects, this means 664 pairs must be generated.
Pairwise
Pairwise comparison is the most simple and directly related to the model. For every pair of objects the user specifies, 1 additional pair is known for the model. In other words, 664 unique pairs must be created in order to achieve the optimal amount of information. Some could be said about the transitive property: if a user says Toy Story 1 > Toy Story 2 and Toy Story 2 > Toy Story 3, then one can assume that Toy Story 1 > Toy Story 3. So in reality, the number of unique pairs might be slightly less.
List
List comparison gives the most information for the least amount of user effort. For every object added to a preference list of n objects, n new pairs are learned. This grows at a rate of (n^2 - n)/2. In a pool of 100 objects, 37 must be ranked to acquire optimal information.
Categorical
Categorical comparison might make sense to a user. It gathers more information from the user than pairwise, but not as much as in list. In fact, unlike pairwise and list, there is a restricted amount of information the model can learn. In a pool of 100 objects, if the user splits all 100 objects among the three categories, only 3267 pairs can be learned. This is because objects within the same category are unaware of each other (if two movies are in the Medium category, no pairs can be learned between them). For every object added, 2/9n^2 - 1/3n ranking pairs are generated. 55 objects must be ranked to achieve the optimal amount of information.
Use Cases
List
A major tech company is looking to expand by building a second headquarters. They want to know where to build in order to attract the most talent as well as save money with state business incentives. They use a publicly available dataset with information as to the number of tech employees, the average starting salary, as well as state tax rates and estimated discount rate for businesses. They use RanKit to input the top states in the US according to number of tech companies. When they calculate their rank, they notice that average starting salary is heavily weighted in the model and that border states are ranked higher than mid-country states.
Categorical
Beth is a high school junior with plans to attend college. She is trying to decide where to apply. She has had the opportunity to tour several schools around her hometown and feels confident making judgments on them. Beth uses the publicly available Scorecard dataset that contains diverse attributes about colleges across the US. She knows she really liked some of the colleges in her area, felt some were ok, and despised others. She ranks these colleges into the high, medium, and low categories using the Categorical view tool. When she clicks "Rank!" she finds that a major predictor of her preferences was the number of undergraduates. Upon further reflection, she decides that number of students doesn't really matter to her. She remembers a smaller university that she forgot to add to the ranking that she really liked, so she returns to the build view to add more preferences. The build view suggests several universities that would help the quality of the ranking for which Beth could add more information. Beth chooses to research these colleges in more depth to form her opinion.
Pairwise
Carl is an active member on a fantasy football league. He wants to make targeted predictions as to which team will be most successful over the season. He uses the Pairwise view on RanKit with a dataset on post-game statistics to encode which teams won and lost over the last week.
Comments
Post a Comment