<?xml version="1.0" encoding="utf-8"?><feed xmlns="http://www.w3.org/2005/Atom" xml:lang="en"><generator uri="https://jekyllrb.com/" version="3.9.3">Jekyll</generator><link href="https://whiting.me/feed.xml" rel="self" type="application/atom+xml" /><link href="https://whiting.me/" rel="alternate" type="text/html" hreflang="en" /><updated>2026-03-06T17:04:31+00:00</updated><id>https://whiting.me/feed.xml</id><title type="html">Mark E. Whiting</title><subtitle>Mark is Chief Scientist at Pareto and research fellow at University of Pennsylvania</subtitle><entry><title type="html">Quantifying common sense</title><link href="https://whiting.me/Quantifying-common-sense/" rel="alternate" type="text/html" title="Quantifying common sense" /><published>2024-01-16T00:00:00+00:00</published><updated>2024-01-16T00:00:00+00:00</updated><id>https://whiting.me/Quantifying%20common%20sense</id><content type="html" xml:base="https://whiting.me/Quantifying-common-sense/">&lt;p&gt;Together with &lt;a href=&quot;https://css.seas.upenn.edu/people/duncan-watts/&quot;&gt;Duncan Watts&lt;/a&gt; and others, I have developed an empirical method for quantifying common sense at the level of an individual or a collective.&lt;/p&gt;

&lt;p&gt;Our &lt;a href=&quot;https://www.pnas.org/doi/10.1073/pnas.2309535121&quot;&gt;paper has been published at PNAS&lt;/a&gt;, and the significance statement from the article is as follows:&lt;/p&gt;

&lt;blockquote&gt;
  &lt;p&gt;Common sense, while often portrayed as universal, is paradoxically also often claimed not to exist. Here, we resolve this puzzling situation by introducing a formal methodology to empirically quantify common sense both at individual and collective levels. We then demonstrate the method with a dataset involving  human raters evaluating claims. We show that common sense varies considerably across types of claims but aligns most closely with plainly worded, factual claims about physical reality; in contrast, does not vary much across different types of people. We also find limited presence of collective common sense, undermining universalist claims and supporting skeptics. Finally, we argue that quantifying common sense is useful both for applications in social science and AI.&lt;/p&gt;
&lt;/blockquote&gt;

&lt;p&gt;The work has received attention at a few sources, which Altmetric summarize nicely:&lt;/p&gt;

&lt;div class=&quot;altmetric-embed altmetric-box&quot; data-badge-type=&quot;medium-donut&quot; data-badge-details=&quot;right&quot; data-altmetric-id=&quot;158425670&quot;&gt;&lt;/div&gt;</content><author><name></name></author><category term="commonsense" /><category term="experiments" /><category term="meta-scientific" /><category term="methods" /><category term="Penn" /><summary type="html">Together with Duncan Watts and others, I have developed an empirical method for quantifying common sense at the level of an individual or a collective.</summary><media:thumbnail xmlns:media="http://search.yahoo.com/mrss/" url="https://whiting.me/commonsense.png" /><media:content medium="image" url="https://whiting.me/commonsense.png" xmlns:media="http://search.yahoo.com/mrss/" /></entry><entry><title type="html">Integrative experiments</title><link href="https://whiting.me/Integrative-experiments/" rel="alternate" type="text/html" title="Integrative experiments" /><published>2022-12-21T00:00:00+00:00</published><updated>2022-12-21T00:00:00+00:00</updated><id>https://whiting.me/Integrative%20experiments</id><content type="html" xml:base="https://whiting.me/Integrative-experiments/">&lt;p&gt;Together with &lt;a href=&quot;http://amaatouq.io/&quot;&gt;Abdullah Almaatouq&lt;/a&gt;, &lt;a href=&quot;https://css.seas.upenn.edu/people/duncan-watts/&quot;&gt;Duncan Watts&lt;/a&gt; and others, we have been developing and using a different apraoch to running experiments. In &lt;a href=&quot;https://doi.org/10.1017/S0140525X22002874&quot;&gt;Beyond Playing 20 Questions with Nature: Integrative Experiment Design in the Social and Behavioral Sciences&lt;/a&gt; we write:&lt;/p&gt;

&lt;blockquote&gt;
  &lt;p&gt;The dominant paradigm of experiments in the social and behavioral sciences views an experiment as a test of a theory, where the theory is assumed to generalize beyond the experiment’s specific conditions. According to this view, which Alan Newell once characterized as “playing twenty questions with nature,” theory is advanced one experiment at a time, and the integration of disparate findings is assumed to happen via the scientific publishing process. In this article, we argue that the process of integration is at best inefficient, and at worst it does not, in fact, occur. We further show that the challenge of integration cannot be adequately addressed by recently proposed reforms that focus on the reliability and replicability of individual findings, nor simply by conducting more or larger experiments. Rather, the problem arises from the imprecise nature of social and behavioral theories and, consequently, a lack of commensurability across experiments conducted under different conditions. Therefore, researchers must fundamentally rethink how they design experiments and how the experiments relate to theory. We specifically describe an alternative framework, integrative experiment design, which intrinsically promotes commensurability and continuous integration of knowledge. In this paradigm, researchers explicitly map the design space of possible experiments associated with a given research question, embracing many potentially relevant theories rather than focusing on just one. The researchers then iteratively generate theories and test them with experiments explicitly sampled from the design space, allowing results to be integrated across experiments. Given recent methodological and technological developments, we conclude that this approach is feasible and would generate more-reliable, more-cumulative empirical and theoretical knowledge than the current paradigm—and with far greater efficiency.&lt;/p&gt;
&lt;/blockquote&gt;</content><author><name></name></author><category term="integrative" /><category term="experiments" /><category term="meta-scientific" /><category term="methods" /><category term="Penn" /><category term="MIT" /><summary type="html">Together with Abdullah Almaatouq, Duncan Watts and others, we have been developing and using a different apraoch to running experiments. In Beyond Playing 20 Questions with Nature: Integrative Experiment Design in the Social and Behavioral Sciences we write:</summary><media:thumbnail xmlns:media="http://search.yahoo.com/mrss/" url="https://whiting.me/integrative%20experiments.png" /><media:content medium="image" url="https://whiting.me/integrative%20experiments.png" xmlns:media="http://search.yahoo.com/mrss/" /></entry><entry><title type="html">COVID dashboards</title><link href="https://whiting.me/COVID-dashboards/" rel="alternate" type="text/html" title="COVID dashboards" /><published>2021-08-15T00:00:00+00:00</published><updated>2021-08-15T00:00:00+00:00</updated><id>https://whiting.me/COVID-dashboards</id><content type="html" xml:base="https://whiting.me/COVID-dashboards/">&lt;p&gt;Together with &lt;a href=&quot;https://css.seas.upenn.edu/people/duncan-watts/&quot;&gt;Duncan Watts&lt;/a&gt; and others on the team at the &lt;a href=&quot;https://css.seas.upenn.edu&quot;&gt;CSSLab&lt;/a&gt; at Penn, and in conjunction with the city of Philadelphia, we built a collection of interactive data dashboards that visually summarize human mobility patterns over time and space for a number of cities, starting with Philadelphia, along with highlighting potentially relevant demographic correlates. The dashboards are available at &lt;a href=&quot;https://covid.seas.upenn.edu&quot;&gt;covid.seas.upenn.edu&lt;/a&gt;.&lt;/p&gt;

&lt;p&gt;The dashboards are built in &lt;a href=&quot;https://observablehq.com/collection/@wattslab/covid-dashboards&quot;&gt;Observable&lt;/a&gt;. The data are a proprietary combination of cell phone GPS data, demographic data derived from the American Community Survey, and COVID-19 caseload data from the New York Times.&lt;/p&gt;</content><author><name></name></author><category term="COVID-19" /><category term="dashboard" /><category term="observable" /><category term="data" /><category term="visualization" /><category term="Penn" /><summary type="html">Together with Duncan Watts and others on the team at the CSSLab at Penn, and in conjunction with the city of Philadelphia, we built a collection of interactive data dashboards that visually summarize human mobility patterns over time and space for a number of cities, starting with Philadelphia, along with highlighting potentially relevant demographic correlates. The dashboards are available at covid.seas.upenn.edu.</summary><media:thumbnail xmlns:media="http://search.yahoo.com/mrss/" url="https://whiting.me/COVID%20dahsboard.jpg" /><media:content medium="image" url="https://whiting.me/COVID%20dahsboard.jpg" xmlns:media="http://search.yahoo.com/mrss/" /></entry><entry><title type="html">Team fracture</title><link href="https://whiting.me/team-fracture/" rel="alternate" type="text/html" title="Team fracture" /><published>2019-11-12T00:00:00+00:00</published><updated>2019-11-12T00:00:00+00:00</updated><id>https://whiting.me/team-fracture</id><content type="html" xml:base="https://whiting.me/team-fracture/">&lt;p&gt;While at &lt;a href=&quot;https://hci.stanford.edu&quot;&gt;Stanford&lt;/a&gt;, &lt;a href=&quot;https://hci.stanford.edu/msb/&quot;&gt;Michael Bernstein&lt;/a&gt; and I, with the help of many great undergrads, designed an experiment to study the consistency of team fracture, a notion we defined as a loss of team viability so severe that the team no longer wants to work together &lt;a class=&quot;button smallCaps&quot; href=&quot;https://hci.stanford.edu/publications/2019/fracture/fracture-cscw2019.pdf&quot;&gt;CSCW’19&lt;/a&gt;.&lt;/p&gt;

&lt;p&gt;We asked, &lt;strong&gt;was a problematic team always doomed to frustration, or could it have ended another way?&lt;/strong&gt;&lt;/p&gt;

&lt;p&gt;To study this, we introduce an online experiment that reconvenes the same team repeatedly, without members realizing that they have worked together before, enabling us to temporarily erase previous team dynamics. We found that reconvened teams are strikingly polarized by task in the consistency of their fracture outcomes — on a creative task, teams might as well have been a completely different set of people: the same teams changed their fracture outcomes at a random chance rate, on the other hand on a cognitive conflict and on an intellective task, the team instead replayed the same dynamics without realizing it, rarely changing their fracture outcomes. Our results indicate that, for some tasks, team fracture can be strongly influenced by interactions in the first moments of a team’s collaboration, and that interventions targeting these initial moments may be critical to scaffolding long-lasting teams.&lt;/p&gt;

&lt;p&gt;In follow up work, we studied the predictability of team fracture &lt;a class=&quot;button smallCaps&quot; href=&quot;https://hci.stanford.edu/publications/2020/viability-prediction/viability-prediction.pdf&quot;&gt;CSCW’21&lt;/a&gt;, using a range of human rated and automatically generated features. We found that automatically generated features alone served as a strong predictor of team fracture. We made the underlying system available at &lt;a href=&quot;https://viability.stanford.edu&quot;&gt;viability.stanford.edu&lt;/a&gt;.&lt;/p&gt;</content><author><name></name></author><category term="fracture" /><category term="teams" /><category term="Stanford" /><category term="study" /><category term="CSCW" /><summary type="html">While at Stanford, Michael Bernstein and I, with the help of many great undergrads, designed an experiment to study the consistency of team fracture, a notion we defined as a loss of team viability so severe that the team no longer wants to work together CSCW’19.</summary><media:thumbnail xmlns:media="http://search.yahoo.com/mrss/" url="https://whiting.me/fracture.png" /><media:content medium="image" url="https://whiting.me/fracture.png" xmlns:media="http://search.yahoo.com/mrss/" /></entry><entry><title type="html">Fair work</title><link href="https://whiting.me/fair-work/" rel="alternate" type="text/html" title="Fair work" /><published>2019-10-30T00:00:00+00:00</published><updated>2019-10-30T00:00:00+00:00</updated><id>https://whiting.me/fair-work</id><content type="html" xml:base="https://whiting.me/fair-work/">&lt;p&gt;While at &lt;a href=&quot;https://hci.stanford.edu&quot;&gt;Stanford&lt;/a&gt;, &lt;a href=&quot;https://linkedin.com/in/grant-hugh&quot;&gt;Grant Hugh&lt;/a&gt;, &lt;a href=&quot;https://hci.stanford.edu/msb/&quot;&gt;Michael Bernstein&lt;/a&gt; and I, released a mechanism to make it possible for Amazon Mechanical Turk requesters to pay workers a more fair wage by adding only one line of code to their tasks &lt;a class=&quot;button smallCaps&quot; href=&quot;https://hci.stanford.edu/publications/2019/fairwork/fairwork-hcomp2019.pdf&quot;&gt;HCOMP’19&lt;/a&gt;. Our paper was awarded an Honorable Mention!&lt;/p&gt;

&lt;p&gt;You can use the service by visiting &lt;a href=&quot;https://fairwork.stanford.edu&quot;&gt;fairwork.stanford.edu&lt;/a&gt;, where you will get a line of code to include in your tasks on MTurk.&lt;/p&gt;

&lt;p&gt;Below I’ve included the paper abstract:&lt;/p&gt;

&lt;blockquote&gt;
  &lt;p&gt;&lt;strong&gt;Fair Work: Crowd Work Minimum Wage with One Line of Code&lt;/strong&gt;&lt;br /&gt;
Accurate task pricing in microtask marketplaces requires substantial effort via trial and error, contributing to a pattern of worker underpayment. In response, we introduce Fair Work, enabling requesters to automatically pay their workers minimum wage by adding a one-line script tag to their task HTML on Amazon Mechanical Turk. &lt;strong&gt;Fair Work automatically surveys workers to find out how long the task takes, then aggregates those self-reports and auto-bonuses workers up to a minimum wage if needed.&lt;/strong&gt; Evaluations demonstrate that the system estimates payments more accurately than requesters and that worker time surveys are close to behaviorally observed time measurements. With this work, we aim to lower the threshold for pro-social work practices in microtask marketplaces.&lt;/p&gt;
&lt;/blockquote&gt;</content><author><name></name></author><category term="Fair" /><category term="work" /><category term="mturk" /><category term="Stanford" /><category term="study" /><category term="system" /><category term="HCOMP" /><summary type="html">While at Stanford, Grant Hugh, Michael Bernstein and I, released a mechanism to make it possible for Amazon Mechanical Turk requesters to pay workers a more fair wage by adding only one line of code to their tasks HCOMP’19. Our paper was awarded an Honorable Mention!</summary><media:thumbnail xmlns:media="http://search.yahoo.com/mrss/" url="https://whiting.me/fairwork.png" /><media:content medium="image" url="https://whiting.me/fairwork.png" xmlns:media="http://search.yahoo.com/mrss/" /></entry><entry><title type="html">Crowdsourcing</title><link href="https://whiting.me/crowdsourcing/" rel="alternate" type="text/html" title="Crowdsourcing" /><published>2017-06-01T00:00:00+00:00</published><updated>2017-06-01T00:00:00+00:00</updated><id>https://whiting.me/crowdsourcing</id><content type="html" xml:base="https://whiting.me/crowdsourcing/">&lt;p&gt;Crowdsourcing tools today offer ad hoc and remote work arrangements at an unprecedented scale, yet many of the perks of traditional jobs have not yet been realized for this community. The &lt;a href=&quot;https://github.com/crowdresearch&quot;&gt;Stanford Crowd Research Collective&lt;/a&gt; developed &lt;a href=&quot;https://www.daemo.org&quot;&gt;Daemo&lt;/a&gt; &lt;a class=&quot;button smallCaps&quot; href=&quot;http://hci.stanford.edu/publications/2017/crowdguilds/CSCWDemo.pdf&quot;&gt;CSCW’17&lt;/a&gt;, a crowdsourcing platform aiming to mitigate many of these challenges. For example we introduced crowd guilds &lt;a class=&quot;button smallCaps&quot; href=&quot;http://hci.stanford.edu/publications/2017/crowdguilds/guilds.pdf&quot;&gt;CSCW’17&lt;/a&gt;, to improve reputation and feedback within crowdsourcing, inspired by traditional worker guilds, &lt;a href=&quot;https://crowdresearch.github.io/open-gov/constitution/2016/11/21/constitution.html&quot;&gt;Daemo Constitution&lt;/a&gt; &lt;a class=&quot;button smallCaps&quot; href=&quot;http://collectiveintelligenceconference.org&quot;&gt;CI’17&lt;/a&gt;, to give workers and requesters agency in platform governance, boomerang &lt;a class=&quot;button smallCaps&quot; href=&quot;http://hci.stanford.edu/publications/2016/boomerang/boomerang-uist.pdf&quot;&gt;UIST’16&lt;/a&gt;, to incentivize honest feedback, and prototype tasks &lt;a class=&quot;button smallCaps&quot; href=&quot;https://arxiv.org/pdf/1707.05645.pdf&quot;&gt;HCOMP’17&lt;/a&gt;, to fix tasks before they launch.&lt;/p&gt;</content><author><name></name></author><category term="jekyll" /><category term="update" /><summary type="html">Crowdsourcing tools today offer ad hoc and remote work arrangements at an unprecedented scale, yet many of the perks of traditional jobs have not yet been realized for this community. The Stanford Crowd Research Collective developed Daemo CSCW’17, a crowdsourcing platform aiming to mitigate many of these challenges. For example we introduced crowd guilds CSCW’17, to improve reputation and feedback within crowdsourcing, inspired by traditional worker guilds, Daemo Constitution CI’17, to give workers and requesters agency in platform governance, boomerang UIST’16, to incentivize honest feedback, and prototype tasks HCOMP’17, to fix tasks before they launch.</summary><media:thumbnail xmlns:media="http://search.yahoo.com/mrss/" url="https://whiting.me/guilds.png" /><media:content medium="image" url="https://whiting.me/guilds.png" xmlns:media="http://search.yahoo.com/mrss/" /></entry><entry><title type="html">Automated design</title><link href="https://whiting.me/automated-design/" rel="alternate" type="text/html" title="Automated design" /><published>2017-01-01T00:00:00+00:00</published><updated>2017-01-01T00:00:00+00:00</updated><id>https://whiting.me/automated-design</id><content type="html" xml:base="https://whiting.me/automated-design/">&lt;p&gt;To automate design, the nuances of design decisions need to be represented computationally. This is challenging because even humans often don’t agree on the decisions going into a design. In my PhD, I introduced an automatic, grammar-based system for structuring design information &lt;a class=&quot;button smallCaps&quot; href=&quot;http://link.springer.com/chapter/10.1007/978-3-319-44989-0_15&quot;&gt;DCC’16&lt;/a&gt;, and showed how it could be applied to design problems &lt;a class=&quot;button smallCaps&quot; href=&quot;http://doi.org/10.1017/S0890060417000464&quot;&gt;AI EDAM&lt;/a&gt;.&lt;/p&gt;</content><author><name></name></author><category term="jekyll" /><category term="update" /><summary type="html">To automate design, the nuances of design decisions need to be represented computationally. This is challenging because even humans often don’t agree on the decisions going into a design. In my PhD, I introduced an automatic, grammar-based system for structuring design information DCC’16, and showed how it could be applied to design problems AI EDAM.</summary><media:thumbnail xmlns:media="http://search.yahoo.com/mrss/" url="https://whiting.me/grammar.png" /><media:content medium="image" url="https://whiting.me/grammar.png" xmlns:media="http://search.yahoo.com/mrss/" /></entry><entry><title type="html">Incentive design</title><link href="https://whiting.me/incentive-design/" rel="alternate" type="text/html" title="Incentive design" /><published>2016-12-01T00:00:00+00:00</published><updated>2016-12-01T00:00:00+00:00</updated><id>https://whiting.me/incentive-design</id><content type="html" xml:base="https://whiting.me/incentive-design/">&lt;p&gt;Using incentives to motivate human behavior is very effective, but poorly designed incentives often lead to worse behavior. For example, the common technique of requiring a certain number of characters of feedback in an online feedback form is more likely to lead to feedback padded with spaces than higher quality responses. With Dilrukshi Gamage, Thejan Rajapakshe, Haritha Thilakarathne, Indika Perera, and Shantha Fernando, we explored how aligning the incentives for peer feedback in online learning helps improve the perceived quality and length of the feedback &lt;a class=&quot;button smallCaps&quot; href=&quot;https://arxiv.org/pdf/1703.06169.pdf&quot;&gt;L@S’17&lt;/a&gt;.&lt;/p&gt;</content><author><name></name></author><category term="jekyll" /><category term="update" /><summary type="html">Using incentives to motivate human behavior is very effective, but poorly designed incentives often lead to worse behavior. For example, the common technique of requiring a certain number of characters of feedback in an online feedback form is more likely to lead to feedback padded with spaces than higher quality responses. With Dilrukshi Gamage, Thejan Rajapakshe, Haritha Thilakarathne, Indika Perera, and Shantha Fernando, we explored how aligning the incentives for peer feedback in online learning helps improve the perceived quality and length of the feedback L@S’17.</summary><media:thumbnail xmlns:media="http://search.yahoo.com/mrss/" url="https://whiting.me/uom.png" /><media:content medium="image" url="https://whiting.me/uom.png" xmlns:media="http://search.yahoo.com/mrss/" /></entry><entry><title type="html">Hybrid systems</title><link href="https://whiting.me/hybrid-systems/" rel="alternate" type="text/html" title="Hybrid systems" /><published>2016-01-01T00:00:00+00:00</published><updated>2016-01-01T00:00:00+00:00</updated><id>https://whiting.me/hybrid-systems</id><content type="html" xml:base="https://whiting.me/hybrid-systems/">&lt;p&gt;Humans are pretty good at chess, computers are better, humans and computers together have been even better. However, we are still not very good at getting the best out of this kind of working relationship. Mentoring undergraduate students at CMU and in the &lt;a href=&quot;http://www.sis.pitt.edu/i3/&quot;&gt;Pitt i3&lt;/a&gt;, we built systems to encourage humans and computers to work together. For example, we explored how socio-technical systems can provide novel insight into new product development opportunities, by performing computational analysis on Amazon product reviews &lt;a class=&quot;button smallCaps&quot; href=&quot;#&quot;&gt;iCONF’16&lt;/a&gt;, and we studied how hybrid tools can help researchers get started with crowdsourcing faster &lt;a class=&quot;button smallCaps&quot; href=&quot;#&quot;&gt;iCONF’18&lt;/a&gt;.&lt;/p&gt;</content><author><name></name></author><category term="jekyll" /><category term="update" /><summary type="html">Humans are pretty good at chess, computers are better, humans and computers together have been even better. However, we are still not very good at getting the best out of this kind of working relationship. Mentoring undergraduate students at CMU and in the Pitt i3, we built systems to encourage humans and computers to work together. For example, we explored how socio-technical systems can provide novel insight into new product development opportunities, by performing computational analysis on Amazon product reviews iCONF’16, and we studied how hybrid tools can help researchers get started with crowdsourcing faster iCONF’18.</summary><media:thumbnail xmlns:media="http://search.yahoo.com/mrss/" url="https://whiting.me/hybrid.png" /><media:content medium="image" url="https://whiting.me/hybrid.png" xmlns:media="http://search.yahoo.com/mrss/" /></entry></feed>