Increasing the sensitivity of A/B tests by utilizing the variance estimates of experimental units

Increasing the sensitivity of A/B tests by utilizing the variance estimates of experimental units

Kevin Liou is a Research Scientist within Core Data Science, a research and development team focused on improving Facebook’s processes, infrastructure, and products.

What we did

Companies routinely turn to A/B testing when evaluating the effectiveness of their product changes. Also known as a randomized field experiment, A/B testing has been used extensively over the past decade to measure the causal impact of product changes or variants of services, and has proved to be an important success factor for businesses making decisions.

With increased adoption of A/B testing, proper analysis of experimental data is crucial to decision quality. Successful A/B tests must exhibit sensitivity — they must be capable of detecting effects that product changes generate. From a hypothesis-testing perspective, experimenters aim to have high statistical power, or the likelihood that the experiment will detect a nonzero effect when such an effect exists.

In our paper, “Variance-weighted estimators to improve sensitivity in online experiments,” we focus on increasing the sensitivity of A/B tests by attempting to understand the inherent uncertainty introduced by individual experimental units. To leverage this information, we propose directly estimating the pre-experiment individual variance for each unit. For example, if our target metric is “time spent by someone on the site per day,” we may want to give more weight to those who previously exhibited lower variance for this metric through their more consistent usage of the product. We can estimate the variance of a person’s daily time spent during the month before the experiment and assign weights that are higher for people with less noisy behaviors.

Applying our approach of using variance-weighted estimators to a corpus of real A/B tests at Facebook, we find opportunity for substantial variance reduction with minimal impact on the bias of treatment effect estimates. Specifically, our results show an average variance reduction of 17 percent, while bias is bounded within 2 percent. In addition, we show that these estimators can achieve improved variance reduction when combined with other standard approaches, such as regression adjustment (also known as CUPED, a commonly used approach at Facebook), demonstrating that this method complements existing work. Our approach has been adopted in several experimental platforms within Facebook.

How we did it

There are several ways in which one can estimate the variance for each unit, and this is still an active area of research. We studied unpooled estimators (using the pre-experiment user-level sampling variance), building a machine learning model to predict out-of-sample variance from features, and using Empirical-Bayes estimators to pool information across those using our platform.

Statistically, we prove that the amount of variance reduction one can achieve when weighting by variance is a function of the coefficient of variation of the variance of experimental users, or roughly, how variable people are in their variability. Details of this proof can be found in our paper.

We tested these approaches on Facebook data and experiments. Figure 1, below, shows how better estimates of in-experiment unit-level variance provide much larger variance reduction. Poorer models of user-level variance can actually increase variance of the estimator, so good estimation is important. To demonstrate that variance-weighted estimators are likely to be useful in practical settings, we collected 12 popular metrics used in A/B tests at Facebook (such as likes, comments, posts shared, and so on) to estimate the predictability of the variance for each metric and its coefficient of variation. The results, shown in Figure 2, indicate that the variance of most of the metrics is highly predictable (as measured using R^2). In addition, the coefficient of variation of the variances is large enough that they can be used effectively in a variance-weighted estimator.

We took a sample of 100 Facebook A/B tests that experimented for an increase in time spent, with the average sample size of each test at around 500,000 users. Before analyzing the results of each test, we assembled the daily time spent for each user in the month prior to the experiment and estimated the variance for each user. To see how accurate the estimated variance of each user was, we compared how well the pre-experiment variance correlated with the post-experiment variance. The results showed an R^2 of 0.696 and a Pearson correlation of 0.754, indicating that the pre-exposed variances, when calculated over an extended period of time, do show reasonable estimations of post-exposed variance.

Next, for each experiment, all users were ranked based on their estimated variance and applied stratification, as in section 4.1 of our paper. To do this, we divided users into quantiles based on pre-experiment estimated variance, and then we calculated the sample variance of the experiment based on various numbers of quantiles. Across all experiments, we found an average of 17 percent decrease in variance with less than 2 percent bias. We also found that our approach worked well with other popular variance reduction approaches, such as CUPED. Table 1, below, shows that we can achieve close to 50 percent variance reduction when both approaches are used together.

What’s next?

There are several opportunities to explore in future work. In particular, there may be significant gains in devising conditional variance models that estimate variance more accurately. Figure 1 showed in simulations how increased estimate qualities can improve variance reduction, suggesting very large gains possible for more precise estimation. Moreover, we would like to understand how variance-weighted estimators may improve the variance reduction observed from other approaches (such as machine learning–based methods), as well as analytically understand the interactions when using multiple variance reduction approaches at once.

The post Increasing the sensitivity of A/B tests by utilizing the variance estimates of experimental units appeared first on Facebook Research.

Read More

Facebook researchers share stories of resilience for Latinx and Hispanic Heritage Month

Latinx and Hispanic Heritage Month is a time to celebrate and honor the Latinx community, elevate underrepresented voices, and hear people’s stories of resilience. As a company, Facebook is celebrating by launching new app features, providing resources and support for Latinx small business owners, and adding new Latinx-focused content on Facebook Watch.

To mark the end of Latinx and Hispanic Heritage Month, we asked four Facebook researchers to share their own journeys and how they celebrate their roots. Through their stories, they illustrate the value of mentorship, intersectionality, community, and resilience for underrepresented minorities in the research community.

Celebrating your roots

At Facebook, we encourage everyone to bring their authentic selves to work. For Nicolas Stier, Co-Director of Core Data Science (CDS), that sense of pride was inspired by his peers in academia.

“I’m an Argentine-born Latino who moved to the U.S. at the end of the nineties. As a new immigrant and a young person, my first reaction in a new culture was to try and assimilate as much as possible. That first reaction allowed me to get a sense of my surroundings, the norms, and the culture of the people around me. However, I eventually noticed people around me were from a diversity of countries and cultures. This reinforced pride in my own culture and made me want to share who I am with my new friends. After all, my identity and roots are an important part of me, and they have shaped my values in life.

“As I was in college pre-internet, I had a very incomplete understanding, to say the least, about how to pursue studies abroad. I had the luck of meeting an Uruguayan-born academic who had been invited to teach a class in a Latin American Computer Science Summer School. After some time working together and exchanging emails, he saw my potential and offered to cover housing and give me a stipend so I could spend six months with him and his group doing research in France. Through that experience, I came to understand that things are indeed possible if one works hard. That gave me the courage to apply to PhD programs in the U.S.

“Celebrating Latinx and Hispanic Heritage Month and other diversity-related events is important because it brings us closer together and helps raise awareness of other minority groups and their perspectives. As a company, we must continue to increase diversity so Latinxs and other underrepresented minorities have a stronger voice.”

Finding your voice

Fostering an inclusive environment is an important part of reaching our “50 in 5” goal, which aims to increase our workforce to at least 50 percent underrepresented people by 2024. Marisol Martinez Escobar, a UX Researcher at WhatsApp, shares the important role that mentorship has played in her career so far, as well as her story of resilience.

“At my previous job, when I was starting my career, I kept following my parents’ advice: Work hard, keep your head down, don’t rock the boat. And those can be great in a way because they help to build community and relationships. But they can be barriers to success as well. I was lucky enough that I had mentors and champions who supported and elevated my voice, and told me to have confidence in my expertise, my voice.

“Through them, I understood the importance of advocating for myself and being more vocal. Because these mentors understood me culturally and where I was coming from, they were able to recognize the ways I needed to grow and offered a safe space to do so.

“I’m a queer Mexican immigrant, and I came to the U.S. when I was 19. I think the Latinx community is a very resilient group of people. I grew up seeing how my family overcame challenge after challenge with very few resources and a lot of humor. And I see this in our Facebook community as we rise to the occasion to become better people, to create a better culture inside and outside of Facebook.”

Acknowledging the intersections of identity

Latinxs are the largest ethnic minority in the U.S., with Hispanic, African, and Indigenous heritage rooted in North, Central, and South America, as well as the Caribbean. Pablo Barberá, Research Scientist at CDS, comments on the diversity inherent in his community.

“One of the aspects of the Latinx and Hispanic community I’m proudest of is its broad diversity in terms of race, country of origin, traditions, and culture, and how our unique differences are embraced and celebrated.

“As someone who recently immigrated to the U.S., [I think] Latinx and Hispanic Heritage Month is a powerful way to celebrate and honor past generations of Hispanic and Latin American descent, and to learn about their important contributions to the history and culture of this country. It is also a good opportunity to remind ourselves of the systemic disadvantages that many members of our community still face today, and to take action that can promote equal access to opportunities.

“I celebrate my roots by giving back to the community. In my previous position as a faculty member at University of Southern California, I made an active effort to offer mentorship opportunities to first-generation Latinx college students. I feel it is so important to acknowledge my privilege and all the help I myself received earlier in my career by paying it forward.

“At Facebook, my current research examines how we can make our platform a space where everyone, no matter who they are, feels empowered to actively participate in political conversations. In this work, embracing my roots means making an active effort to understand the diversity within our community, and how our products should consider those differences to make everyone feel welcome.”

Keeping our communities safe

At Facebook, we are committed to building a product that supports and uplifts marginalized communities, especially when safety is involved. Ignacio Contreras tells us what motivates him in his role as a UX Researcher in Community Integrity.

“A lot of my personal motivation to keep communities safe at Facebook comes from having a strong sense of community myself. This comes directly from my culture, where our community becomes family and we feel a responsibility to keep it safe, healthy, and thriving. My Latinx background has been an asset in what I do, but also in the way I relate to others and help them succeed.

“When I think about working in the tech industry, coming from an underrepresented group can put you at a disadvantage. There are a lot of obstacles we have to overcome to get where we are and keep succeeding. While it’s important to acknowledge that truth, we can’t let it dishearten us and defeat us, so resilience is key. Some of the most resilient people I know who inspire me every day are my Latinx peers and people from other underrepresented groups.

“I immigrated to the U.S. from Chile with my family when I was 18. As a researcher, I am naturally curious about the factors that shape people’s behavior, and culture plays a crucial role. Because I lived half my life in Chile and half in the U.S., I have personally experienced how our behavior is shaped by culture and the way we are raised. The influence of my own family and culture is evident in how I approach the work I do at Facebook and what I value in life.

“Having been raised in a different culture than the one you have to live and work in often comes with challenges, so it’s important to at least take some time during Latinx and Hispanic Heritage month to reflect on the positives and celebrate our culture and the diversity of the Latinx experience in the U.S.”

Tackling the world’s most complex technology challenges requires a diverse set of backgrounds and experiences. Diversity enables us to build better products, make better decisions, and better serve our communities.

We are proud of our attention to the Latinx experience across our apps and technologies, often thanks to the many Latinx people who work at Facebook. However, there is still work to be done. To learn more about our diversity and inclusion efforts at Facebook, visit our website.

The post Facebook researchers share stories of resilience for Latinx and Hispanic Heritage Month appeared first on Facebook Research.

Read More

Announcing the winners of request for proposals on agent-based user interaction simulation

In May, we launched a request for research proposals in agent-based user interaction simulation to find and fix integrity and privacy issues. Today, we’re announcing the recipients of these research awards.
View RFPSoftware systems increasingly support communities of users who interact through the platform, elevating the importance and impact of research on integrity and privacy. How do we ensure that such communities remain safe and their data remains private? To tackle these challenges, Facebook is undertaking research and development on a web-enabled simulation (WES) system called WW.

“The WES research agenda offers so many fascinating new scientific challenges. We cannot hope to tackle them all ourselves,” says Facebook Research Scientist Mark Harman. “We are really excited that the strong response to this call will help to build collaboration and partnerships aimed at tackling these challenges.”

We received 86 proposals from 18 countries and 63 universities. Thank you to everyone who took the time to submit a proposal, and congratulations to the winners.

Research award recipients

A game-theoretic approach to evolving and analysing mechanism design in WES
Aldeida Aleti, Chong Chun Yong, Julian Garcia Gallego (Monash University)

Agent-based simulation for public procurement efficiency
Marcelin Joanis, Andrea Lodi, Igor Sadoune (Polytechnique Montréal)

Empirical game-theoretic analysis for web-enabled simulation
Michael P. Wellman, Mithun Chakraborty (University of Michigan)

Identify metamorphic relations for testing web-enabled simulation systems
Pak Lok Poon, Tsong Yueh Chen (Central Queensland University)

MoCA: Multi-objective co-evolutionary learning agents
Kalyanmoy Deb, Vishnu Boddeti (Michigan State University)

Odbody: An ethics and privacy guardian angel for social media users
Munindar P. Singh, Nirav Ajmeri (North Carolina State University)

Planning to induce emotion labels in a social media network
R. Michael Young (University of Utah)

Simulating a bad actor with knowledge graph-assisted action set generation
Ling Chen, Ivor Tsang (University of Technology Sydney)

Finalists

A bot scheduler for web-enabled simulations
Giovanni Denaro, Martin Tappler, Mauro Pezzè, Valerio Terragni (University of Milano-Bicocca)

AgenTest: A collaborative platform for human testers and test agents
Filippo Ricca, Lorenzo Rosasco, Viviana Mascardi (University of Genova)

Co-evolutionary iterated games to dynamically model bad-actor behaviour
Martin Shepperd (Brunel University)

Co-opetitive game theory for web-enabled simulation
Dr. Shaurya Agarwal (University of Florida)

Combinatorial reap-reward approach to expose privacy and trust attacks
Hyunsook Do (University of North Texas)

Detecting privacy leaks in WES via differential testing and diversification
Kangjie Lu (University of Minnesota Twin Cities)

DOTCOM: Deriving automated tests from conversation mutations
Rumyana Neykova, Giuseppe Destefanis, Stephen Swift, Steve Counsell (Brunel University)

Looking for interactions in the crowd: Using search and self-adaptation
Myra Cohen (Iowa State University Foundation)

MINDSET: Multi-agent-based socio-emotional testing
Rui Filipe Fernandes Prada, Manuel Lopes, Pedro Fernandes, Saba Ansari, Tanja E. J. Vos, Wishnu Prasety (INESC-ID)

Multi-agent-based automated data privacy testing for mobile apps
Yuan Tian, Christian Muise, Xuan-Bach D. Le (Queen’s University)

NLP-driven search-based fuzzing of systems with natural language interfaces
Phil McMinn, Gregory M. Kapfhammer, Mark Stevenson, Owain Parry (University of Sheffield)

SANS-T: Strategic agents network for social testing
Rocco Oliveto, Simone Scalabrino (University of Molise)

Synthesize realistic agents based on behavior examples
Harald C. Gall, Pasquale Salza (University of Zurich)

Taming deep learning (making it faster, more explainable)
Tim Menzies (North Carolina State University)

Towards multi-agent imitation learning in real world
Changyou Chen (University at Buffalo, SUNY)

Using SBSE and web-enabled simulation to detect adversaries
Kevin Leach, Westley Weimer (University of Michigan)

The post Announcing the winners of request for proposals on agent-based user interaction simulation appeared first on Facebook Research.

Read More

Friendship across Europe: How geography and history shape social networks

Social connections shape many aspects of global society. While understanding the geographic structure of these connections is important for a wide range of social science and public policy issues, researchers have traditionally been limited by a lack of large-scale representative data on connectedness. Using the Social Connectedness Index (SCI), an aggregated measure constructed from the friendship networks of Facebook’s more than 2.5 billion monthly active users, we study social connections between European regions. Our results suggest that geographic distance and political borders are important determinants of European connectedness. In fact, we find that the relationship between borders and connectedness persists even long after boundaries change (for example, within former Czechoslovakia and the Austro-Hungarian Empire). We also find that social connections in Europe are stronger between regions with residents of similar ages and education levels, as well as between regions that share a language and religion. In contrast, European region-pairs with dissimilar incomes tend to be more connected, likely due to patterns of migration.

Social Connectedness Index

The SCI uses aggregated friendship connections on Facebook to measure the intensity of connectedness between locations. Locations are assigned to users based on information they provide, connection information, and location services they have opted into. These friendships are used to estimate the probability that a pair of users in these geographies are Facebook friends and mapped to an index score called the Social Connectedness Index. If the SCI is twice as large between two pairs of geographies, it means users in the first geography-pair are about twice as likely to be connected compared with users in the second geography-pair.

More details on the methodology can be found here and in the paper Social Connectedness: Measurement, Determinants, and Effects, published in the Journal of Economic Perspectives.

Analysis

To explore the factors that shape social connectedness in Europe, we looked at SCI between NUTS2 regions, which have between 800,000 and 3 million inhabitants. Using this data, we first constructed a number of case studies. For example, we plotted the social connectedness of the Limburg and Namur regions in Belgium. We found the strongest social connections for both were to other areas nearby within Belgium. Yet, while the capitals of the two regions (Hasselt and Namur, respectively) are less than 70 km apart, the two regions’ connections outside Belgium differ substantially. The official and most commonly spoken language in Limburg is Dutch, whereas in Namur it is French. Accordingly, Limburg is more strongly connected to the entire Netherlands to the north, and Namur is more strongly connected to areas throughout all of France to the south. This suggests that language has an important relationship with patterns of connectedness.

We then sought to understand how patterns of connectedness would be reflected if we created communities of 20 and 50 regions with strong connections to each other (instead of the existing 37 countries). To do so, we created clusters that maximize within-cluster pairwise social connectedness using hierarchical agglomerative linkage clustering.

In the 20-unit map, nearly all the community borders (denoted by a change in area color) line up with country borders (denoted by large black lines). This suggests that individuals are more likely to be connected to distant individuals within their own country than equally distant or closer individuals in other countries. Furthermore, cross-country communities mostly line up with historical borders: For example, every region in the countries that made up Yugoslavia until the early 1990s (NUTS2 regions are defined for Slovenia, Croatia, Serbia, North Macedonia, and Montenegro) are grouped together in one community. We also see the importance of migration in shaping connectedness: Outer London West and North West, which have welcomed a large number of Romanian immigrants in recent years, are grouped together with Romania.

In the 50-unit map, countries begin to break apart internally. Most of these resulting subcountry communities are spatially contiguous, consistent with distance being an important determinant of social connections. We also see linguistic communities form: Belgium splits into French- and Dutch-speaking communities, and Catalan and Andalusian Spanish communities emerge in Spain.

Finally, we used a formal regression approach to assess the relationship between certain factors and European connectedness. Consistent with our exploration, we found that connections are strongest between areas that are physically close to each other: A 10 percent increase in distance is associated with a 13 percent decline in social connectedness. Social connectedness also drops off sharply at country borders. Controlling for geographic distance, the probability of friendship between two individuals living in the same country is five to 18 times as large as it is for two individuals living in different countries.

Using a number of 20th-century European border changes, we also found that this relationship between political borders and connectedness can persist decades after boundary changes. For example, we found higher social connectedness across regions that were originally part of the Austro-Hungarian Empire, even after controlling for distance, current country borders, and a number of other relevant factors.

In addition to distance and political borders, we found that regions that are more similar along demographic measures such as language, religion, education, and age are more socially connected. In particular, social connectedness between two regions with the same most common language is about 4.5 times larger than for two regions without a common language (again controlling for same and border country effects, distance, and other factors). In contrast, we saw that pairs of regions with dissimilar incomes are more connected. Our exploratory analyses suggest this trend may be explained by patterns of migration from regions with lower average incomes to regions with higher average incomes.

A full version of our working paper with additional details on our methodology is available here.

The social connectedness data used is available here.

The post Friendship across Europe: How geography and history shape social networks appeared first on Facebook Research.

Read More

Facebook hosts virtual 2020 Fellowship Summit

The Facebook Fellowship Program supports top PhD students from around the world in fields related to computer science and engineering. The program includes an invitation to the annual Fellowship Summit, which is an opportunity for Fellows to network with one another, present their work, meet Facebook researchers and recruiters, and more.

Apply

Due to COVID-19, the Fellowship Summit was fully virtual and spanned September 8 to September 18. “We’ve embraced this new challenge of planning virtual events, which has provided unique opportunities for the summit,” says Alisa Futriski, Program Coordinator for the Fellowship Program. “For example, because of increased scheduling flexibility and no travel restriction issues, we were able to bring together a particularly robust group of presenters from the Facebook Research community.”

One of the presenters was Facebook Chief Technology Officer Mike Schroepfer, who kicked off the virtual summit with a welcome video. Fellows also heard from research area experts and executives, such as VP of AI Jérôme Pesenti, Novi Head Economist Christian Catalini, Probability Research Scientist Mark Harman, Data for Good Public Policy Research Manager Kelsey Mulcahy, and many more.

“In previous years, Fellows have been given the opportunity to present their research in poster sessions during the summit,” says Sharon Ayalde, Fellowship Program Manager. “This year, in an effort to increase engagement virtually, we asked the Fellows to record presentations of their current research for the summit. These videos were available for all attendees and are now featured for anyone to browse on their Fellow profiles.”

The two-week event also included several Q&As with research-area-specific recruiters for Fellows interested in internships and full-time positions. To complement these Q&As, we organized a panel of several past Fellows who went on to work at Facebook as research interns, full-time researchers, or both: Mark Jeffrey (2017), Eden Litt (2014), Moses Namara (2020), Brandon Schlinker (2016), and Greg Steinbrecher (2017).

Applications for the 2021 Fellowship cohort opened on August 10 with a deadline of October 1, and winners are typically announced in the January after applications close. For more information and to apply, visit the Fellowship page.

The post Facebook hosts virtual 2020 Fellowship Summit appeared first on Facebook Research.

Read More

Robust market equilibria: How to model uncertain buyer preferences

Robust market equilibria: How to model uncertain buyer preferences

What we did

The research in our paper “Robust market equilibria with uncertain preferences,” published at AAAI 2020, is motivated by allocation problems in online markets. Real-world examples of online markets include the following:

  • Online advertising: How should ads be allocated to impressions?
  • Ride allocation: How should ride-sharing platforms allocate drivers to riders?
  • Recommendation systems: How should online recommendation systems (for example, Facebook Jobs, which is a jobs recommendation system) allocate job recommendations to viewers?

In all of these examples, the market consists of buyers (i.e., advertisers) and items (i.e., ad impressions). The buyers have preferences, which are captured by their utility functions. When the platform makes an allocation of an item to a buyer, the buyer receives some utility. In exchange, the buyer often makes a payment of a real or fictitious currency. The allocation rule (who gets what) and the payment rules (pricing of the items) involved are important design choices because they influence whether participants in the market find satisfaction by participating.

A rich body of work, originally by Eisenberg and Gale (1959) in economics and computer science, investigates questions like these. The main insight is the introduction of the notion of a market equilibrium — wherein allocations and prices are determined so that each buyer is “satisfied” with the outcome.

An important challenge within this context is that markets are highly uncertain. At large scales (think billions of buyers and items), it is practically impossible to know the preferences of each buyer perfectly. (This theme has been extensively investigated in “Computing large market equilibria using abstractions,“ and is in a sense a starting point for this work.) In practice, platforms will have machine learning models to predict these preferences from extremely sparse data. These machine learning models are imperfect and make errors in prediction. The challenge we seek to investigate is how do we make allocation and pricing decisions in the face of this uncertainty?

Ideally, we would like our allocation and pricing decisions to be robust against the uncertainty. That is to say, we would like it so that buyers remain “satisfied” when the platform makes decisions with uncertain data. To that end, the work described in our recent paper extends the notion of a market equilibrium to the notion of a robust market equilibrium (RME). The focus of the paper is developing computational tools for RME, as well as applications of the same on some real-world data sets.

How we did it

We start with the decades-old result by Eisenberg and Gale, which answers the question of how to compute market equilibria. (As a caveat, we exclusively work in the so-called divisible goods setting.) They propose doing so by solving a convex optimization program that involves maximizing the geometric mean of all agents’ utilities (also known as Nash Welfare) subject to resource constraints. The key intuition here is that maximizing the geometric mean ensures that allocations cannot be too biased against any individual agent. For example, if we allocate no items to any individual agent, the Nash Welfare is zero.

An analysis of the optimality conditions of the Eisenberg-Gale convex program immediately reveals a host of attractive properties, including the proof that the resulting solution is a market equilibrium.

In order to model uncertain markets, we view the parameters (i.e., the utilities) of agents participating in the market as uncertain. There are many different ways of modeling this uncertainty, resulting in different uncertainty models. We propose and investigate a few natural models in our paper.

We then bring in the main idea of the paper, i.e., the application of a technique called robust optimization to the Eisenberg-Gale convex program. Robust optimization, which is a fairly mature subarea within optimization, deals with solving uncertain optimization problems. In a typical robust optimization problem, the parameters of the problem (the cost function and constraints) are uncertain — or they could be adversarially chosen from a known set.

The technique of robust optimization then seeks to find a solution that produces the best outcome possible against the (worse-case) adversarial choice of parameters from the uncertainty set. Interestingly, many classes of robust optimization problems (including the EG program, as we show in our paper) can be reformulated as (vanilla, certain) optimization problems via convex duality.

After applying the technique of robust optimization, we show a few interesting economic properties of the “robustified” Eisenberg-Gale convex program. One such result is that the solution of the robust variant can be thought of as another market equilibrium where each individual buyer seeks to maximize their uncertain utilities robustly — or, in other words, gain as much utility as possible in the face of adversarial uncertainty. Robust counterparts of a number of other classical properties also follow.

Next, we’ll discuss the application of these ideas to a real-world data set that provides preference information in a recommendation system, namely the MovieLens data set. We perform a thought experiment where we have users choosing movies, with the MovieLens data set indicating how much utility a user would get by watching a particular movie. However, there is a limited supply of movies, so then we would like to allocate movies judiciously, which is using a market equilibrium.

Figure 1
Figure 1 shows the behavior of the Nash Welfare (y-axis) as the size of the uncertainty set (x-axis) grows. The blue line shows the Nash Welfare if we had ignored the uncertainty and simply used the equilibrium based on the Eisenberg-Gale program. The green line shows the Nash Welfare when using the robust solution. There are two trends to note: First, the performance degrades as uncertainty increases, as expected. Second, by a widening margin, the robust solution outperforms the uncertainty-agnostic solution.

Other aspects of the solution can also be investigated. As an example, we consider robust envy — which is a measure of how dissatisfied users are with their own allocation of movies, relative to others under preference uncertainty — can be compared in the uncertain market when using the robust against the vanilla solution.

Figure 2
In Figure 2, the green plot shows the distribution of users’ robust envy using the robust solution in comparison with the blue plot, which shows distribution of envy when using the uncertainty-agnostic solution. The robust envy is significantly lower.

In conclusion, the numerical results reaffirm that modeling uncertainty and accounting for the same when designing allocations result in better market allocations overall.

What’s next

One of our main follow-up questions is how to make such allocations in an online setting. Many practical applications of interest (such as recommender systems) make decisions online, where instead of seeing an entire view of the market and then making allocations, users and buyers arrive online and allocations must be made instantly. Further practical challenges include making allocation decisions within fractions of a second (very low latency), where solving convex optimization problems becomes infeasible.

In conclusion, our research shows that modeling and accounting for uncertainty in markets can dramatically impact the quality of allocations. We introduce the notion of an RME that is uncertainty-aware, show how to compute it numerically, and demonstrate that this uncertainty-aware allocation method is superior to more agnostic allocation methods.

The post Robust market equilibria: How to model uncertain buyer preferences appeared first on Facebook Research.

Read More

Announcing the winners of the Explorations of Trust in AR, VR, and Smart Devices request for proposals

In April, Facebook invited university faculty to respond to the Explorations of Trust in AR, VR, and Smart Devices request for proposals (RFP) to help accelerate research in security, privacy, integrity, and ethics for mixed-reality and smart device products. We were interested in a broad range of topics relating to applications like AR glasses, VR headsets, other AR or VR form-factors, smart home products, and more. Today, we are announcing the recipients and finalists of these research awards.
View RFPThe mixed-reality and smart device industries are continuing to evolve, with unique products, use cases, and devices coming to bear in this new space. With an entirely new category of technologies come entirely new security, privacy, and integrity challenges. But this new category also presents entirely new possibilities and models for considering trust.

The first round of awardees’ research will explore the following:

  • Secure and usable authentication methods in AR and VR devices
  • Unravelling privacy risks of 3D spatial mixed reality data
  • User impacts of novel attacks in AR
  • Building multi-layer defensive frameworks for next-gen intelligent systems
  • Deepfake Detection Methods in 3D mixed reality

Funding these ongoing research efforts will help Facebook Reality Labs and the XR industry better understand and address these nascent risk areas in our efforts to build trustworthy products.

For more information about this RFP, such as topics of interest, eligibility, requirements, and more, visit the application page.

Research award recipients

Principal investigators are listed first unless otherwise noted.

Deepfakes and virtual conferences: Facial motion analysis for ID verification
Ronald Fedkiw (Stanford University)

Experimental analysis of user impacts of novel attacks in augmented reality
Franziska Roesner, Tadayashi Kohno (University of Washington)

Secure and usable authentication for augmented and virtual reality devices
Melanie Volkamer, Peter Mayer, Reyhan Duzgun (Karlsruhe Institute of Technology), Sanchari Das (University of Denver)

SmartShield: Building next generation trustworthy intelligent systems
Sara Rampazzi (University of Florida), Daniel Genkin (University of Michigan)

Unravelling the nascent privacy risks of 3D spatial mixed reality data
Kanchana Thilakarathna, Albert Zomaya (University of Sydney)

Research award finalists

Enforcing spatial content-mediation constraints for augmented reality
Carlos Ernesto Rubio-Medrano (Texas A&M University), Ziming Zhao (University at Buffalo)

Exploring authentication mechanisms for augmented reality glasses
Rahul Chatterjee, Earlence Fernandez (University of Wisconsin – Madison), Yuhang Zhao (Cornell University)

Exploring the design space of trust-worthy voice user interfaces
Yuan Tian (University of Virginia), Sauvik Das (Georgia Institute of Technology)

Intelligent and Interactive Design Fictions to Study and Shape Trust
Elizabeth Murnane (Dartmouth College)

Robust and Efficient Adversarial Machine Learning for Mobile AR
Maria Gorlatova, Neil Gong (Duke University)

Secure Hardware for Trust in AR/VR
G. Edward Suh (Cornell University)

Securing AR/VR and Smart Devices via Cross-domain Low-effort Authentication
Yingying Chen (Rutgers University), Nitesh Saxena (University of Alabama at Birmingham)

Understanding Side-channel Attack Surfaces of AR Devices
Yinqian Zhang (Ohio State University)

The post Announcing the winners of the Explorations of Trust in AR, VR, and Smart Devices request for proposals appeared first on Facebook Research.

Read More

Applying twice: How Facebook Fellow David Pujol adjusted his application for success

The Facebook Fellowship Program supports promising PhD students conducting research in areas related to computer science and engineering. Each year, thousands of PhD students apply to become a Facebook Fellow, and only a handful are selected. To prepare for the end of this year’s application cycle on October 1, we reached out to 2020 Fellow David Pujol to offer some insight for those who may not succeed the first time they apply.
Apply

Pujol is a PhD student at Duke University, advised by Ashwin Machanavajjhala. His research interests lie in the fields of data privacy and algorithmic fairness. Pujol, like other Fellows we’ve chatted with, applied to become a Facebook Fellow in 2018 and was unsuccessful. In 2019, he applied again — and won.

In this Q&A, Pujol tells us about his first approach to his Fellowship application, what changed the second time, what he spent the most time on in his applications, and more.

Q: How did you approach your Fellowship application the first time you applied?

David Pujol: When I first applied, my goal was to impress the reviewers. I tried — in my opinion, unsuccessfully — to make my research sound like some grand project that would change how we view data. In reality, it wasn’t that. I ended up writing a confusing, unfocused, and overly technical piece that failed to convey the information that I wanted it to.

I eventually learned that most projects don’t need to be major, paradigm-changing work. In fact, most are small steps in the right direction.

Q: How did you approach your application the second time around, when you won?

DP: The second time around, I was more process-oriented, and I focused less on making my research look more impressive or more important than it already was. I decided to instead highlight what my research was and why I thought it was important. That meant toning things down — everything from the technical points to the overall vision. Where my first attempt ended with a grand vision of my research, my second attempt highlighted a system that solves a practical problem that has been given little attention.

One of my primary goals was to write in a way that my family (with no technical background) could read my proposal and understand three things: the problem being addressed, why it was important, and why my system was an effective answer.

Q: What made you want to approach things differently?

DP: I think the primary difference was having more experience and more confidence. When I wrote my first proposal, I was completely new to the ins and outs of academia — at least in this field. The second time around, I had a better understanding of how the system at large worked and, more important, I had a bit more confidence in my research. I more fully understood the role my work played and why it was important. I felt better writing about those aspects of the research and didn’t feel the need to overstate their value.

Q: What did you spend the most time on for each application?

DP: Editing, editing, and editing. The first draft of my second application was five pages long. (I think it goes without saying that it was past the word limit.) I asked people to help me edit that draft until it was in a condition that I thought satisfied my standards.

First, I sent my research statement to my adviser, mainly to make sure it was coherent and had no egregious errors. I then edited it myself while taking into consideration the points my adviser had brought up. Then I sent it to one of my friends who had a technical background but not at the PhD level. I asked him to edit so that the draft could be understood by most people, and to make sure nothing went too deep into technicalities. Again, I edited it myself afterwards.

The last round of editing came from my wife, who has no technical background. She adjusted it so that the messaging was clear. We made sure that if I took out everything describing my research, the reader could still tell that there was a problem that needed to be solved and that the solution given addressed that problem and not something else.

I also can’t stress enough how important it is to give yourself some time before editing. It is difficult to be self-critical, especially when you have just finished writing something. Having some time in between edits helps clear up your mind and gives you time to acknowledge your own mistakes.

Q: What advice would you give to someone who doesn’t win the first time they apply?

DP: Just because you didn’t get the fellowship the first time doesn’t mean that your work isn’t relevant enough or that you don’t “deserve it.” It’s always worth trying again once you have some more research experience under your belt. The difference between a good application and a bad one could just be the way you approach things. It might not have anything to do with the research itself. However, it’s different for everyone, so I suggest doing some research to figure out where you could improve.

For me, both of my proposals were about the same project. The only difference is how I went about presenting it.

The post Applying twice: How Facebook Fellow David Pujol adjusted his application for success appeared first on Facebook Research.

Read More