"While algorithm-driven content curations provide online users with convenience and benefits, there are growing concerns about errors, mismatches, and unintended consequences of automations, ranging from creating filter bubbles (Koene et al., 2017; Pariser, 2011), to invading informational privacy (Boyd & Crawford, 2012; Datta et al., 2015), to amplifying human biases and unfairness (Mittelstadt, 2017; Mittelstadt et al., 2016; Schedl et al., 2018). These concerns can be exacerbated by the opaque nature of algorithmic workings and the lack of regulatory frameworks (Pasquale, 2015). In general, individuals do not know what the collection of personal data was based on (i.e., search/purchase history and locations), how or for what purpose their personal information is being exploited by the algorithmic practices, what kinds of biases penetrate machine-learning processes, and the possible discrimination against individual users that could result" (Par 5).