Online information access systems, like recommender systems and search, mediate what information gets exposure and thereby influence their consumption at scale. There is a growing body of evidence that information retrieval (IR) algorithms that narrowly focus on maximizing ranking utility of retrieved items may disparately expose items of similar relevance from the collection. Such disparities in exposure outcome raise concerns of algorithmic fairness and bias of moral import, and may contribute to both representational harms—by reinforcing negative stereotypes and perpetuating inequities in representation of women and other historically marginalized peoples—and allocative harms, from disparate exposure to economic opportunities. In this talk, we present a framework of exposure fairness metrics that model the problem jointly from the perspective of both the consumers and producers. Specifically, we consider group attributes for both types of stakeholders to identify and mitigate fairness concerns that go beyond individual users and items towards more systemic biases in retrieval. The development of expected exposure based metrics also opens up new opportunities and challenges for model optimization. We demonstrate how stochastic ranking policies can be optimized towards target expected exposure and highlight the trade-offs that may exist in optimizing for different fairness dimensions.
Tutorial presented at ACM SIGIR/SIGKDD Africa Summer School on Machine Learning for Data Mining and Search (AFIRM 2020) conference in Cape Town, South Africa.
Learning to rank (LTR) for information retrieval (IR) involves the application of machine learning models to rank artifacts, such as webpages, in response to user's need, which may be expressed as a query. LTR models typically employ training data, such as human relevance labels and click data, to discriminatively train towards an IR objective. The focus of this lecture will be on the fundamentals of neural networks and their applications to learning to rank.
So, You Want to Release a Dataset? Reflections on Benchmark Development, Comm...Bhaskar Mitra
In this talk, I share some of my personal reflections and learnings on benchmark development and community building for making robust scientific progress. This talk is informed by my experience as a developer of the MS MARCO benchmark and as an organizer of the TREC Deep Learning Track. My goal in this talk is to situate the act of releasing a dataset in the context of broader research visions and to draw due attention to considerations of scientific and social outcomes that are invariably salient in the acts of dataset creation and distribution.
Joint Multisided Exposure Fairness for Search and RecommendationBhaskar Mitra
(Slides from my talk at SEA: Search Engines Amsterdam)
Online information access systems, like recommender systems and search, mediate what information gets exposure and thereby influence their consumption at scale. There is a growing body of evidence that information retrieval (IR) algorithms that narrowly focus on maximizing ranking utility of retrieved items may disparately expose items of similar relevance from the collection. Such disparities in exposure outcome raise concerns of algorithmic fairness and bias of moral import, and may contribute to both representational harms—by reinforcing negative stereotypes and perpetuating inequities in representation of women and other historically marginalized peoples—and allocative harms, from disparate exposure to economic opportunities. In this talk, we present a framework of exposure fairness metrics that model the problem jointly from the perspective of both the consumers and producers. Specifically, we consider group attributes for both types of stakeholders to identify and mitigate fairness concerns that go beyond individual users and items towards more systemic biases in retrieval.
Efficient Machine Learning and Machine Learning for Efficiency in Information...Bhaskar Mitra
Emerging machine learning approaches, including deep learning methods, for information retrieval (IR) have recently demonstrated significant improvements in accuracy of relevance estimation at the cost of increasing model complexity and corresponding rise in computational and environmental costs of training and inference. In web search, these costs are further compounded by the necessity to train on large-scale datasets, consume long documents as inputs, and retrieve relevant documents from web-scale collections within milliseconds in response to high volume query traffic. A typical playbook for developing deep learning models for IR involves largely ignoring efficiency concerns during model development and then later scaling these methods by either finding faster approximations of the same models or employing heuristics to reduce the input space over which these models operate. Domain knowledge about the specific IR task and deeper understanding of system design and data structures in whose context these models are deployed can significantly help with not only model simplification but also to inform data-structure specific machine learning model design. Alternatively, predictive machine learning can also be employed specifically to improve efficiency in large scale IR settings. In this talk, I will cover several case studies for both improving efficiency of machine learning models for IR as well as direct application of machine learning to improve retrieval efficiency, and conclude with a brief discussion on potential future directions for efficiency-sensitive benchmarking of machine learning models for IR.
What’s next for deep learning for Search?Bhaskar Mitra
In this talk, I will share some of my personal reflections on the progress in the field of neural IR and some of the ongoing and future research directions that I am personally excited about. This talk will be informed by my own research in this area as well as my experience both as a developer/organizer of the MS MARCO benchmark and the TREC Deep Learning Track and as an applied researcher previously working on web scale search systems at Bing. My goal in this talk would be to move the conversation beyond neural reranking models towards a richer and bolder vision of search powered by deep learning.
Deep neural methods have recently demonstrated significant performance improvements in several IR tasks. In this lecture, we will present a brief overview of deep models for ranking and retrieval.
This is a follow-up lecture to "Neural Learning to Rank" (https://www.slideshare.net/BhaskarMitra3/neural-learning-to-rank-231759858)
Tutorial presented at ACM SIGIR/SIGKDD Africa Summer School on Machine Learning for Data Mining and Search (AFIRM 2020) conference in Cape Town, South Africa.
Learning to rank (LTR) for information retrieval (IR) involves the application of machine learning models to rank artifacts, such as webpages, in response to user's need, which may be expressed as a query. LTR models typically employ training data, such as human relevance labels and click data, to discriminatively train towards an IR objective. The focus of this lecture will be on the fundamentals of neural networks and their applications to learning to rank.
So, You Want to Release a Dataset? Reflections on Benchmark Development, Comm...Bhaskar Mitra
In this talk, I share some of my personal reflections and learnings on benchmark development and community building for making robust scientific progress. This talk is informed by my experience as a developer of the MS MARCO benchmark and as an organizer of the TREC Deep Learning Track. My goal in this talk is to situate the act of releasing a dataset in the context of broader research visions and to draw due attention to considerations of scientific and social outcomes that are invariably salient in the acts of dataset creation and distribution.
Joint Multisided Exposure Fairness for Search and RecommendationBhaskar Mitra
(Slides from my talk at SEA: Search Engines Amsterdam)
Online information access systems, like recommender systems and search, mediate what information gets exposure and thereby influence their consumption at scale. There is a growing body of evidence that information retrieval (IR) algorithms that narrowly focus on maximizing ranking utility of retrieved items may disparately expose items of similar relevance from the collection. Such disparities in exposure outcome raise concerns of algorithmic fairness and bias of moral import, and may contribute to both representational harms—by reinforcing negative stereotypes and perpetuating inequities in representation of women and other historically marginalized peoples—and allocative harms, from disparate exposure to economic opportunities. In this talk, we present a framework of exposure fairness metrics that model the problem jointly from the perspective of both the consumers and producers. Specifically, we consider group attributes for both types of stakeholders to identify and mitigate fairness concerns that go beyond individual users and items towards more systemic biases in retrieval.
Efficient Machine Learning and Machine Learning for Efficiency in Information...Bhaskar Mitra
Emerging machine learning approaches, including deep learning methods, for information retrieval (IR) have recently demonstrated significant improvements in accuracy of relevance estimation at the cost of increasing model complexity and corresponding rise in computational and environmental costs of training and inference. In web search, these costs are further compounded by the necessity to train on large-scale datasets, consume long documents as inputs, and retrieve relevant documents from web-scale collections within milliseconds in response to high volume query traffic. A typical playbook for developing deep learning models for IR involves largely ignoring efficiency concerns during model development and then later scaling these methods by either finding faster approximations of the same models or employing heuristics to reduce the input space over which these models operate. Domain knowledge about the specific IR task and deeper understanding of system design and data structures in whose context these models are deployed can significantly help with not only model simplification but also to inform data-structure specific machine learning model design. Alternatively, predictive machine learning can also be employed specifically to improve efficiency in large scale IR settings. In this talk, I will cover several case studies for both improving efficiency of machine learning models for IR as well as direct application of machine learning to improve retrieval efficiency, and conclude with a brief discussion on potential future directions for efficiency-sensitive benchmarking of machine learning models for IR.
What’s next for deep learning for Search?Bhaskar Mitra
In this talk, I will share some of my personal reflections on the progress in the field of neural IR and some of the ongoing and future research directions that I am personally excited about. This talk will be informed by my own research in this area as well as my experience both as a developer/organizer of the MS MARCO benchmark and the TREC Deep Learning Track and as an applied researcher previously working on web scale search systems at Bing. My goal in this talk would be to move the conversation beyond neural reranking models towards a richer and bolder vision of search powered by deep learning.
Deep neural methods have recently demonstrated significant performance improvements in several IR tasks. In this lecture, we will present a brief overview of deep models for ranking and retrieval.
This is a follow-up lecture to "Neural Learning to Rank" (https://www.slideshare.net/BhaskarMitra3/neural-learning-to-rank-231759858)
Find it! Nail it!Boosting e-commerce search conversions with machine learnin...Rakuten Group, Inc.
The document discusses learning-to-rank models for improving search relevance in e-commerce. It describes how traditional information retrieval models do not scale well to modern needs, while learning-to-rank methods can handle thousands of features and implicit user feedback data. The document reports that using listwise learning-to-rank with NDCG as the loss function improved NDCG by 15.6% and increased conversion rates by 7.5% on e-commerce data. It concludes that deep neural network methods may now outperform traditional machine learning for information retrieval tasks.
This document provides an introduction to polyglot processing using various big data frameworks. It discusses the lambda and kappa architectures for handling batch, micro-batch, and streaming workloads. The document then demonstrates Apache Spark, Storm, Kafka and Redis for stream processing and compares these tools to Flink. It concludes that polyglot processing allows for any data type or workload to be handled and that frameworks like Spark, Storm and Flink each have strengths for distributed, real-time computation.
Anatomy of an eCommerce Search Engine by Mayur DatarNaresh Jain
This document discusses search architecture and optimization for e-commerce platforms. It describes how search is a critical feature that powers recommendations and sales. Key challenges include large catalogs that change frequently, diverse user needs like geo-specific ranking, and balancing multiple objectives. The document outlines the technical infrastructure supporting search, including serving architecture, indexing workflows, and approaches to improve quality like query understanding and personalization.
A fundamental goal of search engines is to identify, given a query, documents that have relevant text. This is intrinsically difficult because the query and the document may use different vocabulary, or the document may contain query words without being relevant. We investigate neural word embeddings as a source of evidence in document ranking. We train a word2vec embedding model on a large unlabelled query corpus, but in contrast to how the model is commonly used, we retain both the input and the output projections, allowing us to leverage both the embedding spaces to derive richer distributional relationships. During ranking we map the query words into the input space and the document words into the output space, and compute a query-document relevance score by aggregating the cosine similarities across all the query-document word pairs.
We postulate that the proposed Dual Embedding Space Model (DESM) captures evidence on whether a document is about a query term in addition to what is modelled by traditional term-frequency based approaches. Our experiments show that the DESM can re-rank top documents returned by a commercial Web search engine, like Bing, better than a term-matching based signal like TF-IDF. However, when ranking a larger set of candidate documents, we find the embeddings-based approach is prone to false positives, retrieving documents that are only loosely related to the query. We demonstrate that this problem can be solved effectively by ranking based on a linear mixture of the DESM and the word counting features.
Learning to rank (LTR) for information retrieval (IR) involves the application of machine learning models to rank artifacts, such as items to be recommended, in response to user's need. LTR models typically employ training data, such as human relevance labels and click data, to discriminatively train towards an IR objective. The focus of this tutorial will be on the fundamentals of neural networks and their applications to learning to rank.
Présentation générales du Big Data et zoom sur des cas d'usage dans l'industrie et les services.
Présentation réalisée à l'occasion de l'événement Big data de Niort du 20 mars 2014
A Multi-Armed Bandit Framework For Recommendations at NetflixJaya Kawale
In this talk, we present a general multi-armed bandit framework for recommendations on the Netflix homepage. We present two example case studies using MABs at Netflix - a) Artwork Personalization to recommend personalized visuals for each of our members for the different titles and b) Billboard recommendation to recommend the right title to be watched on the Billboard.
This document discusses how machine learning can be applied to ecommerce and retail applications. It outlines several problems that ML can address, including search ranking, typeahead, spell correction, cold start recommendations, left-hand navigation, query understanding, related searches, product discovery, image similarity, voice search, attribute extraction, user modeling, title generation, and inventory management. It also provides context on data sizes, user behaviors, and the need for models to have fast prediction speeds and work within memory constraints in a production setting.
Vous êtes responsable MOA ou MOE et vous vous interrogez sur les possibilités du Machine Learning ?
Vous avez déjà rapidement entendu parler de classification supervisée, de prédiction, de recommandation … mais vous n’en comprenez pas réellement les tenants et les aboutissants ?
Cette présentation est faite pour vous!
Vous trouverez :
- une définition concise
- les grands principes du ML
- les problématiques auxquelles répond le ML
- les étapes à suivre
- les prémices d’un projet
- les indicateurs à prendre en compte lors du choix de l’algorithme à utiliser
Aujourd’hui, tous les métiers sont concernés par le Machine Learning, alors n’ayez pas peur de vous lancer! C’est à vous!
Si vous avez des questions, les commentaires sont les bienvenus.
Försäkringskassan: Neo4j as an Information Hub (GraphSummit Stockholm 2023)Neo4j
Having introduced Neo4j for specific applications over time, Försäkringskassan (Swedish Social Insurance Agency) is now leaning heavily on Neo4j as a central component in their data management platform. They are becoming data centric and increasingly centering information around the customer.
A Simple Introduction to Neural Information RetrievalBhaskar Mitra
Neural Information Retrieval (or neural IR) is the application of shallow or deep neural networks to IR tasks. In this lecture, we will cover some of the fundamentals of neural representation learning for text retrieval. We will also discuss some of the recent advances in the applications of deep neural architectures to retrieval tasks.
(These slides were presented at a lecture as part of the Information Retrieval and Data Mining course taught at UCL.)
Unified Approach to Interpret Machine Learning Model: SHAP + LIMEDatabricks
For companies that solve real-world problems and generate revenue from the data science products, being able to understand why a model makes a certain prediction can be as crucial as achieving high prediction accuracy in many applications. However, as data scientists pursuing higher accuracy by implementing complex algorithms such as ensemble or deep learning models, the algorithm itself becomes a blackbox and it creates the trade-off between accuracy and interpretability of a model’s output.
To address this problem, a unified framework SHAP (SHapley Additive exPlanations) was developed to help users interpret the predictions of complex models. In this session, we will talk about how to apply SHAP to various modeling approaches (GLM, XGBoost, CNN) to explain how each feature contributes and extract intuitive insights from a particular prediction. This talk is intended to introduce the concept of general purpose model explainer, as well as help practitioners understand SHAP and its applications.
The document discusses the history of large language models including GPT-1, BERT, GPT-2, T5, and GPT-3 from OpenAI and Google. It provides background information on each model including their size, architecture, and contributions to advancing the state-of-the-art in natural language processing. The document also notes how recent models like GPT-3 have achieved general capabilities without any fine-tuning by leveraging their massive scale.
Building an Enterprise Knowledge Graph @Uber: Lessons from RealityJoshua Shinavier
This document summarizes Uber's experience building an enterprise knowledge graph. It notes that Uber has over 200,000 managed datasets and billions of trips served, making it an ideal testbed for a knowledge graph. However, it also outlines several lessons learned, including that real-world data is messy, an RDF-based approach is difficult, and property graphs alone are insufficient. The document advocates standardizing on shared vocabularies, fitting tools and data models to existing infrastructure, and collaborating across teams.
This document provides an overview of deep learning basics for natural language processing (NLP). It discusses the differences between classical machine learning and deep learning, and describes several deep learning models commonly used in NLP, including neural networks, recurrent neural networks (RNNs), encoder-decoder models, and attention models. It also provides examples of how these models can be applied to tasks like machine translation, where two RNNs are jointly trained on parallel text corpora in different languages to learn a translation model.
안녕하세요 딥논읽 입니다 오늘 소개드릴 논문은 'LayoutLM'입니다 !
여러 회사에서 스캔 된 문서의 텍스트를 추출하여 이해하는 기술에 대한 수요가 증가하고 있습니다. 하지만 뒷받침할 모델들이 많이 학습이 되지 않고 있는 상황입니다
문제는 이제 Label된 Dataset이 극도로 부족한데 이런 문제를 해결하기 위해서
Unlabel Dataset을 활용을 해야 하지만 연구가 충분히 이루어지지 못하고 있습니다
기존의 모델들은 OCR같은 사전에 학습된 CV모델만을 활용하거나 반대로 NLP 모델만 활용을 하고 있고 이 두 개 모델을 같이 활용된 pre-training 모델이 존재하지 않습니다
그래서 이 논문에서는 컴퓨터 비전과 NLP 를 동시에 사용하는 pre-training 모델을 사용하는 LayoutLM에 대해 제안합니다!
오늘 논문 리뷰는 딥논읽 자연어 처리팀 박희수 님이 자세한 리뷰 도와주셨습니다.
오늘도 많은 관심 미리 감사드립니다!
The document discusses various topics related to group influences on consumer behavior, including reference groups, opinion leadership, and the diffusion of innovations. Specifically, it provides information on:
1) The top 5 brands in the US in terms of word-of-mouth are Toyota, Walmart, Honda, iPod, and Chevrolet. The internet has aided word-of-mouth through easy sharing via email, blogs, etc.
2) Reference groups influence consumer behavior through informational, normative, and identification influences. The type of reference group influence depends on factors like the consumption situation.
3) Opinion leaders are influential consumers who provide information to others in specific product categories. Marketers target opinion leaders
Work Deviant Behaviour and Team Cooperation in Selected Manufacturing Compani...ijtsrd
This study determined the effect of work deviant behaviour on team cooperation of manufacturing companies in Enugu, Nigeria. Specifically, the study sought to ascertain the effect of workplace aggression and non compliant to organization policy on team cooperation of manufacturing companies in Enugu, Nigeria. Survey research design was adopted. A sample of one hundred and eighty respondents was purposively selected for the study. Data were generated from the questionnaires administered to the respondents. The hypothesis was tested with regression analysis. Based on this, the study revealed that workplace aggression and non compliant to organization policy have a significant effect on team cooperation of manufacturing companies in Enugu, Nigeria. Based on the outcome of this study, it was recommended that corporate managers should focus on the effective management of attitudinal and behavioural outcome so as to foster conducive work environment and position the organization’s image to the public. Chizoba Bonaventure Okolocha | Onwuchekwa Juliet Anuri "Work Deviant Behaviour and Team Cooperation in Selected Manufacturing Companies in Enugu State, Nigeria" Published in International Journal of Trend in Scientific Research and Development (ijtsrd), ISSN: 2456-6470, Volume-7 | Issue-3 , June 2023, URL: https://www.ijtsrd.com.com/papers/ijtsrd57435.pdf Paper URL: https://www.ijtsrd.com.com/management/other/57435/work-deviant-behaviour-and-team-cooperation-in-selected-manufacturing-companies-in-enugu-state-nigeria/chizoba-bonaventure-okolocha
Find it! Nail it!Boosting e-commerce search conversions with machine learnin...Rakuten Group, Inc.
The document discusses learning-to-rank models for improving search relevance in e-commerce. It describes how traditional information retrieval models do not scale well to modern needs, while learning-to-rank methods can handle thousands of features and implicit user feedback data. The document reports that using listwise learning-to-rank with NDCG as the loss function improved NDCG by 15.6% and increased conversion rates by 7.5% on e-commerce data. It concludes that deep neural network methods may now outperform traditional machine learning for information retrieval tasks.
This document provides an introduction to polyglot processing using various big data frameworks. It discusses the lambda and kappa architectures for handling batch, micro-batch, and streaming workloads. The document then demonstrates Apache Spark, Storm, Kafka and Redis for stream processing and compares these tools to Flink. It concludes that polyglot processing allows for any data type or workload to be handled and that frameworks like Spark, Storm and Flink each have strengths for distributed, real-time computation.
Anatomy of an eCommerce Search Engine by Mayur DatarNaresh Jain
This document discusses search architecture and optimization for e-commerce platforms. It describes how search is a critical feature that powers recommendations and sales. Key challenges include large catalogs that change frequently, diverse user needs like geo-specific ranking, and balancing multiple objectives. The document outlines the technical infrastructure supporting search, including serving architecture, indexing workflows, and approaches to improve quality like query understanding and personalization.
A fundamental goal of search engines is to identify, given a query, documents that have relevant text. This is intrinsically difficult because the query and the document may use different vocabulary, or the document may contain query words without being relevant. We investigate neural word embeddings as a source of evidence in document ranking. We train a word2vec embedding model on a large unlabelled query corpus, but in contrast to how the model is commonly used, we retain both the input and the output projections, allowing us to leverage both the embedding spaces to derive richer distributional relationships. During ranking we map the query words into the input space and the document words into the output space, and compute a query-document relevance score by aggregating the cosine similarities across all the query-document word pairs.
We postulate that the proposed Dual Embedding Space Model (DESM) captures evidence on whether a document is about a query term in addition to what is modelled by traditional term-frequency based approaches. Our experiments show that the DESM can re-rank top documents returned by a commercial Web search engine, like Bing, better than a term-matching based signal like TF-IDF. However, when ranking a larger set of candidate documents, we find the embeddings-based approach is prone to false positives, retrieving documents that are only loosely related to the query. We demonstrate that this problem can be solved effectively by ranking based on a linear mixture of the DESM and the word counting features.
Learning to rank (LTR) for information retrieval (IR) involves the application of machine learning models to rank artifacts, such as items to be recommended, in response to user's need. LTR models typically employ training data, such as human relevance labels and click data, to discriminatively train towards an IR objective. The focus of this tutorial will be on the fundamentals of neural networks and their applications to learning to rank.
Présentation générales du Big Data et zoom sur des cas d'usage dans l'industrie et les services.
Présentation réalisée à l'occasion de l'événement Big data de Niort du 20 mars 2014
A Multi-Armed Bandit Framework For Recommendations at NetflixJaya Kawale
In this talk, we present a general multi-armed bandit framework for recommendations on the Netflix homepage. We present two example case studies using MABs at Netflix - a) Artwork Personalization to recommend personalized visuals for each of our members for the different titles and b) Billboard recommendation to recommend the right title to be watched on the Billboard.
This document discusses how machine learning can be applied to ecommerce and retail applications. It outlines several problems that ML can address, including search ranking, typeahead, spell correction, cold start recommendations, left-hand navigation, query understanding, related searches, product discovery, image similarity, voice search, attribute extraction, user modeling, title generation, and inventory management. It also provides context on data sizes, user behaviors, and the need for models to have fast prediction speeds and work within memory constraints in a production setting.
Vous êtes responsable MOA ou MOE et vous vous interrogez sur les possibilités du Machine Learning ?
Vous avez déjà rapidement entendu parler de classification supervisée, de prédiction, de recommandation … mais vous n’en comprenez pas réellement les tenants et les aboutissants ?
Cette présentation est faite pour vous!
Vous trouverez :
- une définition concise
- les grands principes du ML
- les problématiques auxquelles répond le ML
- les étapes à suivre
- les prémices d’un projet
- les indicateurs à prendre en compte lors du choix de l’algorithme à utiliser
Aujourd’hui, tous les métiers sont concernés par le Machine Learning, alors n’ayez pas peur de vous lancer! C’est à vous!
Si vous avez des questions, les commentaires sont les bienvenus.
Försäkringskassan: Neo4j as an Information Hub (GraphSummit Stockholm 2023)Neo4j
Having introduced Neo4j for specific applications over time, Försäkringskassan (Swedish Social Insurance Agency) is now leaning heavily on Neo4j as a central component in their data management platform. They are becoming data centric and increasingly centering information around the customer.
A Simple Introduction to Neural Information RetrievalBhaskar Mitra
Neural Information Retrieval (or neural IR) is the application of shallow or deep neural networks to IR tasks. In this lecture, we will cover some of the fundamentals of neural representation learning for text retrieval. We will also discuss some of the recent advances in the applications of deep neural architectures to retrieval tasks.
(These slides were presented at a lecture as part of the Information Retrieval and Data Mining course taught at UCL.)
Unified Approach to Interpret Machine Learning Model: SHAP + LIMEDatabricks
For companies that solve real-world problems and generate revenue from the data science products, being able to understand why a model makes a certain prediction can be as crucial as achieving high prediction accuracy in many applications. However, as data scientists pursuing higher accuracy by implementing complex algorithms such as ensemble or deep learning models, the algorithm itself becomes a blackbox and it creates the trade-off between accuracy and interpretability of a model’s output.
To address this problem, a unified framework SHAP (SHapley Additive exPlanations) was developed to help users interpret the predictions of complex models. In this session, we will talk about how to apply SHAP to various modeling approaches (GLM, XGBoost, CNN) to explain how each feature contributes and extract intuitive insights from a particular prediction. This talk is intended to introduce the concept of general purpose model explainer, as well as help practitioners understand SHAP and its applications.
The document discusses the history of large language models including GPT-1, BERT, GPT-2, T5, and GPT-3 from OpenAI and Google. It provides background information on each model including their size, architecture, and contributions to advancing the state-of-the-art in natural language processing. The document also notes how recent models like GPT-3 have achieved general capabilities without any fine-tuning by leveraging their massive scale.
Building an Enterprise Knowledge Graph @Uber: Lessons from RealityJoshua Shinavier
This document summarizes Uber's experience building an enterprise knowledge graph. It notes that Uber has over 200,000 managed datasets and billions of trips served, making it an ideal testbed for a knowledge graph. However, it also outlines several lessons learned, including that real-world data is messy, an RDF-based approach is difficult, and property graphs alone are insufficient. The document advocates standardizing on shared vocabularies, fitting tools and data models to existing infrastructure, and collaborating across teams.
This document provides an overview of deep learning basics for natural language processing (NLP). It discusses the differences between classical machine learning and deep learning, and describes several deep learning models commonly used in NLP, including neural networks, recurrent neural networks (RNNs), encoder-decoder models, and attention models. It also provides examples of how these models can be applied to tasks like machine translation, where two RNNs are jointly trained on parallel text corpora in different languages to learn a translation model.
안녕하세요 딥논읽 입니다 오늘 소개드릴 논문은 'LayoutLM'입니다 !
여러 회사에서 스캔 된 문서의 텍스트를 추출하여 이해하는 기술에 대한 수요가 증가하고 있습니다. 하지만 뒷받침할 모델들이 많이 학습이 되지 않고 있는 상황입니다
문제는 이제 Label된 Dataset이 극도로 부족한데 이런 문제를 해결하기 위해서
Unlabel Dataset을 활용을 해야 하지만 연구가 충분히 이루어지지 못하고 있습니다
기존의 모델들은 OCR같은 사전에 학습된 CV모델만을 활용하거나 반대로 NLP 모델만 활용을 하고 있고 이 두 개 모델을 같이 활용된 pre-training 모델이 존재하지 않습니다
그래서 이 논문에서는 컴퓨터 비전과 NLP 를 동시에 사용하는 pre-training 모델을 사용하는 LayoutLM에 대해 제안합니다!
오늘 논문 리뷰는 딥논읽 자연어 처리팀 박희수 님이 자세한 리뷰 도와주셨습니다.
오늘도 많은 관심 미리 감사드립니다!
The document discusses various topics related to group influences on consumer behavior, including reference groups, opinion leadership, and the diffusion of innovations. Specifically, it provides information on:
1) The top 5 brands in the US in terms of word-of-mouth are Toyota, Walmart, Honda, iPod, and Chevrolet. The internet has aided word-of-mouth through easy sharing via email, blogs, etc.
2) Reference groups influence consumer behavior through informational, normative, and identification influences. The type of reference group influence depends on factors like the consumption situation.
3) Opinion leaders are influential consumers who provide information to others in specific product categories. Marketers target opinion leaders
Work Deviant Behaviour and Team Cooperation in Selected Manufacturing Compani...ijtsrd
This study determined the effect of work deviant behaviour on team cooperation of manufacturing companies in Enugu, Nigeria. Specifically, the study sought to ascertain the effect of workplace aggression and non compliant to organization policy on team cooperation of manufacturing companies in Enugu, Nigeria. Survey research design was adopted. A sample of one hundred and eighty respondents was purposively selected for the study. Data were generated from the questionnaires administered to the respondents. The hypothesis was tested with regression analysis. Based on this, the study revealed that workplace aggression and non compliant to organization policy have a significant effect on team cooperation of manufacturing companies in Enugu, Nigeria. Based on the outcome of this study, it was recommended that corporate managers should focus on the effective management of attitudinal and behavioural outcome so as to foster conducive work environment and position the organization’s image to the public. Chizoba Bonaventure Okolocha | Onwuchekwa Juliet Anuri "Work Deviant Behaviour and Team Cooperation in Selected Manufacturing Companies in Enugu State, Nigeria" Published in International Journal of Trend in Scientific Research and Development (ijtsrd), ISSN: 2456-6470, Volume-7 | Issue-3 , June 2023, URL: https://www.ijtsrd.com.com/papers/ijtsrd57435.pdf Paper URL: https://www.ijtsrd.com.com/management/other/57435/work-deviant-behaviour-and-team-cooperation-in-selected-manufacturing-companies-in-enugu-state-nigeria/chizoba-bonaventure-okolocha
The nature of the workforce across the organizations is going through a change because of the increasing significance of the internet in daily life, forces of globalization and various other factors. Companies have to deal with changes in their workforce very effectively so as to survive in current highly competitive marketplace. Specifically, programs and initiatives are to be devised by firms which aim to handle the issues related to ageing population of workforce in the developed nations (Enns 2005). Moreover, they are also aimed to ensure diversity in the workforce so as to ensure the overall growth of the employees in the firm by interacting with employees of different cultural background. Companies are left with no choice but to accommodate workforce who have a preference of new pattern of working because of the change in pattern of employment which includes job sharing and working from home (Remery 2003).
Course Introduction and Lecture slides for "Practice of International Trade", Department of International Business, Tunghai University. Utilizing the Export Import Management System V. 2.0 from JAI International (USA).
This document summarizes and discusses recommender systems. It begins by defining recommender systems and their purpose of presenting personalized recommendations of items likely to interest users based on their profiles and preferences. It then outlines three main recommendation techniques: content-based filtering which uses item attributes to make recommendations; collaborative filtering which identifies similar users to make recommendations; and hybrid filtering which combines the two approaches. Finally, it discusses challenges for non-personalized recommendation systems in serving diverse user groups and notes that personalized approaches may help address this.
The document discusses disability inclusion in the workplace and provides a framework to foster and facilitate disability inclusion called the "A List". The A List consists of 5 parts: Broaden Access, Raise Awareness, Foster Advocacy, Integrate Actions, and Ensure Accountability. It also highlights examples of companies that are leaders in disability inclusion, such as how Dell provides coaching and mentoring to neurodiverse new hires to help them succeed.
Gender board diversity spillovers and the public eyeGRAPE
A range of policy recommendations mandating gender board quotas is based on the idea that "women help women". We analyze potential gender diversity spillovers from supervisory to top managerial positions over three decades in Europe. Contrary to previous studies which worked with stock listed firms or were region locked, we use a large data base of roughly 2 000 000 firms. We find evidence that women do not help women in corporate Europe, unless the firm is stock listed. Only within public firms, going from no woman to at least one woman on supervisory position is associated with a 10-15\% higher probability of appointing at least one woman to the executive position. This pattern aligns with the Public Eye Managerial Theory, suggesting that external visibility influences corporate gender diversity practices. The study implies that diversity policies, while impactful in public firms, have limited effectiveness in promoting gender diversity in corporate Europe.
Fairness-aware Machine Learning: Practical Challenges and Lessons Learned (WS...Krishnaram Kenthapadi
Researchers and practitioners from different disciplines have highlighted the ethical and legal challenges posed by the use of machine learned models and data-driven systems, and the potential for such systems to discriminate against certain population groups, due to biases in algorithmic decision-making systems. This tutorial presents an overview of algorithmic bias / discrimination issues observed over the last few years and the lessons learned, key regulations and laws, and evolution of techniques for achieving fairness in machine learning systems. We will motivate the need for adopting a "fairness by design" approach (as opposed to viewing algorithmic bias / fairness considerations as an afterthought), when developing machine learning based models and systems for different consumer and enterprise applications. Then, we will focus on the application of fairness-aware machine learning techniques in practice by presenting non-proprietary case studies from different technology companies. Finally, based on our experiences working on fairness in machine learning at companies such as Facebook, Google, LinkedIn, and Microsoft, we will present open problems and research directions for the data mining / machine learning community.
Please cite as:
Sarah Bird, Ben Hutchinson, Krishnaram Kenthapadi, Emre Kiciman, and Margaret Mitchell. Fairness-Aware Machine Learning: Practical Challenges and Lessons Learned. WSDM 2019.
This document provides an outline for a research proposal that will examine the labor market effects of mandatory benefit regulations for maids in Ecuador. The researcher will use a difference-in-difference methodology to analyze how mandatory social security enrollment for maids has impacted their labor market outcomes and coverage rates. They will also examine how the increased costs of maid services have impacted work decisions of female heads of households. The study will use employment survey data from Ecuador and may include a qualitative survey of maids. The expected results could provide insights into how social security policies impact vulnerable labor groups and lessons for the design of labor and social protection policies in developing countries.
Excelsior College Business External Environment Report.docxbkbk37
The document discusses various frameworks for analyzing the external environment of a business, including PESTEL, SWOT, Porter's Five Forces model, and strategic group mapping. It provides an overview of each framework and how they can be used to identify opportunities and threats outside a company's control that influence its strategy. The frameworks analyze political, economic, social, technological, environmental, legal factors, as well as competition, suppliers, buyers, substitutes, and new entrants in a company's industry.
From Hype to Reality: AI in Market Research Tom De Ruyck
Galvin is an AI assistant that can help companies better utilize consumer insights by providing marketers easy access to all consumer research data [SENTENCE 1]. Galvin allows companies to have direct access to consumer insights and gives marketers the right insights anywhere, anytime to help address the challenge of properly activating insights [SENTENCE 2]. Galvin can impersonate consumer personas to allow employees to have simulated chats with consumers to better understand them or can provide daily inspiration to help create a consumer-connected mindset [SENTENCE 3].
Alibabas Internal( just internal) EnvironmentTimothy .docxgalerussel59292
Alibaba's Internal( just internal?) Environment
Timothy W.C. Burke, Firas Faraj, Leon Edwards, and Hinde Ahzi
STR/581
January 22, 2015
Dr. Kamal L. Ranasinghe Ph.D., D.B.A.
Running head: ALIBABA'S INTERNAL ENVIRONMENT
1
ALIBABA'S INTERNAL ENVIRONMENT
2
Alibaba's Internal Environment
Alibaba Group Ltd. is the world's leading company in online trade. It helps exporters around the globe connect with buyers in over 190 countries. The research team noted that, Alibaba’s core competency is the choice of leadership in the company. The top executives of Alibaba are educated at prestigious schools is the United States and China. The internal competitive environmental scan will incorporate the environmental factors, strengths, weaknesses and the future of the organizational structure.
External Environmental Factors
Economic and Social Factors: The fundamental information that affects a businessHmmm obviously, my good intent notwithstanding …some pay only scant attention to my reviews . and the environment in which a company operates concerning the direction of the economy is considered economic factors.)) Yes …well , it would , would it not . What else could it be … )) “At the national and international level, it must consider the general availability of credit, the level of disposable income, and the propensity of people to spend” (Pearce & Robinson, 2009). China has been gaining momentum for years to become the world next financial super power. “The social factors that effect a firm involve the beliefs, values, attitudes, opinions, and lifestyles of persons in the firm’s external environment, as developed from cultural, ecological, demographic, religious, educational, and ethnic conditioning” (Wang, 2012). “One of the most profound social changes in recent years has been the entry of significant members of women into the Chinese labor market” (Pearce & Robinson, 2009). In this researcher’s opinion,Hmmm I was under the impression this was intended to be a collaborative research effort BY A TEAM of researchers ?
getting more women in the labor market will increase the productivity of companies in China. Hmmm An overtly sweeping assertion folk . Did you all read this before allowing it to be in this paper ? I speak now to the absolute requirement to substantiate any such sweeping assertion with current , irrefutable, globally-acknowledged citations. Whoever included this , in this paper , please do raise in class and defend this assertion .
Goods and services require expanding and emerging economy resources and infrastructure in order to facilitate increasing sales.
Political and Technological Factors: “The direction and stability of political factors is a major consideration for managers in formulating company strategy. Political factors define the legal and regulatory parameters within which Alibaba operates” (Pearce & Robinson, 2009). China’s regulatory environment has encouraged, but also impedes company’s s.
IRJET- Survey on Facial Feature Detection Methods in Offline Recommendation b...IRJET Journal
This document discusses using facial feature extraction for offline recommendation systems. It proposes using age and gender detection from facial images as input for recommendation algorithms without accessing private user data. Age and gender can help categorize users into basic demographic groups associated with different interests. The paper outlines assumptions for the system, including having a camera and processing power for image analysis. It describes common methods for facial feature extraction and age/gender classification, including using global, local, and hybrid facial features. One proposed approach uses multi-level local phase quantization and support vector machines to classify gender from images. The findings could enable targeted recommendations in physical spaces like stores based on visual analysis of users.
AI & DEI: With Great Opportunities Comes Great HR ResponsibilityAggregage
https://www.humanresourcestoday.com/frs/26184029/ai---dei--with-great-opportunities-comes-great-hr-responsibility
The promise of AI for today’s organizations is real, yet in a frenzied state of experimentation, many stumble to get to a full-scale enterprise. As companies race to discover what generative AI can do, HR must lead conversations about how to balance cutting-edge innovations with integrity, trust, and diversity. Globally, organizations are at a critical intersection of Diversity, Equity, Inclusion, and AI acceleration. We will explore how AI is rapidly transforming workplace dynamics and decision-making processes. The safety and protection of the workforce have never been more important and need to be co-led by HR to prevent biases and ensure fair and equitable representation in systems, hiring, and the workforce evolution.
We'll cover:
• The opportunities that AI presents and the responsibility of HR
• How to enhance diverse perspectives in use cases
• Increasing collaboration between AI Developers, HR, Legal and IT
Paper for 2nd International Conference on Lean Six Sigma for Higher Education
Lean as a management strategy: universal or domain specific adaptation required
Mba2216 business research week 3 research methodology 0613Stephen Ong
The document discusses research methodology and key concepts in developing theories. It defines theory as a formal explanation of how things relate that allows for predictions. The goals of theory are to gain understanding of relationships between phenomena and enable prediction based on that understanding. The document also defines important research terms like concepts, constructs, propositions, variables, and hypotheses that are used to build theories through reviewing literature and logical deduction.
1. IntroductionIn the modern society, all the enterprises in the.docxSONU61709
1. Introduction
In the modern society, all the enterprises in the face of increasingly complex social environment, fickle and variable current social situation in economy, politics, culture and other factors bring both opportunities and challenges to enterprises. To develop in intense market competitions, be prepared is the foundation of all the business activities. To walk on the right way by a clear direction to reach a right goal, planning must be an indispensable and significant process of the business operation.
Strategy is defined as a plan of action designed to achieve a long-term or overall aim by Oxford Dictionary. A strategy is a series of integrated, coordinated action that designed to help organization develop core competencies and access to competitive advantage. Business model is an abstract representation of some aspects of a firm’s strategy; it helps people to understand how a firm can successfully deliver value to its customers. Unlike strategy, business models do not consider competitive positioning. Business strategic model could help organization to set direction and priorities in less resource and time by a scientific way.
In this report, mainly to explore two different strategic models(SWOT analysis, Porter’s value chain) from several aspects including introduction of strategic model, application of models, analysis and comparison about two models. At last, make a short reflective conclusion about the strategic models.
2. SWOT Analysis
2.1 What is SWOT Analysis model
Today’s organizations find themselves operating in an environment that is changing faster than ever before, SWOT analysis is a method for organization analyze the implications of these changes and modify the way to react to the changes. SWOT analysis is a method based on internal and external environment of the organization analysis or a procedural or structural components analysis thereof embodied in establishing the main strengths, weaknesses, opportunities and threats. (Verboncu, 2016). As a basis of strategic analysis and formulation,
SWOT analysis by analyzing the competitive advantage of the firm itself (Strength), competitive disadvantage (weakness), opportunities and threat, this model helps the enterprise get better know about the opportunities and challenges facing themselves, at the same time, it has vital significance for the development of the public division and formulate the future development strategy. Figure 1. SWOT matrix
SWOT matrix (Figure 1) is made for considering four components; important SWOT analysis’s results of the enterprise should be listed in the table. The combination of the four elements defining this model (the model of "fitting" or the model of "alignment") generates, according to experts, four modes of analysis of internal and external factors, as the basis of specific strategies. (Verboncu, 2016) These four factors come into being different strategy combination like SO, ST, WO and WT. (Figure 2)
Figure 2. SWOT Matr ...
The document discusses the role of sustainability in the exercise equipment industry. It outlines how SportsArt ECO-POWRTM harnesses human-generated power from exercise to feed energy back into the power grid. Assessing the cost-benefits of this system can help stakeholders understand the benefits of sustainability practices compared to non-sustainable equipment lines. It also discusses the growing exercise equipment market and competitive industry landscape, as well as definitions and key drivers of sustainability practices.
The document provides instructions for using slides that introduce generative AI and its use in assessments. The slides are designed as a 1-hour lecture that can be delivered live or circulated asynchronously. Staff should delete instruction slides before using the presentation with students. The AI and You teaching toolkit provides further guidance.
This document discusses fairness in machine learning recommendation systems. It begins by outlining the need for fairness to avoid disparate impact on protected groups. It then discusses various sources of bias that can lead to unfair recommendations if not addressed. The document outlines best practices for building fair recommendation systems, including identifying biases, mitigating their impact, monitoring performance metrics, and incorporating fairness constraints. It also discusses how to evaluate and achieve different notions of fairness, including individual, group, and multi-sided fairness. Overall, the document provides an overview of the key challenges and approaches for developing recommendation systems that treat all user groups fairly.
Semelhante a Multisided Exposure Fairness for Search and Recommendation (20)
Search and Society: Reimagining Information Access for Radical FuturesBhaskar Mitra
The field of Information retrieval (IR) is currently undergoing a transformative shift, at least partly due to the emerging applications of generative AI to information access. In this talk, we will deliberate on the sociotechnical implications of generative AI for information access. We will argue that there is both a critical necessity and an exciting opportunity for the IR community to re-center our research agendas on societal needs while dismantling the artificial separation between the work on fairness, accountability, transparency, and ethics in IR and the rest of IR research. Instead of adopting a reactionary strategy of trying to mitigate potential social harms from emerging technologies, the community should aim to proactively set the research agenda for the kinds of systems we should build inspired by diverse explicitly stated sociotechnical imaginaries. The sociotechnical imaginaries that underpin the design and development of information access technologies needs to be explicitly articulated, and we need to develop theories of change in context of these diverse perspectives. Our guiding future imaginaries must be informed by other academic fields, such as democratic theory and critical theory, and should be co-developed with social science scholars, legal scholars, civil rights and social justice activists, and artists, among others.
Neural Information Retrieval: In search of meaningful progressBhaskar Mitra
The emergence of deep learning based methods for search poses several challenges and opportunities not just for modeling, but also for benchmarking and measuring progress in the field. Some of these challenges are new, while others have evolved from existing challenges in IR benchmarking exacerbated by the scale at which deep learning models operate. Evaluation efforts such as the TREC Deep Learning track and the MS MARCO public leaderboard are intended to encourage research and track our progress, addressing big questions in our field. The goal is not simply to identify which run is "best" but to move the field forward by developing new robust techniques, that work in many different settings, and are adopted in research and practice. This entails a wider conversation in the IR community about what constitutes meaningful progress, how benchmark design can encourage or discourage certain outcomes, and about the validity of our findings. In this talk, I will present a brief overview of what we have learned from our work on MS MARCO and the TREC Deep Learning track--and reflect on the state of the field and the road ahead.
Conformer-Kernel with Query Term Independence @ TREC 2020 Deep Learning TrackBhaskar Mitra
We benchmark Conformer-Kernel models under the strict blind evaluation setting of the TREC 2020 Deep Learning track. In particular, we study the impact of incorporating: (i) Explicit term matching to complement matching based on learned representations (i.e., the “Duet principle”), (ii) query term independence (i.e., the “QTI assumption”) to scale the model to the full retrieval setting, and (iii) the ORCAS click data as an additional document description field. We find evidence which supports that all three aforementioned strategies can lead to improved retrieval quality.
Lecture slides presented at Northeastern University (December, 2020).
Learning to rank (LTR) for information retrieval (IR) involves the application of machine learning models to rank artifacts, such as webpages, in response to user's need, which may be expressed as a query. LTR models typically employ training data, such as human relevance labels and click data, to discriminatively train towards an IR objective. The focus of this lecture will be on the fundamentals of neural networks and their applications to learning to rank.
This report discusses three submissions based on the Duet architecture to the Deep Learning track at TREC 2019. For the document retrieval task, we adapt the Duet model to ingest a "multiple field" view of documents—we refer to the new architecture as Duet with Multiple Fields (DuetMF). A second submission combines the DuetMF model with other neural and traditional relevance estimators in a learning-to-rank framework and achieves improved performance over the DuetMF baseline. For the passage retrieval task, we submit a single run based on an ensemble of eight Duet models.
Benchmarking for Neural Information Retrieval: MS MARCO, TREC, and BeyondBhaskar Mitra
The emergence of deep learning-based methods for information retrieval (IR) poses several challenges and opportunities for benchmarking. Some of these are new, while others have evolved from existing challenges in IR exacerbated by the scale at which deep learning models operate. In this talk, I will present a brief overview of what we have learned from our work on MS MARCO and the TREC Deep Learning track, and reflect on the road ahead.
The document summarizes a workshop on recommender systems held from August 20-23, 2019 at HEC Montréal. It discusses the objectives of the workshop which include providing a quick recap of neural networks and learning to rank, as well as learning to rank with deep neural networks. Bhaskar Mitra will present on neural learning to rank as the principal applied scientist at Microsoft and PhD student at University College London.
Adversarial and reinforcement learning-based approaches to information retrievalBhaskar Mitra
Traditionally, machine learning based approaches to information retrieval have taken the form of supervised learning-to-rank models. Recently, other machine learning approaches—such as adversarial learning and reinforcement learning—have started to find interesting applications in retrieval systems. At Bing, we have been exploring some of these methods in the context of web search. In this talk, I will share couple of our recent work in this area that we presented at SIGIR 2018.
5 Lessons Learned from Designing Neural Models for Information RetrievalBhaskar Mitra
Slides from my keynote talk at the Recherche d'Information SEmantique (RISE) workshop at CORIA-TALN 2018 conference in Rennes, France.
(Abstract)
Neural Information Retrieval (or neural IR) is the application of shallow or deep neural networks to IR tasks. Unlike classical IR models, these machine learning (ML) based approaches are data-hungry, requiring large scale training data before they can be deployed. Traditional learning to rank models employ supervised ML techniques—including neural networks—over hand-crafted IR features. By contrast, more recently proposed neural models learn representations of language from raw text that can bridge the gap between the query and the document vocabulary.
Neural IR is an emerging field and research publications in the area has been increasing in recent years. While the community explores new architectures and training regimes, a new set of challenges, opportunities, and design principles are emerging in the context of these new IR models. In this talk, I will share five lessons learned from my personal research in the area of neural IR. I will present a framework for discussing different unsupervised approaches to learning latent representations of text. I will cover several challenges to learning effective text representations for IR and discuss how latent space models should be combined with observed feature spaces for better retrieval performance. Finally, I will conclude with a few case studies that demonstrates the application of neural approaches to IR that go beyond text matching.
Neural Models for Information RetrievalBhaskar Mitra
In the last few years, neural representation learning approaches have achieved very good performance on many natural language processing (NLP) tasks, such as language modelling and machine translation. This suggests that neural models may also yield significant performance improvements on information retrieval (IR) tasks, such as relevance ranking, addressing the query-document vocabulary mismatch problem by using semantic rather than lexical matching. IR tasks, however, are fundamentally different from NLP tasks leading to new challenges and opportunities for existing neural representation learning approaches for text.
In this talk, I will present my recent work on neural IR models. We begin with a discussion on learning good representations of text for retrieval. I will present visual intuitions about how different embeddings spaces capture different relationships between items, and their usefulness to different types of IR tasks. The second part of this talk is focused on the applications of deep neural architectures to the document ranking task.
The document discusses a neural model called Duet for ranking documents based on their relevance to a query. Duet uses both a local model that operates on exact term matches between queries and documents, and a distributed model that learns embeddings to match queries and documents in the embedding space. The two models are combined using a linear combination and trained jointly on labeled query-document pairs. Experimental results show Duet performs significantly better at document ranking and other IR tasks compared to using the local and distributed models individually. The amount of training data is also important, with larger datasets needed to learn better representations.
Neural Models for Information RetrievalBhaskar Mitra
In the last few years, neural representation learning approaches have achieved very good performance on many natural language processing (NLP) tasks, such as language modelling and machine translation. This suggests that neural models will also yield significant performance improvements on information retrieval (IR) tasks, such as relevance ranking, addressing the query-document vocabulary mismatch problem by using semantic rather than lexical matching. IR tasks, however, are fundamentally different from NLP tasks leading to new challenges and opportunities for existing neural representation learning approaches for text.
We begin this talk with a discussion on text embedding spaces for modelling different types of relationships between items which makes them suitable for different IR tasks. Next, we present how topic-specific representations can be more effective than learning global embeddings. Finally, we conclude with an emphasis on dealing with rare terms and concepts for IR, and how embedding based approaches can be augmented with neural models for lexical matching for better retrieval performance. While our discussions are grounded in IR tasks, the findings and the insights covered during this talk should be generally applicable to other NLP and machine learning tasks.
The document welcomes attendees to Neu-IR'17, the Second SIGIR Workshop on Neural Information Retrieval. It provides details on the organizers and notes that almost 1 in 4 papers at SIGIR 2017 were related to neural IR, showing the growth of interest in the topic. Statistics on the Neu-IR'17 workshop are given, including 178 registrations and a 76% acceptance rate for the 19 accepted papers out of 25 submissions. The day will include oral presentations, posters, panels and discussions to discuss, share and learn about neural information retrieval.
This document presents the Duet model for document ranking. The Duet model uses a combination of local and distributed representations of text to perform both exact and inexact matching of queries to documents. The local model operates on a term interaction matrix to model exact matches, while the distributed model projects text into an embedding space for inexact matching. Results show the Duet model, which combines these approaches, outperforms models using only local or distributed representations. The Duet model benefits from training on large datasets and can effectively handle queries containing rare terms or needing semantic matching.
Neural Text Embeddings for Information Retrieval (WSDM 2017)Bhaskar Mitra
The document describes a tutorial on using neural networks for information retrieval. It discusses an agenda for the tutorial that includes fundamentals of IR, word embeddings, using word embeddings for IR, deep neural networks, and applications of neural networks to IR problems. It provides context on the increasing use of neural methods in IR applications and research.
Query Expansion with Locally-Trained Word Embeddings (ACL 2016)Bhaskar Mitra
This document discusses using locally-trained word embeddings for query expansion to improve search results. It trains word embeddings on either the full corpus (global) or just the topically-relevant documents (local) and finds the local approach works better. It experiments on three datasets and finds local embeddings consistently outperform global embeddings and no expansion for search result ranking based on NDCG.
Query Expansion with Locally-Trained Word Embeddings (Neu-IR 2016)Bhaskar Mitra
This document discusses using locally-trained word embeddings for query expansion. It shows that training word embeddings on documents relevant to a query (local model) provides a better representation than training globally on the entire corpus. In experiments on three datasets, the local model improved average NDCG@10 scores over using global embeddings or no expansion. The local model identifies query and expansion terms more closely related to relevant documents. Future work could improve effectiveness and efficiency of the local model approach.
Discover top-tier mobile app development services, offering innovative solutions for iOS and Android. Enhance your business with custom, user-friendly mobile applications.
The Department of Veteran Affairs (VA) invited Taylor Paschal, Knowledge & Information Management Consultant at Enterprise Knowledge, to speak at a Knowledge Management Lunch and Learn hosted on June 12, 2024. All Office of Administration staff were invited to attend and received professional development credit for participating in the voluntary event.
The objectives of the Lunch and Learn presentation were to:
- Review what KM ‘is’ and ‘isn’t’
- Understand the value of KM and the benefits of engaging
- Define and reflect on your “what’s in it for me?”
- Share actionable ways you can participate in Knowledge - - Capture & Transfer
"$10 thousand per minute of downtime: architecture, queues, streaming and fin...Fwdays
Direct losses from downtime in 1 minute = $5-$10 thousand dollars. Reputation is priceless.
As part of the talk, we will consider the architectural strategies necessary for the development of highly loaded fintech solutions. We will focus on using queues and streaming to efficiently work and manage large amounts of data in real-time and to minimize latency.
We will focus special attention on the architectural patterns used in the design of the fintech system, microservices and event-driven architecture, which ensure scalability, fault tolerance, and consistency of the entire system.
Taking AI to the Next Level in Manufacturing.pdfssuserfac0301
Read Taking AI to the Next Level in Manufacturing to gain insights on AI adoption in the manufacturing industry, such as:
1. How quickly AI is being implemented in manufacturing.
2. Which barriers stand in the way of AI adoption.
3. How data quality and governance form the backbone of AI.
4. Organizational processes and structures that may inhibit effective AI adoption.
6. Ideas and approaches to help build your organization's AI strategy.
High performance Serverless Java on AWS- GoTo Amsterdam 2024Vadym Kazulkin
Java is for many years one of the most popular programming languages, but it used to have hard times in the Serverless community. Java is known for its high cold start times and high memory footprint, comparing to other programming languages like Node.js and Python. In this talk I'll look at the general best practices and techniques we can use to decrease memory consumption, cold start times for Java Serverless development on AWS including GraalVM (Native Image) and AWS own offering SnapStart based on Firecracker microVM snapshot and restore and CRaC (Coordinated Restore at Checkpoint) runtime hooks. I'll also provide a lot of benchmarking on Lambda functions trying out various deployment package sizes, Lambda memory settings, Java compilation options and HTTP (a)synchronous clients and measure their impact on cold and warm start times.
Conversational agents, or chatbots, are increasingly used to access all sorts of services using natural language. While open-domain chatbots - like ChatGPT - can converse on any topic, task-oriented chatbots - the focus of this paper - are designed for specific tasks, like booking a flight, obtaining customer support, or setting an appointment. Like any other software, task-oriented chatbots need to be properly tested, usually by defining and executing test scenarios (i.e., sequences of user-chatbot interactions). However, there is currently a lack of methods to quantify the completeness and strength of such test scenarios, which can lead to low-quality tests, and hence to buggy chatbots.
To fill this gap, we propose adapting mutation testing (MuT) for task-oriented chatbots. To this end, we introduce a set of mutation operators that emulate faults in chatbot designs, an architecture that enables MuT on chatbots built using heterogeneous technologies, and a practical realisation as an Eclipse plugin. Moreover, we evaluate the applicability, effectiveness and efficiency of our approach on open-source chatbots, with promising results.
[OReilly Superstream] Occupy the Space: A grassroots guide to engineering (an...Jason Yip
The typical problem in product engineering is not bad strategy, so much as “no strategy”. This leads to confusion, lack of motivation, and incoherent action. The next time you look for a strategy and find an empty space, instead of waiting for it to be filled, I will show you how to fill it in yourself. If you’re wrong, it forces a correction. If you’re right, it helps create focus. I’ll share how I’ve approached this in the past, both what works and lessons for what didn’t work so well.
Connector Corner: Seamlessly power UiPath Apps, GenAI with prebuilt connectorsDianaGray10
Join us to learn how UiPath Apps can directly and easily interact with prebuilt connectors via Integration Service--including Salesforce, ServiceNow, Open GenAI, and more.
The best part is you can achieve this without building a custom workflow! Say goodbye to the hassle of using separate automations to call APIs. By seamlessly integrating within App Studio, you can now easily streamline your workflow, while gaining direct access to our Connector Catalog of popular applications.
We’ll discuss and demo the benefits of UiPath Apps and connectors including:
Creating a compelling user experience for any software, without the limitations of APIs.
Accelerating the app creation process, saving time and effort
Enjoying high-performance CRUD (create, read, update, delete) operations, for
seamless data management.
Speakers:
Russell Alfeche, Technology Leader, RPA at qBotic and UiPath MVP
Charlie Greenberg, host
Session 1 - Intro to Robotic Process Automation.pdfUiPathCommunity
👉 Check out our full 'Africa Series - Automation Student Developers (EN)' page to register for the full program:
https://bit.ly/Automation_Student_Kickstart
In this session, we shall introduce you to the world of automation, the UiPath Platform, and guide you on how to install and setup UiPath Studio on your Windows PC.
📕 Detailed agenda:
What is RPA? Benefits of RPA?
RPA Applications
The UiPath End-to-End Automation Platform
UiPath Studio CE Installation and Setup
💻 Extra training through UiPath Academy:
Introduction to Automation
UiPath Business Automation Platform
Explore automation development with UiPath Studio
👉 Register here for our upcoming Session 2 on June 20: Introduction to UiPath Studio Fundamentals: https://community.uipath.com/events/details/uipath-lagos-presents-session-2-introduction-to-uipath-studio-fundamentals/
zkStudyClub - LatticeFold: A Lattice-based Folding Scheme and its Application...Alex Pruden
Folding is a recent technique for building efficient recursive SNARKs. Several elegant folding protocols have been proposed, such as Nova, Supernova, Hypernova, Protostar, and others. However, all of them rely on an additively homomorphic commitment scheme based on discrete log, and are therefore not post-quantum secure. In this work we present LatticeFold, the first lattice-based folding protocol based on the Module SIS problem. This folding protocol naturally leads to an efficient recursive lattice-based SNARK and an efficient PCD scheme. LatticeFold supports folding low-degree relations, such as R1CS, as well as high-degree relations, such as CCS. The key challenge is to construct a secure folding protocol that works with the Ajtai commitment scheme. The difficulty, is ensuring that extracted witnesses are low norm through many rounds of folding. We present a novel technique using the sumcheck protocol to ensure that extracted witnesses are always low norm no matter how many rounds of folding are used. Our evaluation of the final proof system suggests that it is as performant as Hypernova, while providing post-quantum security.
Paper Link: https://eprint.iacr.org/2024/257
How information systems are built or acquired puts information, which is what they should be about, in a secondary place. Our language adapted accordingly, and we no longer talk about information systems but applications. Applications evolved in a way to break data into diverse fragments, tightly coupled with applications and expensive to integrate. The result is technical debt, which is re-paid by taking even bigger "loans", resulting in an ever-increasing technical debt. Software engineering and procurement practices work in sync with market forces to maintain this trend. This talk demonstrates how natural this situation is. The question is: can something be done to reverse the trend?
inQuba Webinar Mastering Customer Journey Management with Dr Graham HillLizaNolte
HERE IS YOUR WEBINAR CONTENT! 'Mastering Customer Journey Management with Dr. Graham Hill'. We hope you find the webinar recording both insightful and enjoyable.
In this webinar, we explored essential aspects of Customer Journey Management and personalization. Here’s a summary of the key insights and topics discussed:
Key Takeaways:
Understanding the Customer Journey: Dr. Hill emphasized the importance of mapping and understanding the complete customer journey to identify touchpoints and opportunities for improvement.
Personalization Strategies: We discussed how to leverage data and insights to create personalized experiences that resonate with customers.
Technology Integration: Insights were shared on how inQuba’s advanced technology can streamline customer interactions and drive operational efficiency.
Main news related to the CCS TSI 2023 (2023/1695)Jakub Marek
An English 🇬🇧 translation of a presentation to the speech I gave about the main changes brought by CCS TSI 2023 at the biggest Czech conference on Communications and signalling systems on Railways, which was held in Clarion Hotel Olomouc from 7th to 9th November 2023 (konferenceszt.cz). Attended by around 500 participants and 200 on-line followers.
The original Czech 🇨🇿 version of the presentation can be found here: https://www.slideshare.net/slideshow/hlavni-novinky-souvisejici-s-ccs-tsi-2023-2023-1695/269688092 .
The videorecording (in Czech) from the presentation is available here: https://youtu.be/WzjJWm4IyPk?si=SImb06tuXGb30BEH .
Skybuffer SAM4U tool for SAP license adoptionTatiana Kojar
Manage and optimize your license adoption and consumption with SAM4U, an SAP free customer software asset management tool.
SAM4U, an SAP complimentary software asset management tool for customers, delivers a detailed and well-structured overview of license inventory and usage with a user-friendly interface. We offer a hosted, cost-effective, and performance-optimized SAM4U setup in the Skybuffer Cloud environment. You retain ownership of the system and data, while we manage the ABAP 7.58 infrastructure, ensuring fixed Total Cost of Ownership (TCO) and exceptional services through the SAP Fiori interface.
Must Know Postgres Extension for DBA and Developer during MigrationMydbops
Mydbops Opensource Database Meetup 16
Topic: Must-Know PostgreSQL Extensions for Developers and DBAs During Migration
Speaker: Deepak Mahto, Founder of DataCloudGaze Consulting
Date & Time: 8th June | 10 AM - 1 PM IST
Venue: Bangalore International Centre, Bangalore
Abstract: Discover how PostgreSQL extensions can be your secret weapon! This talk explores how key extensions enhance database capabilities and streamline the migration process for users moving from other relational databases like Oracle.
Key Takeaways:
* Learn about crucial extensions like oracle_fdw, pgtt, and pg_audit that ease migration complexities.
* Gain valuable strategies for implementing these extensions in PostgreSQL to achieve license freedom.
* Discover how these key extensions can empower both developers and DBAs during the migration process.
* Don't miss this chance to gain practical knowledge from an industry expert and stay updated on the latest open-source database trends.
Mydbops Managed Services specializes in taking the pain out of database management while optimizing performance. Since 2015, we have been providing top-notch support and assistance for the top three open-source databases: MySQL, MongoDB, and PostgreSQL.
Our team offers a wide range of services, including assistance, support, consulting, 24/7 operations, and expertise in all relevant technologies. We help organizations improve their database's performance, scalability, efficiency, and availability.
Contact us: info@mydbops.com
Visit: https://www.mydbops.com/
Follow us on LinkedIn: https://in.linkedin.com/company/mydbops
For more details and updates, please follow up the below links.
Meetup Page : https://www.meetup.com/mydbops-databa...
Twitter: https://twitter.com/mydbopsofficial
Blogs: https://www.mydbops.com/blog/
Facebook(Meta): https://www.facebook.com/mydbops/
Must Know Postgres Extension for DBA and Developer during Migration
Multisided Exposure Fairness for Search and Recommendation
1. Joint Multisided Exposure Fairness
for Search and Recommendation
Bhaskar Mitra
Microsoft Research, Canada
bmitra@microsoft.com
Pre-print: https://arxiv.org/pdf/2205.00048.pdf
(Paper accepted @ SIGIR’22)
Joint work with Haolun Wu, Chen Ma,
Fernando Diaz, and Xue Liu
2. Digital information
access and exposure
Traditional IR is concerned with ranking
of items according to relevance
These information access systems
deployed at web-scale mediate what
information gets exposure
The exposure-framing of IR raises several
fairness concerns, new opportunities for
ranking optimization, and can be
relevant to other FATE considerations
(e.g., privacy and transparency)
3. Sweeney. Discrimination in online ad delivery. Commun. ACM. (2013)
Crawford. The Trouble with Bias. NeurIPS. (2017)
Singh and Joachims. Fairness of Exposure in Rankings. In KDD, ACM. (2018)
Harms of disparate exposure
Several past studies have pointed out representational
and allocative harms from disparate exposure
Concerns of fairness in the context of IR/ML systems are
inherently interdisciplinary and sociotechnical; and these
concerns span beyond just questions of system design
The role of IR/ML in this process is to deconstruct their
own measures and models in ways that allows a broad
range of researchers and stakeholders to critically
analyze and shape these technologies
In traditional IR, we have made progress in
modeling, measuring, and optimizing for
individual user satisfaction; a key challenge ahead
is to model, measure, and optimize IR systems with
respect to impact on populations of users and
consider disparate impact across subpopulations
4. Exposure fairness is a multisided problem
It is important to ask not just whether specific content receives
exposure, but who it is exposed to and in what context
Haolun, Mitra, Ma, and Liu. Joint Multisided Exposure Fairness for Recommendation. Under review for SIGIR, ACM. (2022)
5. Exposure fairness is a multisided problem
Take the example of a job recommendation system
Group-of-users-to-group-of-items fairness (GG-F)
Are groups of items under/over-exposed to groups
of users?
E.g., men being disproportionately recommended
high-paying jobs and women low-paying jobs.
Individual-user-to-Individual-item fairness (II-F)
Are Individual items under/over-exposed to
Individual users?
Individual-user-to-group-of-items fairness (IG-F)
Are groups of items under/over-exposed to
individual users?
E.g., a specific user being disproportionately
recommended low-paying jobs.
Group-of-users-to-Individual-item fairness (GI-F)
Are Individual items under/over-exposed to groups
of users?
E.g., a specific job being disproportionately
recommended to men and not to women and
non-binary people.
All-users-to-Individual-item fairness (AI-F)
Are Individual items under/over-exposed to all users
overall?
E.g., a specific job being disproportionately under-
exposed to all users.
All-users-to-group-of-items fairness (AG-F)
Are groups of items under/over-exposed to all users
overall?
E.g., jobs at Black-owned businesses being
disproportionately under-exposed to all users.
6. Exposure fairness is a multisided problem
Take the example of a job recommendation system
Group-of-users-to-group-of-items fairness (GG-F)
Are groups of items under/over-exposed to groups
of users?
E.g., men being disproportionately recommended
high-paying jobs and women low-paying jobs.
Individual-user-to-Individual-item fairness (II-F)
Are Individual items under/over-exposed to
Individual users?
Individual-user-to-group-of-items fairness (IG-F)
Are groups of items under/over-exposed to
individual users?
E.g., a specific user being disproportionately
recommended low-paying jobs.
Group-of-users-to-Individual-item fairness (GI-F)
Are Individual items under/over-exposed to groups
of users?
E.g., a specific job being disproportionately
recommended to men and not to women and
non-binary people.
All-users-to-Individual-item fairness (AI-F)
Are Individual items under/over-exposed to all users
overall?
E.g., a specific job being disproportionately under-
exposed to all users.
All-users-to-group-of-items fairness (AG-F)
Are groups of items under/over-exposed to all users
overall?
E.g., jobs at Black-owned businesses being
disproportionately under-exposed to all users.
7. Exposure fairness is a multisided problem
Take the example of a job recommendation system
Group-of-users-to-group-of-items fairness (GG-F)
Are groups of items under/over-exposed to groups
of users?
E.g., men being disproportionately recommended
high-paying jobs and women low-paying jobs.
Individual-user-to-Individual-item fairness (II-F)
Are Individual items under/over-exposed to
Individual users?
Individual-user-to-group-of-items fairness (IG-F)
Are groups of items under/over-exposed to
individual users?
E.g., a specific user being disproportionately
recommended low-paying jobs.
Group-of-users-to-Individual-item fairness (GI-F)
Are Individual items under/over-exposed to groups
of users?
E.g., a specific job being disproportionately
recommended to men and not to women and
non-binary people.
All-users-to-Individual-item fairness (AI-F)
Are Individual items under/over-exposed to all users
overall?
E.g., a specific job being disproportionately under-
exposed to all users.
All-users-to-group-of-items fairness (AG-F)
Are groups of items under/over-exposed to all users
overall?
E.g., jobs at Black-owned businesses being
disproportionately under-exposed to all users.
8. Exposure fairness is a multisided problem
Take the example of a job recommendation system
Group-of-users-to-group-of-items fairness (GG-F)
Are groups of items under/over-exposed to groups
of users?
E.g., men being disproportionately recommended
high-paying jobs and women low-paying jobs.
Individual-user-to-Individual-item fairness (II-F)
Are Individual items under/over-exposed to
Individual users?
Individual-user-to-group-of-items fairness (IG-F)
Are groups of items under/over-exposed to
individual users?
E.g., a specific user being disproportionately
recommended low-paying jobs.
Group-of-users-to-Individual-item fairness (GI-F)
Are Individual items under/over-exposed to groups
of users?
E.g., a specific job being disproportionately
recommended to men and not to women and
non-binary people.
All-users-to-Individual-item fairness (AI-F)
Are Individual items under/over-exposed to all users
overall?
E.g., a specific job being disproportionately under-
exposed to all users.
All-users-to-group-of-items fairness (AG-F)
Are groups of items under/over-exposed to all users
overall?
E.g., jobs at Black-owned businesses being
disproportionately under-exposed to all users.
9. Exposure fairness is a multisided problem
Take the example of a job recommendation system
Group-of-users-to-group-of-items fairness (GG-F)
Are groups of items under/over-exposed to groups
of users?
E.g., men being disproportionately recommended
high-paying jobs and women low-paying jobs.
Individual-user-to-Individual-item fairness (II-F)
Are Individual items under/over-exposed to
Individual users?
Individual-user-to-group-of-items fairness (IG-F)
Are groups of items under/over-exposed to
individual users?
E.g., a specific user being disproportionately
recommended low-paying jobs.
Group-of-users-to-Individual-item fairness (GI-F)
Are Individual items under/over-exposed to groups
of users?
E.g., a specific job being disproportionately
recommended to men and not to women and
non-binary people.
All-users-to-Individual-item fairness (AI-F)
Are Individual items under/over-exposed to all users
overall?
E.g., a specific job being disproportionately under-
exposed to all users.
All-users-to-group-of-items fairness (AG-F)
Are groups of items under/over-exposed to all users
overall?
E.g., jobs at Black-owned businesses being
disproportionately under-exposed to all users.
10. Exposure fairness is a multisided problem
Take the example of a job recommendation system
Group-of-users-to-group-of-items fairness (GG-F)
Are groups of items under/over-exposed to groups
of users?
E.g., men being disproportionately recommended
high-paying jobs and women low-paying jobs.
Individual-user-to-Individual-item fairness (II-F)
Are Individual items under/over-exposed to
Individual users?
Individual-user-to-group-of-items fairness (IG-F)
Are groups of items under/over-exposed to
individual users?
E.g., a specific user being disproportionately
recommended low-paying jobs.
Group-of-users-to-Individual-item fairness (GI-F)
Are Individual items under/over-exposed to groups
of users?
E.g., a specific job being disproportionately
recommended to men and not to women and
non-binary people.
All-users-to-Individual-item fairness (AI-F)
Are Individual items under/over-exposed to all users
overall?
E.g., a specific job being disproportionately under-
exposed to all users.
All-users-to-group-of-items fairness (AG-F)
Are groups of items under/over-exposed to all users
overall?
E.g., jobs at Black-owned businesses being
disproportionately under-exposed to all users.
11. Exposure fairness is a multisided problem
Take the example of a job recommendation system
Group-of-users-to-group-of-items fairness (GG-F)
Are groups of items under/over-exposed to groups
of users?
E.g., men being disproportionately recommended
high-paying jobs and women low-paying jobs.
Individual-user-to-Individual-item fairness (II-F)
Are Individual items under/over-exposed to
Individual users?
Individual-user-to-group-of-items fairness (IG-F)
Are groups of items under/over-exposed to
individual users?
E.g., a specific user being disproportionately
recommended low-paying jobs.
Group-of-users-to-Individual-item fairness (GI-F)
Are Individual items under/over-exposed to groups
of users?
E.g., a specific job being disproportionately
recommended to men and not to women and
non-binary people.
All-users-to-Individual-item fairness (AI-F)
Are Individual items under/over-exposed to all users
overall?
E.g., a specific job being disproportionately under-
exposed to all users.
All-users-to-group-of-items fairness (AG-F)
Are groups of items under/over-exposed to all users
overall?
E.g., jobs at Black-owned businesses being
disproportionately under-exposed to all users.
12. User browsing models and exposure
User browsing models are simplified models of how users inspect
and interact with retrieved results
It estimates the probability that the user inspects a particular item
in a ranked list of items—i.e., the item is exposed to the user
In IR, user models have been implicitly and explicitly employed in
metric definitions and for estimating relevance from historical
logs of user behavior data
For example, let’s consider the RBP user model…
NDCG
RBP
Probability of exposure at different ranks according
to NDCG and RBP user browsing models
exposure event
an item
a ranked list of items
rank of the item in the ranked list
patience factor
13. Stochastic ranking and expected exposure
In recommendation, Diaz et al. (2020) define a stochastic ranking policy 𝜋𝑢, conditioned on user
𝑢 ∈ U, as a probability distribution over all permutations of items in the collection
The expected exposure of an item 𝑑 for user 𝑢 can then be computed as follows:
Here, 𝑝(𝜖|𝑑,𝜎) can be computed using a user browsing model like RBP as discussed previously
Note: The above formulation can also be applied to search by replacing user with query
Diaz, Mitra, Ekstrand, Biega, and Carterette. Evaluating stochastic rankings with expected exposure. In CIKM, ACM. (2020)
14. System, target, and random exposure
System exposure. The user-item expected exposure distribution corresponding to a stochastic
ranking policy 𝜋. Correspondingly, we can define a |U|×|D| matrix E, such that E𝑖𝑗 = 𝑝(𝜖|D𝑗 ,𝜋U𝑖
).
Target exposure. The user-item expected exposure distribution corresponding to an ideal
stochastic ranking policy 𝜋*, as defined by some desirable principle (e.g., the equal expected
exposure principle). We denote the corresponding expected exposure matrix as E*.
Random exposure. The user-item expected exposure distribution corresponding to a stochastic
ranking policy 𝜋~ that samples rankings from a uniform distribution over all item permutations.
We denote the corresponding expected exposure matrix as E~.
The deviation of E from E* gives us a quantitative measure of the suboptimality of the retrieval
system under consideration.
15. Joint multisided exposure (JME) fairness metrics
Haolun, Mitra, Ma, and Liu. Joint Multisided Exposure Fairness for Recommendation. Under review for SIGIR, ACM. (2022)
16. Toy example
Let, there be 4 candidates (𝑢𝑎1
, 𝑢𝑎2
, 𝑢𝑏1
, 𝑢𝑏2
) and
4 jobs (𝑑𝑥1
, 𝑑𝑥2
, 𝑑𝑦1
, 𝑑𝑦2
)
All 4 jobs are relevant to each of the 4
candidates
The candidates belong to 2 groups 𝑎 (𝑢𝑎1
, 𝑢𝑎2
)
and 𝑏 (𝑢𝑏1
, 𝑢𝑏2
)—e.g., based on gender—and
similarly the jobs belong to 2 groups 𝑥 (𝑑𝑥1
, 𝑑𝑥2
)
and 𝑦 (𝑑𝑦1
, 𝑑𝑦2
)—say based on whether they
pay high or low salaries
Let’s assume that the recommender system
displays only one result at a time and our simple
user model assumes that the user always
inspects the displayed result—i.e., the
probability of exposure is 1 for the displayed
item and 0 for all other items for a given
impression
In this setting, an ideal recommender should
expose each of the four jobs to each candidate
with a probability of 0.25
17. Toy example
Let, there be 4 candidates (𝑢𝑎1
, 𝑢𝑎2
, 𝑢𝑏1
, 𝑢𝑏2
) and
4 jobs (𝑑𝑥1
, 𝑑𝑥2
, 𝑑𝑦1
, 𝑑𝑦2
)
All 4 jobs are relevant to each of the 4
candidates
The candidates belong to 2 groups 𝑎 (𝑢𝑎1
, 𝑢𝑎2
)
and 𝑏 (𝑢𝑏1
, 𝑢𝑏2
)—e.g., based on gender—and
similarly the jobs belong to 2 groups 𝑥 (𝑑𝑥1
, 𝑑𝑥2
)
and 𝑦 (𝑑𝑦1
, 𝑑𝑦2
)—say based on whether they
pay high or low salaries
Let’s assume that the recommender system
displays only one result at a time and our simple
user model assumes that the user always
inspects the displayed result—i.e., the
probability of exposure is 1 for the displayed
item and 0 for all other items for a given
impression
In this setting, an ideal recommender should
expose each of the four jobs to each candidate
with a probability of 0.25
All of them are equally II-Unfair
18. Let, there be 4 candidates (𝑢𝑎1
, 𝑢𝑎2
, 𝑢𝑏1
, 𝑢𝑏2
) and
4 jobs (𝑑𝑥1
, 𝑑𝑥2
, 𝑑𝑦1
, 𝑑𝑦2
)
All 4 jobs are relevant to each of the 4
candidates
The candidates belong to 2 groups 𝑎 (𝑢𝑎1
, 𝑢𝑎2
)
and 𝑏 (𝑢𝑏1
, 𝑢𝑏2
)—e.g., based on gender—and
similarly the jobs belong to 2 groups 𝑥 (𝑑𝑥1
, 𝑑𝑥2
)
and 𝑦 (𝑑𝑦1
, 𝑑𝑦2
)—say based on whether they
pay high or low salaries
Let’s assume that the recommender system
displays only one result at a time and our simple
user model assumes that the user always
inspects the displayed result—i.e., the
probability of exposure is 1 for the displayed
item and 0 for all other items for a given
impression
In this setting, an ideal recommender should
expose each of the four jobs to each candidate
with a probability of 0.25
Only (b), (e), and (f) are IG-Unfair
Toy example
19. Toy example
Let, there be 4 candidates (𝑢𝑎1
, 𝑢𝑎2
, 𝑢𝑏1
, 𝑢𝑏2
) and
4 jobs (𝑑𝑥1
, 𝑑𝑥2
, 𝑑𝑦1
, 𝑑𝑦2
)
All 4 jobs are relevant to each of the 4
candidates
The candidates belong to 2 groups 𝑎 (𝑢𝑎1
, 𝑢𝑎2
)
and 𝑏 (𝑢𝑏1
, 𝑢𝑏2
)—e.g., based on gender—and
similarly the jobs belong to 2 groups 𝑥 (𝑑𝑥1
, 𝑑𝑥2
)
and 𝑦 (𝑑𝑦1
, 𝑑𝑦2
)—say based on whether they
pay high or low salaries
Let’s assume that the recommender system
displays only one result at a time and our simple
user model assumes that the user always
inspects the displayed result—i.e., the
probability of exposure is 1 for the displayed
item and 0 for all other items for a given
impression
In this setting, an ideal recommender should
expose each of the four jobs to each candidate
with a probability of 0.25
Only (c), (d), (e), and (f) are GI-Unfair
20. Toy example
Let, there be 4 candidates (𝑢𝑎1
, 𝑢𝑎2
, 𝑢𝑏1
, 𝑢𝑏2
) and
4 jobs (𝑑𝑥1
, 𝑑𝑥2
, 𝑑𝑦1
, 𝑑𝑦2
)
All 4 jobs are relevant to each of the 4
candidates
The candidates belong to 2 groups 𝑎 (𝑢𝑎1
, 𝑢𝑎2
)
and 𝑏 (𝑢𝑏1
, 𝑢𝑏2
)—e.g., based on gender—and
similarly the jobs belong to 2 groups 𝑥 (𝑑𝑥1
, 𝑑𝑥2
)
and 𝑦 (𝑑𝑦1
, 𝑑𝑦2
)—say based on whether they
pay high or low salaries
Let’s assume that the recommender system
displays only one result at a time and our simple
user model assumes that the user always
inspects the displayed result—i.e., the
probability of exposure is 1 for the displayed
item and 0 for all other items for a given
impression
In this setting, an ideal recommender should
expose each of the four jobs to each candidate
with a probability of 0.25
Only (e) and (f) are GG-Unfair
21. Let, there be 4 candidates (𝑢𝑎1
, 𝑢𝑎2
, 𝑢𝑏1
, 𝑢𝑏2
) and
4 jobs (𝑑𝑥1
, 𝑑𝑥2
, 𝑑𝑦1
, 𝑑𝑦2
)
All 4 jobs are relevant to each of the 4
candidates
The candidates belong to 2 groups 𝑎 (𝑢𝑎1
, 𝑢𝑎2
)
and 𝑏 (𝑢𝑏1
, 𝑢𝑏2
)—e.g., based on gender—and
similarly the jobs belong to 2 groups 𝑥 (𝑑𝑥1
, 𝑑𝑥2
)
and 𝑦 (𝑑𝑦1
, 𝑑𝑦2
)—say based on whether they
pay high or low salaries
Let’s assume that the recommender system
displays only one result at a time and our simple
user model assumes that the user always
inspects the displayed result—i.e., the
probability of exposure is 1 for the displayed
item and 0 for all other items for a given
impression
In this setting, an ideal recommender should
expose each of the four jobs to each candidate
with a probability of 0.25
Only (d) and (f) are AI-Unfair
Toy example
22. Toy example
Let, there be 4 candidates (𝑢𝑎1
, 𝑢𝑎2
, 𝑢𝑏1
, 𝑢𝑏2
) and
4 jobs (𝑑𝑥1
, 𝑑𝑥2
, 𝑑𝑦1
, 𝑑𝑦2
)
All 4 jobs are relevant to each of the 4
candidates
The candidates belong to 2 groups 𝑎 (𝑢𝑎1
, 𝑢𝑎2
)
and 𝑏 (𝑢𝑏1
, 𝑢𝑏2
)—e.g., based on gender—and
similarly the jobs belong to 2 groups 𝑥 (𝑑𝑥1
, 𝑑𝑥2
)
and 𝑦 (𝑑𝑦1
, 𝑑𝑦2
)—say based on whether they
pay high or low salaries
Let’s assume that the recommender system
displays only one result at a time and our simple
user model assumes that the user always
inspects the displayed result—i.e., the
probability of exposure is 1 for the displayed
item and 0 for all other items for a given
impression
In this setting, an ideal recommender should
expose each of the four jobs to each candidate
with a probability of 0.25
Only (f) is AG-Unfair
23. Relationship between
different JME metrics
Based on the metric definitions, we can
show that a system that is II-Fair (i.e., II-
F=0) will also be fair along the other
five JME-fairness dimensions
Similarly, IG-Fair and GI-Fair
independently implies GG-Fair, and
GG-Fair and AI-Fair implies AG-Fair
Finally, all the other metrics can be
viewed as specific instances of GG-F,
with different (extreme) definitions of
groups on user and item side
II-F=0
IG-F=0 GI-F=0
GG-F=0 AI-F=0
AG-F=0
24. Disparity and
relevance
Each of our proposed JME-fairness metrics can be
decomposed into a disparity and a relevance component,
such that increasing randomness in the model would
decrease disparity (good!) but also decrease relevance (bad!)
25. Disparity and
relevance
Each of our proposed JME-fairness metrics can be
decomposed into a disparity and a relevance component,
such that increasing randomness in the model would
decrease disparity (good!) but also decrease relevance (bad!)
26. Disparity and
relevance
Each of our proposed JME-fairness metrics can be
decomposed into a disparity and a relevance component,
such that increasing randomness in the model would
decrease disparity (good!) but also decrease relevance (bad!)
Different models have
different disparity-relevance
trade-off for each of the
different JME-fairness metrics
27. How correlated are different
JME-fairness dimensions?
Recall that all six JME-Fairness metrics can be seen as
specific instances of GG-F
For this analysis using MovieLens we had 2 groups
by gender* and 7 groups by age on the user side
and 18 genres on the item side
When we have small number of large groups
“Individual” and “Group” analysis will diverge, and
vice versa
* The gender attribute is available in the MovieLens dataset as a binary annotation. We recognize that
this does not reflect the full spectrum of gender identities, and this is a short-coming of our work.
28. New metrics, new optimization opportunity!
How can we optimize ranking models for target exposure?
Diaz, Mitra, Ekstrand, Biega, and Carterette. Evaluating stochastic rankings with expected exposure. In CIKM, ACM. (2020)
29. Stochastic ranking
A stochastic ranking model samples a ranking from a probability distribution over all possible permutations
of items in the collection—i.e., for the same intent it returns a slightly different ranking on each impression
Given a static ranking policy, we can generate a stochastic equivalent using Plackett-Luce sampling—for
example, given items 𝑑1, 𝑑2, 𝑑3, 𝑑4 the probability of sampling a particular ranking 𝑑2, 𝑑1, 𝑑4, 𝑑3 is:
𝜋: a ranking, 𝜙: a transformation, e.g., exponential over score 𝑠𝑖 for document 𝑑𝑖
Equivalent to sequentially sampling documents without replacement with probability 𝜙 𝑠𝑖
restaurants in montreal restaurants in montreal
restaurants in montreal
restaurants in montreal
Luce. Individual Choice Behavior. (1959)
Plackett. The Analysis of Permutations. Journal of the Royal Statistical Society: Series C (Applied Statistics). (1975)
30. Gradient-based optimization for target exposure
Approach
1. Use the target model to score the items
2. Compute PL sampling probability as a
function of the item scores
3. Sample multiple rankings
4. Compute expected system exposure
across sampled rankings
5. Compute the loss as a difference between
system and target exposure
6. Backpropagate!
Challenges and solutions
The key challenge is the proposed approach is
that both the sampling and the ranking steps
are non-differentiable!
For sampling, we can use Gumbel sampling
as a differentiable approximation
For ranking, we can employ SmoothRank /
ApproxRank as differentiable approximations
of the ranking step
Wu, Chang, Zheng, and Zha. Smoothing DCG for learning to rank: A novel approach using smoothed hinge functions. In Proc. CIKM, ACM. (2009)
Qin, Liu, and Li. A general approximation framework for direct optimization of information retrieval measures. Information retrieval. (2010)
Bruch, Han, Bendersky, and Najork. A stochastic treatment of learning to rank scoring functions. In Proc. WSDM, ACM. (2020)
,
31. Gradient-based optimization for target exposure
add independently
sampled Gumbel noise
neural scoring
function
compute smooth
rank value
compute exposure
using user model
compute loss with
target exposure
compute average
exposure
items target
exposure
Diaz, Mitra, Ekstrand, Biega, and Carterette. Evaluating stochastic rankings with expected exposure. In CIKM, ACM. (2020)
32. Trading-off different JME-fairness metrics
We can simultaneously optimize for multiple exposure metrics by
combining them linearly
For example,
Preliminary experiments indicate that we can significantly
minimize GG-F with minimal degradation to II-F and relevance
33. Discussion
True vs. observed relevance labels. The computation of target exposure itself raises fairness questions. E.g., the equal expected
exposure principle assumes we have access to true relevance labels, but in fact the observed labels reflect huge historical social
biases. E.g., In the job recommendation scenario, it may be more appropriate to define GG-F target exposure for high and low
paying jobs to be uniform across user groups, irrespective of historical disparities reflected in the data.
Choice of group attributes. The choice of group attributes necessitates reflecting on historical and socioeconomic contexts. We
note that our formulation can also be extended to handling multiple group attributes on each side. However, that raises questions
of intersectional fairness that we haven’t yet studied in our work.
Beyond two-sided exposure fairness. While we have primarily focused on two-sided exposure fairness so far, we envision that
extending that to additional stakeholder may also be important. E.g., in product search exposure fairness may concern with being
fair to consumers, manufacturers, and retailers.
Incorporating model uncertainty. The stochastic ranking policies we have considered so far involves randomizing a static policy
with model-independent sampling of noise. In contrast, the stochasticity could also be informed by the model’s own uncertainty
in its prediction. This is an area for potential future work.