photo

 Shib Sankar Dasgupta

  PhD Student
  University of Massachusetts Amherst
  ssdasgupta@cs.umass.edu
  Google Scholar
  Github , GitLab
  LinkedIn

Bio

I am currently a 5th-year Ph.D. student in the College of Information and Computer Sciences at UMass Amherst, working with Prof. Andrew McCallum at IESL.
My research is focused on developing set-based representation learning techniques, with applications in natural language processing and information retrieval. I earned my Master’s degree from the Indian Institute of Science (IISc) in 2018, where I was part of the Machine and Language Learning (MALL) Lab, working with Prof. Partha Talukdar. In addition to my academic research, I have gained valuable industry experience through work/internships with leading researchers, including Dr. Tijmen Tieleman at Minds.ai, Dr. Steffen Rendle at Google Research, Dr. Achille Fokoue at IBM Research, and Dr. Tobias Schnabel at Microsoft Research.

Research Areas

Representation Learning: My PhD thesis focuses on set-based representation learning using Box Embeddings – trainable Venn diagrams that model logical reasoning and set-based similarity. Box embeddings provide a structured alternative to vector-based similarity measures like dot product by encoding inclusion, intersection, and exclusion relationships. I improved the training stability and local identifiability of box embeddings (Improving Local Identifiability), enabling more accurate modeling of complex set-theoretic relationships. In Word2Box, I showed that box embeddings outperform traditional word embeddings in representing hierarchical and compositional semantics. My work on Box-to-Box Transformations demonstrated their effectiveness for joint hierarchical modeling in knowledge graph completion and reasoning.

Information Retrieval: Box embeddings improve real-world IR tasks like faceted search, recommendation systems, and web browsing by better handling compositional queries. In Answering Compositional Queries with Set-Theoretic Embeddings, I proposed a framework for handling queries like "Jazz but not Smooth Jazz" or "Animation ∩ Monsters," where traditional vector models often fail. My work on Personalized Recommendation with Set-Theoretic Constraints explored attribute-aware recommendations, enhancing performance when user preferences follow set-based patterns. I also integrated box embeddings with transformer-based dual encoders, improving search and personalized recommendation performance (results coming soon). At Microsoft, I developed a hybrid approach combining traditional collaborative filtering with LLM-based similarity estimation to enhance recommendation accuracy.

Knowledge Base Representation: My research explores improving knowledge base representation by developing methods that capture both structural and temporal dependencies in large-scale knowledge graphs. In HyTE, I proposed a hyperplane-based embedding method to model the temporal evolution of facts in knowledge graphs. HyTE introduced a novel temporal axis in embedding space, allowing for more accurate reasoning over time-dependent relationships. My work on Neural Dater applied graph convolutional networks (GCNs) to predict the dates of historical documents, showing that structured representations improve temporal inference. Building on my Ph.D. research, I have extended box embeddings to probabilistic knowledge bases, where uncertainty and hierarchical relationships are key challenges. Box embeddings provide a natural geometric representation for encoding uncertainty and capturing the complex structure of multi-relational data. In Probabilistic Box Embeddings for Knowledge Graph Reasoning, I showed that box-based models outperform traditional embeddings in capturing hierarchical and compositional properties of knowledge bases, improving both link prediction and query answering performance. My work demonstrates that box embeddings can effectively handle partial and uncertain information, making them suitable for real-world knowledge graphs.

Education

P.h.D. in Computer Science
University of Massachusetts Amherst.
Started PhD in Fall 2019.
Thesis on Set Theoretic Representation Learning. Supervised by Prof. Andrew McCallum. Published and presented in multiple top-tier ML/NLP conferences.
Master of Technology in System Science and Automation
Indian Institute of Science, Bangalore. M.Tech. 2018.
Thesis on Temporal Information Processing. Published fundamental research work on temporal representation of knowledge bases.
Bachelor of Engineering in Electrical Engineering
Jadavpur University B.E. 2015.

Experience

  1. Microsoft Research Summer 2024
  2. IBM Research Summer 2023
  3. Google Research Summer 2022, Fall 2022
  4. IBM Research Spring 2022
  5. Adobe Research Summer 2021
  6. Minds.ai 2018 - 2019

Publications

Please visit my Google Scholar page for an updated list of publications.

  1. Shib Dasgupta, Michael Boratko, Andrew McCallum. A Geometric Approach to Personalized Recommendation with Set-Theoretic Constraints Using Box Embeddings.
    In-Submission: ICML 2025Recommendation Systems
  2. B. Rozonoyer, M. Boratko, D. Patel, W. Zhao, Shib Dasgupta, H. Le, A. McCallum. Learning Representations for Hierarchies with Minimal Support.
    NeurIPS 2024Machine Learning
  3. Shib Dasgupta, Andrew McCallum, Steffen Rendle, Li Zhang. Answering Compositional Queries with Set-Theoretic Embeddings.
    ArXiv 2023Information Retrieval
  4. Shib Dasgupta, Michael Boratko, S. Atmakuri, Xiang Lorraine Li, D. Patel, Andrew McCallum. Word2Box: Capturing Set-Theoretic Semantics of Words using Box Embeddings.
    ACL 2022NLP
    👉 [Talk Link]
  5. EunJeong Hwang, Veronika Thost, Shib Sankar Dasgupta, Tengfei Ma. An Analysis of Virtual Nodes in Graph Neural Networks for Link Prediction.
    LoG 2022Graph Neural Networks
  6. Michael Boratko, D. Patel, Shib Dasgupta, Andrew McCallum. Measure-Theoretic Set Representation Learning.
    ArXiv 2022ML Theory
  7. Michael Boratko, Javier Burroni, Shib Sankar Dasgupta, Andrew McCallum. Min/Max Stability and Box Distributions.
    UAI 2021ML Theory
  8. Shib Dasgupta, Xiang Lorraine Li, Michael Boratko, Dongxu Zhang, Andrew McCallum. Box-To-Box Transformations for Modeling Joint Hierarchies.
    ACL (Rep4NLP) 2021Graph Learning
  9. Xuelu Chen, Michael Boratko, Muhao Chen, Shib Dasgupta, Xiang Li, Andrew McCallum. Probabilistic Box Embeddings for Uncertain Knowledge Graph Reasoning.
    NAACL 2021Graph Learning
  10. T. Chheda, P. Goyal, T. Trang, D. Patel, M. Boratko, Shib Dasgupta, Andrew McCallum. Box Embeddings: An Open-source Library for Representation Learning using Geometric Structures.
    EMNLP (Demo Track) 2021Open Source Contribution
  11. Shib Dasgupta*, Michael Boratko*, Dongxu Zhang, Luke Vilnis, Xiang Li, Andrew McCallum. Improving Local Identifiability for Probabilistic Box Embeddings.
    NeurIPS 2020Machine Learning
    👉 [Talk Link]
  12. Dhruvesh Patel*, Shib Sankar Dasgupta*, Michael Boratko, Xiang Li, Luke Vilnis, Andrew McCallum. Representing Joint Hierarchies with Box Embeddings.
    AKBC 2020Graph Learning
  13. Shib Dasgupta, Swayambhu Nath Ray, Partha Talukdar. HyTE: Hyperplane-based Temporally aware Knowledge Graph Embedding.
    EMNLP 2018Graph Learning
  14. Shikhar Vashishth, Shib Sankar Dasgupta, Swayambhu Nath Ray, Partha Talukdar. Dating Documents using Graph Convolution Networks.
    ACL 2018Graph Learning
  15. Swayambhu Nath Ray, Shib Sankar Dasgupta, Partha Talukdar. AD3: Attentive Deep Document Dater.
    EMNLP 2018Graph Learning

Skills

Published Software and Datasets

Mentorship Experience

Awards & Achievements