December 15-18, 2023

From Theory to Practice: Workshop on Large Language

and Foundation Models

Workshop held in conjunction with IEEE Big Data 2023

Foundation models and large language models have demonstrated impressive outcomes across diverse applications, including Chatbot design, Natural Language Interaction, Text Summarization, and more. Among the crucial components contributing to the success of these models, the Transformer stands out as a central focus of extensive research and development. We cordially invite researchers and practitioners to submit their original work to the Workshop on Foundation Models and Large Language Models. The workshop aims to foster collaboration among experts from both academia and industry, providing a platform to delve into the most recent advancements, address challenges, and explore opportunities in the realm of foundation models and large language models."

Topics of interest for the workshop include, but are not limited to:

Model Training and Optimization:

  • - Techniques to deal with Hallucinations
  • - Training data for LLMs
  • - Efficient and Stable training of LLMs
  • - Efficient Tuning Methods
  • - Scalable approaches for distributed model training
  • - Middleware for scale out data preparation for LLM training
  • - Workflow orchestration for end-to-end LLM life cycle
  • - Resource management for compute and energy efficient model training
  • - Masking techniques
  • - Representation learning

Model Utilization and Integration:

  • - Using LLMs effectively as tools for Reinforcement Learning
  • - Enhancing LLM capabilities by using external tools (Search engines, calculators, etc)
  • - Visual Prompt Tuning and in-context learning
  • - Enable easy experimentation with high utilization to train foundational models in the cloud.
  • - Strategies to scale resources for training/fine-tuning foundational models
  • - Instruction tuning including generation of instruction tuning data.
  • - Parallel training: data model tensor (attention and weights)
  • - Distributed workflows for data cleansing and model usage (Langchain)
  • - Next arch such as sparse transformers
  • - Principled AI

Application-Specific Models:

  • - Math LLM
  • - MultiModal Foundation Models
  • - Trustworthy Foundation Models
  • - Large-scale Visual Foundation Models
  • - Timeseries foundation models for forecasting, prediction and control

Knowledge Incorporation and Adaptation:

  • - Approaches to deal with knowledge recency to effectively update knowledge within LLMs.
  • - Incorporating domain knowledge

Evaluation and Benchmarking:

  • - Additional Benchmarks to fill gap between human and automatic reference-based evaluation

Call for Papers

Paper Submission Deadline: Nov 01, 2023, 11:59 PM AoE.

Paper Notification: Nov. 15, 2023, 11:59 PM AoE.

Camera Ready Version: Nov. 22, 2023, 11:59 PM AoE.

Workshop: Dec-15 2023

Authors are invited to submit their papers electronically. Please submit a full-length paper (up to 10 page IEEE 2-column format, reference pages dont count in the 10 pages) through the online submission system at: Submission Formatting Instructions: Papers should be formatted to IEEE Computer Society Proceedings Manuscript Formatting Guidelines.

Accepted Papers with Registration

Program

1. Wealth of Nations, Wealth of Data: How GDP Shapes Diverse Large Language Models like ChatGPT
2. Enabling Dataspaces Using Foundation Models: Technical, Legal and Ethical Considerations and Future Trends
3. Model-as-a-Service (MaaS): A Survey
4. TRANSQLATION: TRANsformer-based SQL RecommendATION
5. GPT-in-the-Loop: Supporting Adaptation in Multiagent Systems
6. Hallucination-minimized Data-to-answer Framework for Financial Decision-makers
7. Unsupervised Style Transfer of Modern Hebrew using Generative Language Modeling and Zero-Shot Prompting
8. Towards Automated Regulatory Compliance Verification in Financial Auditing with Large Language Models
9. Comparing Generative Chatbots Based on Process Requirements: A Case Study
10. Generating Prototypes for Contradiction Detection Using Large Language Models and Linguistic Rules
11. Anaphoric Ambiguity Resolution in Software Requirement Texts

PC Members

ID Name Affiliation Country
1 Paulo Alencar Alencar None USA
2 Samuel Belkadi, Belkadi None UK
3 Armin Berger Fraunhofer IAIS Germany
4 Christodoulos Christodoulos USA
5 Wensheng Gan Jinan University China
6 Amir H. Payberah None USA
7 Jie Hong Hong None USA
8 Srideepika Jayaraman IBM United States
9 Pavel Kaganovich None USA
10 Alex Kaplunovich University of Maryland USA
11 Luis Fernando Lins University of Waterloo Canada
12 Cristiano Malossi IBM USA
13 Nathalia Nascimento University of Waterloo Canada
14 Dhaval Patel IBM USA
15 Maren Pielka Fraunhofer IAIS Germany
16 Punit Punit None USA
17 Sifa Rafet Fraunhofer IAIS and University of Bonn Germany
18 Sohini Roychowdhury None United States
19 Ahmet Soylu Oslo Metropolitan University Norway
20 Shirin Tahmasebinotarki KTH Royal Institute of Technology Sweden
21 Nianjun Zhou IBM USA

WORKSHOP ORGANIZERS

Workshop Chairs

Image

Dr. Dhaval Patel

IBM Research
Image

Prof. Dr. Rafet Sifa

University of Bonn and Fraunhofer Institute for Intelligent Analysis and Information Systems

Workshop Area Chairs

Image

Dr. Cristiano Malossi

IBM Research
Image

Dr. Carlos Costa

IBM Research
Image

Dr. Sameep Mehta

IBM Research
Image

Dr. Chandra Reddy

IBM Research
Image

Dr. Kaoutar El Maghraoui

IBM Research
Image

Dr. Bradley Eck

IBM Research
Image

Dr. Yikang Shen

IBM Research

Steering Chairs

Image

Dr. Jayant Kalangnanam

IBM Research
Image

Prof. Nitesh Chawla

University of Notre Dame

Contact: pateldha@us.ibm.com