Workshop

The First IEEE International Workshop on
Testing and Evaluation of Large Language Models (TELLMe 2024)
In Conjunction with
The Sixth IEEE International Conference on Artificial Intelligence Testing (AITest 2024)
15- 18 July 2024, Shanghai, China

Call for Papers

KEY DATES

  • June 7, 2024: Submission deadline (Firm)
  • June 15, 2024: Notification of acceptance/reject
  • June 20, 2024: Final paper submission (camera-ready) and authors registration
  • July 15- 18, 2024: Workshop dates

1. TOPICS OF INTEREST

The workshop focuses on bridging the gap between the theories and practice of testing and evaluation of LLMs and their applications. The topics cover all aspects of testing and evaluation of LLMs, which include, but are not limited to, the following:

1) Methodology: The testing and evaluation of LLMs may take place in different contexts, for example, as a part of the training, fine tuning, and application development. The methodological aspects of to perform testing and evaluation of LLM in such contexts include: 

  • processes of testing and evaluation,
  • quality assurance methodology for LLMs, such as quality attributes and performance metrics of LLMs,
  • benchmark construction, validation, and application methodology,
  • integration of testing and evaluation of LLMs with machine learning research methodology such as machine learning development and operations (MLDevOps), AI application development methodology and software engineering methodology,

2) Technology: Techniques and methods used in various activities in the testing and evaluation of LLMs, and support the methodology, which include

  • techniques for test data generation, selection, cleaning, labelling, balancing, etc. for testing and evaluation of LLMs
  • techniques for statistical analysis, visualisation, and comparison of test results,
  • techniques for test scenario identification, formulation, representation, combination, and coverage,
  • techniques for test adequacy and requirements definition and measurement,
  • techniques for testing and evaluation of various applications of LLMs, such as in program code generation, in various other software engineering tasks, in audio, video and text generations, etc.
  • techniques for testing and evaluation of LLM algorithms, such as ablation studies
  • techniques for testing and evaluation of LLM capabilities on various text and multi-modal language processing tasks
  • techniques for testing and evaluation of LLM on various specific aspects, such as robustness, bias, hallucination, safety, privacy, and other ethical issues,

3) Tools and environments: Issues in the development, operation, maintenance and evolution of testing and evaluation tools, platforms, library code, infrastructures and environments that enables the testing and evaluation of LLMs, such as feature stores, open-source platforms, etc.

2. TYPES OF SUBMISSIONS

  • The workshop will accept the following types of submissions.
    Regular Papers (6 pages): Two types of regular papers are encouraged. The first is research papers that report on unpublished original research results on a relevant topic of the workshop. The second is industry case studies that reports the current practice in the industry. Test practitioners are especially encouraged to contribute their experiences and insights on testing and evaluation of LLMs.
  • Work-In-Progress Reports (4 pages): Short papers on the research project in progress to report the following aspects of project:
    o The research problem to be solved, e.g. research questions of the project and/or the hypothesis of the research,
    o The project’s goals and objectives,
    o The methodology, the theoretical foundation, and/or the technical approach,
    o The work plan, and
    o The current state of the project,
    o Funding budget and cost of the project, etc.
  • Tool Demonstration (Flexible, <6 pages): The workshop will run a tool demonstration session. To participate in the tool demonstration session, the demonstrator should submit an extended abstract to describe the key features and functionality of the tool together with an outline of the demonstration.
  • Position statement for charette discussion (1 – 2 pages): The workshop will run one or two charette discussion sessions on topics of interests to the participants. In order to be a presenter in such a discussion session, you must submit a position statement (one two pages) to state the problem that you are interested in, the importance of the problem, and your vision on the potential solutions of the problem, etc.

3. SUBMISSIONS AND REVIEWS
All papers submitted to the workshop must be unpublished original work and must not have been submitted anywhere else for publication. The paper submissions must be in English and conform to the IEEE Computer Society’s conference proceedings format. 

All submissions will be reviewed by three PC members of the workshop. Regular research papers will be judged based on their relevance, originality, contribution, significance, and clarity. Industry case study, work-in-progress report and tool demonstration papers will be judged based on the relevance, soundness, interests to the audience, and clarity. Position statement will be judged based on their relevance of the topic, importance of the problem, convincing arguments, and soundness of the vision.

Please note that the workshop will comply to IEEE Publications Policy about AI-Generated Contents. The use of content generated by artificial intelligence (AI) in an article (including but not limited to text, figures, images, and code) shall be disclosed in the acknowledgements section of any article submitted to an IEEE publication. The sections of the paper that use AI-generated text shall have a citation to the AI system used to generate the text. Using AI systems for editing and grammar enhancement is common practice and, as such, is generally outside the intent of the above policy. This case also recommends the same disclosure. For additional information, please read: https://conferences.ieeeauthorcenter.ieee.org/author-ethics/guidelines-and-policies/submission-policies/

4. KEY DATES

  • June 7, 2024: Submission deadline (Firm)
  • June 15, 2024: Notification of acceptance/reject
  • June 20, 2024: Final paper submission (camera-ready) and authors registration
  • July 15- 18, 2024: Workshop dates

5. PUBLICATION OF PAPERS AND PARTICIPATION
Each accepted paper is required to have a unique author registration to the workshop, who must give a presentation in person at the workshop and participate in person at the sessions of the workshop. All accepted submissions that satisfy the above requirements will be published in the IEEE AITest 2024 Conference Proceedings by IEEE and included in the IEEE Digital Library Xplore.

People without accepted submissions can also participate in the workshop through audience registration. All participants will have access to the AITest 2024 conference proceedings, attend all sessions of the workshop and engage in discussions.