ADaSci Banner 2024

Top 5 AI Papers Published This Week

Recent AI research highlights advancements in multi-token prediction, open-source evaluation models, and flexible frameworks for specialized tasks.

The AI research community has been abuzz with several groundbreaking papers published this week. These papers highlight significant advancements in large language models (LLMs), model evaluation techniques, conversational question answering, and innovative approaches to adapting models for specialized tasks. Here’s a look at some of the most notable contributions.

Better & Faster Large Language Models via Multi-token Prediction

Authors: Fabian Gloeckle, Badr Youbi Idrissi, Baptiste Rozière, David Lopez-Paz, Gabriel Synnaeve
Source: arXiv:2404.19737

This paper proposes a novel training approach for large language models (LLMs) by predicting multiple future tokens simultaneously instead of the traditional next-token prediction. The authors suggest that multi-token prediction can lead to higher sample efficiency, especially beneficial for larger models and generative tasks. Key highlights include:

  • Multi-token Prediction: The model predicts multiple future tokens at each position using independent output heads on a shared model trunk.
  • Improved Performance: This approach improves downstream capabilities without additional training time, particularly excelling in generative benchmarks like coding.
  • Efficiency Gains: Models trained with 4-token prediction can be up to three times faster during inference, demonstrating significant performance enhancements.

Prometheus 2: An Open Source Language Model Specialized in Evaluating Other Language Models

Authors: Seungone Kim, Juyoung Suk, Shayne Longpre, Bill Yuchen Lin, Jamin Shin, Sean Welleck, Graham Neubig, Moontae Lee, Kyungjae Lee, Minjoon Seo
Source: arXiv:2405.01535

Prometheus 2 addresses the need for transparent, controllable, and affordable evaluation models by introducing an open-source language model specialized in evaluating other LLMs. The paper highlights the limitations of existing evaluator models and presents Prometheus 2 as a superior alternative:

  • Human-like Judgments: Prometheus 2 closely mirrors human and GPT-4 judgments, ensuring high correlation and agreement with human assessments.
  • Flexible Evaluation: The model supports both direct assessment and pairwise ranking, accommodating user-defined evaluation criteria.
  • Public Availability: All models, code, and data are publicly accessible, promoting transparency and further research.

ChatQA: Surpassing GPT-4 on Conversational QA and RAG

Authors: Zihan Liu, Wei Ping, Rajarshi Roy, Peng Xu, Chankyu Lee, Mohammad Shoeybi, Bryan Catanzaro
Source: arXiv:2401.10225

ChatQA introduces a suite of models that outperform GPT-4 in retrieval-augmented generation (RAG) and conversational question answering (QA). The paper details a two-stage instruction tuning method and a dense retriever optimized for conversational QA:

  • Enhanced Generation: The proposed instruction tuning method significantly boosts RAG performance.
  • Effective Retrieval: The dense retriever achieves results comparable to state-of-the-art query rewriting models with reduced deployment costs.
  • Comprehensive Evaluation: The ChatRAG Bench evaluates the models on ten datasets, demonstrating that ChatQA models built on Llama2 and Llama3 outperform GPT-4.

Phi-3 Technical Report: A Highly Capable Language Model Locally on Your Phone

Authors: Marah Abdin, Sam Ade Jacobs, Ammar Ahmad Awan, Jyoti Aneja, Ahmed Awadallah, Hany Awadalla, Nguyen Bach, Amit Bahree, Arash Bakhtiari, Harkirat Behl, Alon Benhaim, Misha Bilenko, Johan Bjorck, Sébastien Bubeck, Martin Cai, Caio César Teodoro Mendes, Weizhu Chen, Vishrav Chaudhary, Parul Chopra, Allie Del Giorno, Gustavo de Rosa, Matthew Dixon, Ronen Eldan, Dan Iter, Amit Garg, Abhishek Goswami, Suriya Gunasekar, Emman Haider, Junheng Hao, Russell J. Hewett, Jamie Huynh, Mojan Javaheripi, Xin Jin, Piero Kauffmann, Nikos Karampatziakis, Dongwoo Kim, Mahoud Khademi, Lev Kurilenko, James R. Lee, Yin Tat Lee, Yuanzhi Li, Chen Liang, Weishung Liu, Eric Lin, Zeqi Lin, Piyush Madan, Arindam Mitra, Hardik Modi, Anh Nguyen, Brandon Norick, Barun Patra, Daniel Perez-Becker, Thomas Portet, Reid Pryzant, Heyang Qin, Marko Radmilac, Corby Rosset, Sambudha Roy, Olatunji Ruwase, Olli Saarikivi, Amin Saied, Adil Salim, Michael Santacroce, Shital Shah, Ning Shang, Hiteshi Sharma, Xia Song, Masahiro Tanaka, Xin Wang, Rachel Ward, Guanhua Wang, Philipp Witte, Michael Wyatt, Can Xu, Jiahang Xu, Sonali Yadav, Fan Yang, Ziyi Yang, Donghan Yu, Chengruidong Zhang, Cyril Zhang, Jianwen Zhang, Li Lyna Zhang, Yi Zhang, Yue Zhang, Yunan Zhang, Xiren Zhou
Source: arXiv:2404.14219

Phi-3-mini is a 3.8 billion parameter language model designed to run on a phone, achieving performance comparable to larger models like GPT-3.5. The innovation lies in the scaled-up dataset used for training and the model’s alignment for robustness and safety:

  • High Performance: Phi-3-mini achieves 69% on MMLU and 8.38 on MT-bench, rivaling Mixtral 8x7B and GPT-3.5.
  • Scalability: The paper also introduces phi-3-small and phi-3-medium models, showing significant performance improvements with larger datasets.
  • Practical Deployment: The model’s small size allows for deployment on mobile devices, making advanced AI capabilities more accessible.

X-LoRA: Mixture of Low-Rank Adapter Experts, a Flexible Framework for Large Language Models with Applications in Protein Mechanics and Molecular Design

Authors: Eric L. Buehler, Markus J. Buehler
Source: arXiv:2402.07148

X-LoRA introduces a mixture of expert strategy using low-rank adaptation (LoRA) to create fine-tuned large language models. The framework is designed for applications in protein mechanics and molecular design, leveraging biological principles of universality and diversity:

  • Dynamic Layer Mixing: The model dynamically mixes adapted layers using hidden states, enabling diverse and novel combinations for task solving.
  • Broad Applicability: X-LoRA can be implemented on any existing LLM without modifying the underlying structure, enhancing domain-specific capabilities.
  • Scientific Capabilities: The model excels in tasks like protein mechanics analysis, molecular design, and knowledge graph construction, providing quantitative predictions and reasoning over results.

These papers represent significant advancements in AI research, offering new methodologies, enhanced performance, and practical applications across various domains. The open-source nature of several of these contributions ensures that the broader research community can build upon these innovations, driving further progress in the field.

Picture of Association of Data Scientists

Association of Data Scientists

The Chartered Data Scientist Designation

Achieve the highest distinction in the data science profession.

Elevate Your Team's AI Skills with our Proven Training Programs

Strengthen Critical AI Skills with Trusted Generative AI Training by Association of Data Scientists.

Our Accreditations

Get global recognition for AI skills

Chartered Data Scientist (CDS™)

The highest distinction in the data science profession. Not just earn a charter, but use it as a designation.

Certified Data Scientist - Associate Level

Global recognition of data science skills at the beginner level.

Certified Generative AI Engineer

An upskilling-linked certification initiative designed to recognize talent in generative AI and large language models

Join thousands of members and receive all benefits.

Become Our Member

We offer both Individual & Institutional Membership.