
Mastering Long Context AI through MiniMax-01
MiniMax-01 achieves up to 4M tokens with lightning attention and MoE, setting new standards for
MiniMax-01 achieves up to 4M tokens with lightning attention and MoE, setting new standards for
Constitutional Classifiers provide a robust framework to defend LLMs against universal jailbreaks, leveraging adaptive filtering
Author(s): Mohamed Azharudeen M, Balaji Dhamodharan
Attention-Based Distillation efficiently compresses large language models by aligning attention patterns between teacher and student.
Author(s): Deewakar Thakyal, Adwait Kelkar, Biswajit Biswas
Author(s): Anik Chakraborty, Sayantani Ghosh, Raktim Chakraborty,Dr. Indranil Mitra, Prasun Nandy
Author(s): Jaiyesh Chahar, Pravar Kulbhushan, Rohini Das, Indrajit Kar
Author(s): Ranti Dev Sharma, Aditya Bhashkar, Divakar Roy, Anubhav Srivastava, Saravanan Murugan, Aparna Prabhu
Author(s): Sabeesh Ethiraj, Bharath Kumar Bolla
Author(s): Prithwis Mukerjee
Author(s): Yashaswini Viswanath, Sudha Jamthe, Suresh Lokiah
Author(s): Sabeesh Ethiraj, Bharath Kumar Bolla
Author(s): Jaydip Sen, Saikat Mondal, Sidra Mehtab
We noticed you're visiting from India. We've updated our prices to Indian rupee for your shopping convenience. Use United States (US) dollar instead. Dismiss