About 6,160,000 results
Open links in new tab
  1. Evaluating and documenting AI models and data across performance dimensions informed by system and mission constraints. Provides a high-level overview of critical T&E concepts that will be …

  2. To evaluate DeepSeek models, which are available as open-weight models, CAISI downloaded their model weights from the model sharing platform Hugging Face and deployed the models on CAISI’s …

  3. Large Language Models Pretraining (and how to train transformers for language modeling) Pretraining The big idea that underlies all the amazing performance of language models retrain a transf

  4. Large-language models (LLMs) and multi-modal models like text and image are enabling new capabilities, from code generation to the creation of images based on natural language descriptions.

  5. Dual-use foundation models with widely available model weights (referred to in this Report as open foun-dation models) introduce a wide spectrum of benefits. They diversify and expand the array of actors, …

  6. Large language models (LLMs) are powerful but static; they lack mechanisms to adapt their weights in response to new tasks, knowledge, or examples. We introduce Self-Adapting LLMs (SEAL), a …

  7. In this paper we explore instruction finetuning with a particular focus on (1) scaling the number of tasks, (2) scaling the model size, and (3) finetuning on chain-of-thought data.