Jun 22 – 25, 2025
University of South Dakota
US/Central timezone

Fine-Tuning DeepSeek-R1 for AI-Driven Materials Discovery Using Unsloth and Domain-Specific Supervised Instruction

Not scheduled
20m
University of South Dakota

University of South Dakota

The Role of LLMs, Scientific ML, and Data-Driven Approaches in Materials Innovation

Speaker

David Zeng (Dakota State University)

Description

The application of large language models (LLMs) in scientific domains, particularly materials discovery, remains underexplored compared to general natural language tasks. This study presents a systematic fine-tuning of the DeepSeek-R1 model using Unsloth’s memory-efficient framework and domain-specific supervised instruction techniques. Three publicly available datasets—GPT-Narratives-for-Materials, MDF-LLM Resources, and MatChem-LLM—were curated and harmonized to develop a fine-grained instruction-response corpus relevant to material science properties, experimental conditions, and compositional analysis.
The fine-tuning process leveraged SFTTrainer with instruction tuning templates, a sequence length of 2048 tokens, and efficient gradient accumulation to adapt DeepSeek-R1 for expert-level retrieval, summarization, and hypothesis generation tasks in material science. Preliminary evaluations using a simulated GLUE-style protocol, including BLEU-based output quality assessment, indicate significant improvement in domain-specific language generation, with BLEU scores exceeding expectations on validation samples.
This work demonstrates that strategically fine-tuned LLMs can serve as foundational models for AI-driven material discovery knowledge bases, potentially accelerating hypothesis testing, experimental planning, and scientific reasoning. Future directions include integrating regression-based materials property prediction benchmarks and open-domain QA fine-tuning for broader applicability.

Primary author

David Zeng (Dakota State University)

Presentation materials

There are no materials yet.