The widespread adoption of electric vehicles is intrinsically linked to the advancements in lithium-ion battery technology. These power sources are favored for their high energy density and relatively good cycle life. However, concerns regarding safety and range, particularly in extreme conditions, persist and are fundamentally tied to the materials within the battery, especially the electrolyte. Conventional lithium-ion battery systems employ liquid electrolytes, which are flammable and can lead to thermal runaway under conditions like overcharge or physical damage. This has catalyzed intense research into solid-state electrolytes (SSEs), which promise enhanced safety, higher energy density, and better performance across a wider temperature range. The development of novel SSE materials is crucial for the next generation of high-performance, safe lithium-ion battery systems.

The scientific literature contains a vast, ever-expanding repository of knowledge on synthesizing these promising SSE materials. This information, however, is locked within unstructured text. Manually extracting structured data—such as precise precursors, synthesis methods, processing steps, and resulting compounds—from thousands of publications is prohibitively time-consuming, labor-intensive, and prone to inconsistency. This creates a significant bottleneck. To accelerate the discovery and optimization of SSEs through data-driven methods like machine learning, we require efficient, accurate, and automated pipelines to transform this textual knowledge into structured, computable data. Traditional rule-based extraction systems lack the flexibility to handle the diverse and complex language used in scientific writing, while supervised machine learning methods demand large, expensively annotated datasets.
The emergence of large language models (LLMs) presents a transformative opportunity. Pre-trained on colossal text corpora, these models possess a profound understanding of language context and semantics. When fine-tuned on specific tasks with relatively small, high-quality datasets, they can exhibit remarkable performance in information extraction and text classification. This study explores the application of LLMs to automate the extraction of synthesis information for lithium-ion battery solid-state electrolytes from scientific literature. The proposed framework involves a two-stage process: first, identifying paragraphs that describe SSE synthesis within experimental sections, and second, extracting structured synthesis protocols from those identified paragraphs.
Methodology: A Two-Stage Extraction Pipeline
The overall workflow for extracting solid-state electrolyte synthesis information is designed to be systematic and automated. The pipeline consists of five core stages: Literature Acquisition & Parsing, Paragraph Classification, Synthesis Information Extraction, Data Structuring, and optionally, Synthesis Visualization.
First, relevant scientific articles are programmatically retrieved. To ensure high relevance, searches are typically constrained to titles containing keywords like “solid lithium battery” or “solid Li battery”. The downloaded articles, often in structured XML format, are then parsed to isolate the textual content most likely to contain synthesis details: namely, all paragraphs under standard section headings such as “Experimental”, “Methods”, or “Materials and Methods”. This initial filtering step is crucial for reducing the amount of text for subsequent, more computationally intensive analysis.
The core intelligence of the pipeline resides in the next two stages, both powered by fine-tuned language models. The parsed experimental section contains descriptions of various procedures, including electrode preparation, cell assembly, and characterization, alongside the target SSE synthesis. Therefore, the first model is tasked with Paragraph Classification. It analyzes each paragraph and classifies it as either describing a solid-state electrolyte synthesis process (positive class) or describing other experimental procedures (negative class). Only paragraphs classified as positive are passed forward.
The second model performs the Synthesis Information Extraction. It takes a synthesis paragraph as input and outputs a structured record. The output schema is predefined to capture key aspects of the synthesis protocol. For a lithium-ion battery solid-state electrolyte, this typically includes:
- Product: The final compound or material synthesized (e.g., Li7La3Zr2O12).
- Precursors: The starting chemicals and their quantities or molar ratios.
- Method: The general synthesis technique (e.g., solid-state reaction, sol-gel, mechanochemical milling).
- Steps: A sequential list of key actions and conditions (e.g., “mix precursors”, “calcine at 900°C for 6 h”, “pelletize under 300 MPa”).
The extracted structured information can be stored in a database or JSON format, creating a growing, queryable knowledge base of SSE synthesis. Furthermore, this structured data can be used to automatically generate synthesis route diagrams, providing a quick visual summary of the process.
Model Selection and Fine-Tuning Strategy
To evaluate different approaches, several models were selected and fine-tuned for the two distinct tasks. The choice balances performance, computational efficiency, and the growing ecosystem of open-source models.
For the paragraph classification task, we compare a traditional machine learning baseline with modern pre-trained language models:
- Naive Bayes with TF-IDF (NB-TFIDF): A strong traditional baseline for text classification. Term Frequency-Inverse Document Frequency (TF-IDF) is used to convert paragraphs into numerical feature vectors. The TF-IDF weight for a term \(t\) in document \(d\) from corpus \(D\) is calculated as:
$$ \text{tf-idf}(t, d, D) = \text{tf}(t, d) \times \text{idf}(t, D) $$
where \(\text{tf}(t, d)\) is the frequency of \(t\) in \(d\), and \(\text{idf}(t, D) = \log \frac{N}{|\{d \in D: t \in d\}|}\) with \(N\) being the total number of documents. A Multinomial Naive Bayes classifier is then trained on these features. - BERT (Bidirectional Encoder Representations from Transformers): A transformer-based model pre-trained on a large corpus using masked language modeling. It captures deep bidirectional context and is fine-tuned for the binary classification task.
- Large Language Models (LLMs): We fine-tune several instruction-tuned, open-source LLMs of varying sizes: LLaMA-3.2-3B-Instruct (3B), Gemma-7B-Instruct (7B), and LLaMA-3.1-8B-Instruct (8B). Instruction tuning allows them to better follow task-specific prompts.
For the synthesis information extraction task, which requires more nuanced language understanding and structured generation, we employ the same suite of LLMs (3B, 7B, 8B).
Fine-tuning these large models, especially the LLMs, requires efficient techniques to manage computational cost. We employ Parameter-Efficient Fine-Tuning (PEFT), specifically the Low-Rank Adaptation (LoRA) method. Instead of updating all billions of parameters, LoRA freezes the pre-trained model weights and injects trainable rank-decomposition matrices into transformer layers, dramatically reducing the number of trainable parameters and memory requirements. The training objective is to minimize the cross-entropy loss between the model’s predictions and the ground-truth labels (for classification) or tokens (for information extraction). For a classification task with true label \(y\) and predicted probability distribution \(\hat{y}\), the loss \(L\) for a sample is:
$$ L = -\sum_{c=1}^{C} y_c \log(\hat{y}_c) $$
where \(C\) is the number of classes (2 for our classifier).
Data Preparation for Model Training
The performance of supervised models hinges on the quality of the training data. We constructed specialized datasets for each task.
1. Paragraph Classification Dataset: A dataset of 300 paragraphs was curated from the parsed “Experimental” sections of SSE-related literature. Each paragraph was labeled as ‘1’ (SSE synthesis) or ‘0’ (other procedures). To ensure accuracy, labels were assigned using a combination of heuristic rules (e.g., paragraph sub-headings), LLM-assisted pre-annotation (using a model like ChatGPT for initial suggestions), and finally, meticulous human verification. The dataset was split into 200 samples for training/validation and 100 held-out samples for testing. A sample data point includes the paragraph text and its binary label.
2. Synthesis Information Extraction Dataset: This dataset requires structured input-output pairs. We leveraged and adapted a previously built dataset from the domain of inorganic catalyst synthesis, which shares similar linguistic patterns for describing solid-state synthesis protocols. The dataset contains 150 examples in a standardized format (e.g., OpenAI’s chat format). Each example consists of:
- System Prompt: Instructions defining the extraction task and output schema (Product, Precursors, Method, Steps).
- User Input (Synthesis Paragraph): The raw text describing a synthesis.
- Assistant Output (Structured Record): The correct JSON-like structured extraction.
This dataset was used entirely for fine-tuning the LLMs for the extraction task. For final evaluation, 50 genuine SSE synthesis paragraphs, identified by the classification model, were used.
| Model Type | Specific Model | Key Characteristics | Training Mechanism |
|---|---|---|---|
| Traditional ML | Naive Bayes + TF-IDF | Statistical, bag-of-words representation. | Train on TF-IDF vectors. |
| Pre-trained Encoder | BERT-large-uncased | Deep bidirectional context, 340M parameters. | Full fine-tuning of classification head. |
| Large Language Model | LLaMA-3.2-3B-Instruct | Instruction-tuned, 3B parameters. | Parameter-Efficient Fine-Tuning (LoRA). |
| Large Language Model | Gemma-7B-Instruct | Instruction-tuned, ~8B parameters. | Parameter-Efficient Fine-Tuning (LoRA). |
| Large Language Model | LLaMA-3.1-8B-Instruct | Instruction-tuned, 8B parameters. | Parameter-Efficient Fine-Tuning (LoRA). |
| Model | Learning Rate | Batch Size | Gradient Accumulation Steps | Epochs |
|---|---|---|---|---|
| LLaMA-3.2-3B-Instruct | 5.0×10–5 | 1 | 8 | 20 |
| Gemma-7B-Instruct | 5.0×10–5 | 1 | 8 | 20 |
| LLaMA-3.1-8B-Instruct | 5.0×10–5 | 1 | 8 | 20 |
| BERT-large | 3.0×10–6 | 1 | 8 | 20 |
Results and Discussion
Performance on Paragraph Classification
The primary goal of this stage is to reliably filter out irrelevant text. We evaluate the models using standard metrics: Precision (fraction of predicted SSE paragraphs that are correct), Recall (fraction of all true SSE paragraphs that are found), and their harmonic mean, the F1 Score. The F1 score is defined as:
$$ F1 = 2 \times \frac{\text{Precision} \times \text{Recall}}{\text{Precision} + \text{Recall}} $$
A higher F1 score indicates a better balance between precision and recall.
The NB-TFIDF model serves as a competent baseline but is outperformed by the deep learning-based models. Its reliance on word statistics without deep semantic understanding limits its ability to disambiguate complex descriptions. For instance, a paragraph discussing “mixing the electrolyte with the electrode powder” might contain keywords associated with synthesis but is actually describing cell assembly.
BERT and the LLMs, having been pre-trained on vast and diverse text, demonstrate a superior grasp of context. They can understand that a phrase like “the solid electrolyte was prepared by…” is a strong indicator of synthesis, even if the specific keywords are common. Fine-tuning further specializes this capability.
| Model | Precision | Recall | F1 Score |
|---|---|---|---|
| Naive Bayes + TF-IDF | 0.82 | 0.84 | 0.83 |
| BERT-large-uncased | 0.88 | 0.86 | 0.87 |
| LLaMA-3.2-3B-Instruct | 0.89 | 0.88 | 0.885 |
| Gemma-7B-Instruct | 0.92 | 0.90 | 0.91 |
| LLaMA-3.1-8B-Instruct | 0.88 | 0.90 | 0.89 |
The results show that the Gemma-7B model achieved the highest F1 score, exceeding 0.9, indicating excellent classification performance. The LLaMA 3B and 8B models also performed very well, with F1 scores above 0.88. It is noteworthy that the smaller 3B model’s performance is comparable to its larger 8B counterpart. This suggests that for this specific, well-defined classification task, a smaller, efficiently tuned model can be sufficient, offering advantages in deployment speed and cost. The confusion matrices revealed that the most common error was misclassifying a non-SSE paragraph (label 0) as an SSE paragraph (label 1). This often occurred in paragraphs describing electrode slurry preparation, which involves “mixing” and “drying” steps semantically similar to some solid-state synthesis procedures. This is a challenging edge case that requires nuanced context to resolve.
| Model | True Negative (TN) | False Positive (FP) | False Negative (FN) | True Positive (TP) |
|---|---|---|---|---|
| Naive Bayes | 41 | 9 | 8 | 42 |
| BERT | 39 | 11 | 2 | 48 |
| LLaMA-3.2-3B | 42 | 8 | 6 | 44 |
| Gemma-7B | 45 | 5 | 5 | 45 |
| LLaMA-3.1-8B | 43 | 7 | 5 | 45 |
Performance on Synthesis Information Extraction
This task is more complex, requiring the model to identify specific entities and their relations within a paragraph and format them into a strict schema. We evaluate extraction performance at the entity level. For each of the four entity types (Product, Precursors, Method, Steps) in a paragraph, we calculate Precision, Recall, and F1 Score. A predicted entity is considered correct (True Positive) if it matches a ground-truth entity. The final scores for a model are micro-averages across all entities in the 50 test paragraphs.
Remarkably, all three fine-tuned LLMs performed exceptionally well on this task, with overall F1 scores exceeding 0.9. This underscores the power of LLMs’ instruction-following and text comprehension abilities when guided by appropriate fine-tuning. The models successfully learned to parse dense scientific prose, ignore irrelevant details, and output clean structured data.
| Model | Product Entities | Precursor Entities | Method Entities | Step Entities | Total Entities Extracted |
|---|---|---|---|---|---|
| LLaMA-3.2-3B-Instruct | 50 | 187 | 52 | 215 | 504 |
| Gemma-7B-Instruct | 49 | 182 | 51 | 208 | 490 |
| LLaMA-3.1-8B-Instruct | 50 | 185 | 53 | 212 | 500 |
An equally important finding is the demonstrated strong generalization capability. The LLMs were fine-tuned on a dataset originally constructed for inorganic catalyst synthesis, not specifically for lithium-ion battery solid-state electrolytes. Their high performance on the SSE test set indicates they did not merely memorize patterns from the catalyst domain. Instead, they learned the fundamental task of “reading a synthesis paragraph and extracting product, precursors, method, and steps” as a generalizable skill. This is a significant advantage, reducing the need for massive, domain-specific annotation efforts for every new material class.
Analysis of errors showed that lower precision was typically caused by the model extracting entities outside the predefined schema (e.g., extracting a characterization temperature as a synthesis step). Lower recall was often due to the model missing some precursor compounds listed in a complex mixture or consolidating multiple detailed actions into a single, broader step. The 3B model slightly outperformed the 7B and 8B models in this extraction task, reinforcing the observation that larger parameter count does not automatically guarantee better performance for focused, fine-tuned applications. The model architecture, training data quality, and fine-tuning strategy are critical co-factors.
| Model | Average Precision | Average Recall | Average F1 Score |
|---|---|---|---|
| LLaMA-3.2-3B-Instruct | 0.92 | 0.91 | 0.915 |
| Gemma-7B-Instruct | 0.90 | 0.89 | 0.895 |
| LLaMA-3.1-8B-Instruct | 0.91 | 0.90 | 0.905 |
Conclusion and Implications
This work successfully demonstrates a functional and effective pipeline for automatically extracting structured synthesis information for lithium-ion battery solid-state electrolytes from scientific literature. The two-stage approach, leveraging fine-tuned large language models, addresses a critical bottleneck in materials informatics.
Key findings include:
- High-Performance Classification: Modern language models (BERT and LLMs) significantly outperform traditional TF-IDF based methods in identifying synthesis-relevant paragraphs, with F1 scores above 0.85 and reaching over 0.9. This provides a reliable filter for downstream processing.
- Accurate and Generalizable Extraction: Fine-tuned LLMs excel at the complex task of parsing synthesis text and outputting structured records. With F1 scores exceeding 0.9, they prove capable of high-precision information extraction. Crucially, they exhibit strong generalization, effectively transferring knowledge learned from one materials domain (catalysts) to another (solid-state electrolytes for lithium-ion battery).
- Efficiency of Smaller Models: The comparable, and sometimes superior, performance of the 3-billion parameter model relative to the 7B and 8B models challenges the notion that bigger is always better for specialized tasks. Smaller models fine-tuned with techniques like LoRA offer a compelling balance of performance, speed, and lower computational resource requirements, making them highly practical for research and deployment.
The implications for the field of lithium-ion battery research are substantial. This pipeline can be scaled to process thousands of publications, building a comprehensive, structured database of solid-state electrolyte synthesis protocols. Such a database is an invaluable resource for data-driven discovery. Researchers can use it to:
- Perform large-scale meta-analyses to identify correlations between synthesis parameters and material properties (ionic conductivity, stability).
- Train machine learning models to predict optimal synthesis conditions for target properties.
- Discover novel, promising synthesis pathways by analyzing patterns across the literature.
- Rapidly survey the state-of-the-art for a specific class of SSEs.
Future work will focus on expanding the extraction schema to include more detailed parameters (e.g., exact temperatures, times, atmospheric conditions), handling multi-paragraph synthesis descriptions, and integrating the extracted data directly into materials property prediction models. The success of this approach paves the way for a new paradigm of accelerated knowledge extraction and materials design for safer, higher-performance lithium-ion battery technologies.
