Improve model card: Add pipeline tag, library name, paper, and code links

#2
by nielsr HF Staff - opened
Files changed (1) hide show
  1. README.md +13 -4
README.md CHANGED
@@ -1,13 +1,22 @@
1
  ---
2
- license: llama3.1
 
3
  datasets:
4
  - allenai/winogrande
5
  - allenai/ai2_arc
6
  - google/boolq
7
  - wentingzhao/obqa
8
- base_model:
9
- - meta-llama/Llama-3.1-8B
10
  tags:
11
  - peft
12
  - bayesian
13
- ---
 
 
 
 
 
 
 
 
 
 
1
  ---
2
+ base_model:
3
+ - meta-llama/Llama-3.1-8B
4
  datasets:
5
  - allenai/winogrande
6
  - allenai/ai2_arc
7
  - google/boolq
8
  - wentingzhao/obqa
9
+ license: llama3.1
 
10
  tags:
11
  - peft
12
  - bayesian
13
+ pipeline_tag: text-generation
14
+ library_name: transformers
15
+ ---
16
+
17
+ This repository contains a low-rank adapter model, based on [Meta-Llama-3.1-8B](https://huggingface.co/meta-llama/Llama-3.1-8B), which was presented in the paper [Training-Free Bayesianization for Low-Rank Adapters of Large Language Models](https://huggingface.co/papers/2412.05723).
18
+
19
+ **Training-Free Bayesianization (TFB)** is a simple yet theoretically grounded framework that efficiently transforms trained low-rank adapters into Bayesian ones without additional training. TFB systematically searches for the maximally acceptable level of variance in the weight posterior, constrained within a family of low-rank isotropic Gaussian distributions. This approach aims to achieve superior uncertainty estimation and generalization compared to existing methods, while eliminating the need for complex Bayesianization training procedures.
20
+
21
+ For the code, installation instructions, and further details on how to use the TFB framework, please refer to the official GitHub repository:
22
+ [https://github.com/Wang-ML-Lab/bayesian-peft](https://github.com/Wang-ML-Lab/bayesian-peft)