urchade commited on
Commit
839d6dc
·
verified ·
1 Parent(s): 0176603

Upload GLiNER2

Browse files
README.md CHANGED
@@ -1,3 +1,6 @@
 
 
 
1
  ## Model Description
2
 
3
  GLiNER2 extends the original GLiNER architecture to support multi-task information extraction with a schema-driven interface. This base model provides efficient CPU-based inference while maintaining high accuracy across diverse extraction tasks.
 
1
+ ---
2
+ {}
3
+ ---
4
  ## Model Description
5
 
6
  GLiNER2 extends the original GLiNER architecture to support multi-task information extraction with a schema-driven interface. This base model provides efficient CPU-based inference while maintaining high accuracy across diverse extraction tasks.
config.json CHANGED
@@ -4,5 +4,6 @@
4
  "max_width": 8,
5
  "model_name": "microsoft/deberta-v3-base",
6
  "model_type": "extractor",
7
- "transformers_version": "4.51.3"
 
8
  }
 
4
  "max_width": 8,
5
  "model_name": "microsoft/deberta-v3-base",
6
  "model_type": "extractor",
7
+ "token_pooling": "first",
8
+ "transformers_version": "4.51.0"
9
  }
encoder_config/config.json CHANGED
@@ -27,7 +27,7 @@
27
  "relative_attention": true,
28
  "share_att_key": true,
29
  "torch_dtype": "float32",
30
- "transformers_version": "4.51.3",
31
  "type_vocab_size": 0,
32
  "vocab_size": 128011
33
  }
 
27
  "relative_attention": true,
28
  "share_att_key": true,
29
  "torch_dtype": "float32",
30
+ "transformers_version": "4.51.0",
31
  "type_vocab_size": 0,
32
  "vocab_size": 128011
33
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3a22b85f3b81a66869eb6c61c960f26cac5606f3911255ddbb39f76d6d04d0f6
3
  size 833938108
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:845fc4bd93c525b86124c58ab4f56c9eacf8587953086b14c501fab25957c007
3
  size 833938108
tokenizer_config.json CHANGED
@@ -145,7 +145,7 @@
145
  "sep_token": "[SEP]",
146
  "sp_model_kwargs": {},
147
  "split_by_punct": false,
148
- "tokenizer_class": "DebertaV2Tokenizer",
149
  "unk_token": "[UNK]",
150
  "vocab_type": "spm"
151
  }
 
145
  "sep_token": "[SEP]",
146
  "sp_model_kwargs": {},
147
  "split_by_punct": false,
148
+ "tokenizer_class": "DebertaV2TokenizerFast",
149
  "unk_token": "[UNK]",
150
  "vocab_type": "spm"
151
  }