| { | |
| "_name_or_path": "unum-cloud/uform-gen2-qwen-500m", | |
| "architectures": [ | |
| "VLMForCausalLM" | |
| ], | |
| "auto_map": { | |
| "AutoConfig": "configuration_uform_gen.VLMConfig", | |
| "AutoModel": "modeling_uform_gen.VLMForCausalLM", | |
| "AutoProcessor": "unum-cloud/uform-gen2-qwen-500m--processing_uform_gen.VLMProcessor" | |
| }, | |
| "image_encoder_hidden_size": 1280, | |
| "image_encoder_name_or_path": "unum-cloud/uform-vl-english-big", | |
| "image_encoder_num_heads": 16, | |
| "image_encoder_num_layers": 32, | |
| "image_encoder_patch_size": 14, | |
| "image_encoder_pooling": "cls", | |
| "image_pooler_intermediate_size": 3200, | |
| "image_pooler_num_attn_heads": 16, | |
| "image_size": 336, | |
| "image_token_id": 151646, | |
| "initializer_range": 0.02, | |
| "model_type": "vlm", | |
| "num_image_latents": 256, | |
| "text_decoder_name_or_path": "Qwen/Qwen1.5-0.5B-Chat", | |
| "torch_dtype": "float32", | |
| "transformers_version": "4.38.2", | |
| "use_cache": true | |
| } | |