OrlandoHugBot commited on
Commit
69c997a
·
verified ·
1 Parent(s): 4d8b1f3

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +2 -92
README.md CHANGED
@@ -9,7 +9,7 @@ pipeline_tag: image-text-to-text
9
  <img src="skywork-logo.png" alt="Introduction Image" width="500" height="400">
10
  </div>
11
 
12
- ## 📖 [Technical Report](https://github.com/SkyworkAI/Skywork-R1V/blob/main/Skywork_R1V2.pdf) | 💻 [GitHub](https://github.com/SkyworkAI/Skywork-R1V)
13
 
14
  <div align="center">
15
 
@@ -257,99 +257,9 @@ If you use this model in your research, please cite:
257
  primaryClass={cs.CV},
258
  url={https://arxiv.org/abs/2504.05599},
259
  }
260
-
261
- @misc{chris2025skyworkr1v2multimodalhybrid,
262
- title={Skywork R1V2: Multimodal Hybrid Reinforcement Learning for Reasoning},
263
- author={Chris and Yichen Wei and Yi Peng and Xiaokun Wang and Weijie Qiu and Wei Shen and Tianyidan Xie and Jiangbo Pei and Jianhao Zhang and Yunzhuo Hao and Xuchen Song and Yang Liu and Yahui Zhou},
264
- year={2025},
265
- eprint={2504.16656},
266
- archivePrefix={arXiv},
267
- primaryClass={cs.CV},
268
- url={https://arxiv.org/abs/2504.16656},
269
- }
270
- ```
271
-
272
- # Skywork-R1V2-38B-AWQ (中文说明)
273
-
274
-
275
- ## 使用方法
276
- 您可以使用不同的推理框架来使用这个量化模型:
277
-
278
- ### 使用 VLLM
279
-
280
- #### Python API
281
-
282
- ```python
283
- import os
284
- from vllm import LLM, SamplingParams
285
- from vllm.entrypoints.chat_utils import load_chat_template
286
- model_name = "Skywork/Skywork-R1V2-38B-AWQ" # 或本地路径
287
- llm = LLM(model_name,
288
- dtype='float16',
289
- quantization="awq",
290
- gpu_memory_utilization=0.85,
291
- max_model_len=4096,
292
- trust_remote_code=True,
293
- )
294
- # 在此添加您的推理代码
295
- ```
296
-
297
- #### OpenAI 兼容的 API 服务器
298
-
299
- ```bash
300
- MODEL_ID="Skywork/Skywork-R1V2-38B-AWQ" # 或本地路径
301
- CUDA_VISIBLE_DEVICES=0 \
302
- python -m vllm.entrypoints.openai.api_server \
303
- --model $MODEL_ID \
304
- --dtype float16 \
305
- --quantization awq \
306
- --port 23334 \
307
- --max-model-len 12000 \
308
- --gpu-memory-utilization 0.9 \
309
- --trust-remote-code
310
- ```
311
-
312
- ### 使用 LMDeploy
313
-
314
- ```python
315
- import os
316
- from lmdeploy import pipeline, TurbomindEngineConfig, ChatTemplateConfig
317
- from lmdeploy.vl import load_image
318
- model_path = "Skywork/Skywork-R1V2-38B-AWQ" # 或本地路径
319
- engine_config = TurbomindEngineConfig(cache_max_entry_count=0.75)
320
- chat_template_config = ChatTemplateConfig(model_name=model_path)
321
- pipe = pipeline(model_path,
322
- backend_config=engine_config,
323
- chat_template_config=chat_template_config,
324
- )
325
- # 示例:多模态推理
326
- image = load_image('table.jpg')
327
- response = pipe(('描述这个图片?', image))
328
- print(response.text)
329
  ```
330
 
331
- ## 硬件要求
332
-
333
- 与原始 FP16 模型相比,AWQ 量化减少了内存占用。我们建议:
334
-
335
- - 至少一块 30GB+ 显存的 GPU 用于推理
336
- - 对于更长上下文的最佳性能,建议使用 40GB+ 显存
337
-
338
- ## 引用
339
-
340
- 如果您在研究中使用此模型,请引用:
341
-
342
  ```bibtex
343
- @misc{peng2025skyworkr1vpioneeringmultimodal,
344
- title={Skywork R1V: Pioneering Multimodal Reasoning with Chain-of-Thought},
345
- author={Yi Peng and Chris and Xiaokun Wang and Yichen Wei and Jiangbo Pei and Weijie Qiu and Ai Jian and Yunzhuo Hao and Jiachun Pan and Tianyidan Xie and Li Ge and Rongxian Zhuang and Xuchen Song and Yang Liu and Yahui Zhou},
346
- year={2025},
347
- eprint={2504.05599},
348
- archivePrefix={arXiv},
349
- primaryClass={cs.CV},
350
- url={https://arxiv.org/abs/2504.05599},
351
- }
352
-
353
  @misc{chris2025skyworkr1v2multimodalhybrid,
354
  title={Skywork R1V2: Multimodal Hybrid Reinforcement Learning for Reasoning},
355
  author={Chris and Yichen Wei and Yi Peng and Xiaokun Wang and Weijie Qiu and Wei Shen and Tianyidan Xie and Jiangbo Pei and Jianhao Zhang and Yunzhuo Hao and Xuchen Song and Yang Liu and Yahui Zhou},
@@ -359,4 +269,4 @@ print(response.text)
359
  primaryClass={cs.CV},
360
  url={https://arxiv.org/abs/2504.16656},
361
  }
362
- ```
 
9
  <img src="skywork-logo.png" alt="Introduction Image" width="500" height="400">
10
  </div>
11
 
12
+ ## 📖 [Technical Report](https://arxiv.org/abs/2504.16656) | 💻 [GitHub](https://github.com/SkyworkAI/Skywork-R1V)
13
 
14
  <div align="center">
15
 
 
257
  primaryClass={cs.CV},
258
  url={https://arxiv.org/abs/2504.05599},
259
  }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
260
  ```
261
 
 
 
 
 
 
 
 
 
 
 
 
262
  ```bibtex
 
 
 
 
 
 
 
 
 
 
263
  @misc{chris2025skyworkr1v2multimodalhybrid,
264
  title={Skywork R1V2: Multimodal Hybrid Reinforcement Learning for Reasoning},
265
  author={Chris and Yichen Wei and Yi Peng and Xiaokun Wang and Weijie Qiu and Wei Shen and Tianyidan Xie and Jiangbo Pei and Jianhao Zhang and Yunzhuo Hao and Xuchen Song and Yang Liu and Yahui Zhou},
 
269
  primaryClass={cs.CV},
270
  url={https://arxiv.org/abs/2504.16656},
271
  }
272
+ ```