Summarization
Transformers
Safetensors
English
phi
text-generation
arxiv
custom_code
text-generation-inference
Instructions to use AlgorithmicResearchGroup/phi-physics with libraries, inference providers, notebooks, and local apps. Follow these links to get started.
- Libraries
- Transformers
How to use AlgorithmicResearchGroup/phi-physics with Transformers:
# Use a pipeline as a high-level helper # Warning: Pipeline type "summarization" is no longer supported in transformers v5. # You must load the model directly (see below) or downgrade to v4.x with: # 'pip install "transformers<5.0.0' from transformers import pipeline pipe = pipeline("summarization", model="AlgorithmicResearchGroup/phi-physics", trust_remote_code=True)# Load model directly from transformers import AutoTokenizer, AutoModelForCausalLM tokenizer = AutoTokenizer.from_pretrained("AlgorithmicResearchGroup/phi-physics", trust_remote_code=True) model = AutoModelForCausalLM.from_pretrained("AlgorithmicResearchGroup/phi-physics", trust_remote_code=True) - Notebooks
- Google Colab
- Kaggle
| license: apache-2.0 | |
| language: | |
| - en | |
| pipeline_tag: summarization | |
| widget: | |
| - text: What is the peak phase of T-eV? | |
| example_title: Question Answering | |
| tags: | |
| - arxiv | |
| # Table of Contents | |
| 0. [TL;DR](#TL;DR) | |
| 1. [Model Details](#model-details) | |
| 2. [Usage](#usage) | |
| 3. [Uses](#uses) | |
| 4. [Citation](#citation) | |
| # TL;DR | |
| This is a Phi-1_5 model trained on [camel-ai/physics](https://huggingface.co/datasets/camel-ai/physics). This model is for research purposes only and ***should not be used in production settings***. | |
| ## Model Description | |
| - **Model type:** Language model | |
| - **Language(s) (NLP):** English | |
| - **License:** Apache 2.0 | |
| - **Related Models:** [Phi-1_5](https://huggingface.co/microsoft/phi-1_5) | |
| # Usage | |
| Find below some example scripts on how to use the model in `transformers`: | |
| ## Using the Pytorch model | |
| ```python | |
| from huggingface_hub import notebook_login | |
| from datasets import load_dataset, Dataset | |
| from transformers import AutoModelForCausalLM, AutoTokenizer, TextStreamer | |
| model = "ArtifactAI/phi-physics" | |
| model = AutoModelForCausalLM.from_pretrained(base_model, trust_remote_code= True) | |
| tokenizer = AutoTokenizer.from_pretrained(base_model, trust_remote_code=True) | |
| def generate(prompt): | |
| inputs = tokenizer(f'''Below is an instruction that describes a task. Write a response that appropriately completes the request If you are adding additional white spaces, stop writing".\n\n### Instruction:\n{prompt}.\n\n### Response:\n ''', return_tensors="pt", return_attention_mask=False) | |
| streamer = TextStreamer(tokenizer, skip_prompt= True) | |
| _ = model.generate(**inputs, streamer=streamer, max_new_tokens=500) | |
| generate("What are the common techniques used in identifying a new species, and how can scientists accurately categorize it within the existing taxonomy system?") | |
| ``` | |
| ## Training Data | |
| The model was trained on [camel-ai/phi-physics](https://huggingface.co/datasets/camel-ai/physics), a dataset of question/answer pairs. | |
| ## Training procedure | |
| The following `bitsandbytes` quantization config was used during training: | |
| - quant_method: bitsandbytes | |
| - load_in_8bit: False | |
| - load_in_4bit: True | |
| - llm_int8_threshold: 6.0 | |
| - llm_int8_skip_modules: None | |
| - llm_int8_enable_fp32_cpu_offload: False | |
| - llm_int8_has_fp16_weight: False | |
| - bnb_4bit_quant_type: nf4 | |
| - bnb_4bit_use_double_quant: True | |
| - bnb_4bit_compute_dtype: float16 | |
| ### Framework versions | |
| - PEFT 0.6.2 | |
| ## Training procedure | |
| The following `bitsandbytes` quantization config was used during training: | |
| - quant_method: bitsandbytes | |
| - load_in_8bit: False | |
| - load_in_4bit: True | |
| - llm_int8_threshold: 6.0 | |
| - llm_int8_skip_modules: None | |
| - llm_int8_enable_fp32_cpu_offload: False | |
| - llm_int8_has_fp16_weight: False | |
| - bnb_4bit_quant_type: nf4 | |
| - bnb_4bit_use_double_quant: True | |
| - bnb_4bit_compute_dtype: float16 | |
| ### Framework versions | |
| - PEFT 0.6.2 | |
| # Citation | |
| ``` | |
| @misc{phi-math, | |
| title={phi-physics}, | |
| author={Matthew Kenney}, | |
| year={2023} | |
| } | |
| ``` | |