mirror of
				https://github.com/explosion/spaCy.git
				synced 2025-10-30 23:47:31 +03:00 
			
		
		
		
	Merge pull request #12994 from explosion/docs/llm_main
Synch `llm_develop` with `llm_main`
This commit is contained in:
		
						commit
						830eba5426
					
				|  | @ -19,26 +19,20 @@ prototyping** and **prompting**, and turning unstructured responses into | ||||||
| An LLM component is implemented through the `LLMWrapper` class. It is accessible | An LLM component is implemented through the `LLMWrapper` class. It is accessible | ||||||
| through a generic `llm` | through a generic `llm` | ||||||
| [component factory](https://spacy.io/usage/processing-pipelines#custom-components-factories) | [component factory](https://spacy.io/usage/processing-pipelines#custom-components-factories) | ||||||
| as well as through task-specific component factories: | as well as through task-specific component factories: `llm_ner`, `llm_spancat`, `llm_rel`, | ||||||
| 
 | `llm_textcat`, `llm_sentiment` and `llm_summarization`. | ||||||
| - `llm_ner` |  | ||||||
| - `llm_spancat` |  | ||||||
| - `llm_rel` |  | ||||||
| - `llm_textcat` |  | ||||||
| - `llm_sentiment` |  | ||||||
| - `llm_summarization` |  | ||||||
| 
 | 
 | ||||||
| ### LLMWrapper.\_\_init\_\_ {id="init",tag="method"} | ### LLMWrapper.\_\_init\_\_ {id="init",tag="method"} | ||||||
| 
 | 
 | ||||||
| > #### Example | > #### Example | ||||||
| > | > | ||||||
| > ```python | > ```python | ||||||
| > # Construction via add_pipe with default GPT3.5 model and NER task | > # Construction via add_pipe with the default GPT 3.5 model and an explicitly defined task | ||||||
| > config = {"task": {"@llm_tasks": "spacy.NER.v3", "labels": ["PERSON", "ORGANISATION", "LOCATION"]}} | > config = {"task": {"@llm_tasks": "spacy.NER.v3", "labels": ["PERSON", "ORGANISATION", "LOCATION"]}} | ||||||
| > llm = nlp.add_pipe("llm") | > llm = nlp.add_pipe("llm", config=config) | ||||||
| > | > | ||||||
| > # Construction via add_pipe with task-specific factory and default GPT3.5 model | > # Construction via add_pipe with a task-specific factory and default GPT3.5 model | ||||||
| > parser = nlp.add_pipe("llm-ner", config=config) | > llm = nlp.add_pipe("llm-ner") | ||||||
| > | > | ||||||
| > # Construction from class | > # Construction from class | ||||||
| > from spacy_llm.pipeline import LLMWrapper | > from spacy_llm.pipeline import LLMWrapper | ||||||
|  | @ -956,6 +950,8 @@ provider's API. | ||||||
| > config = {"temperature": 0.0} | > config = {"temperature": 0.0} | ||||||
| > ``` | > ``` | ||||||
| 
 | 
 | ||||||
|  | Currently, these models are provided as part of the core library: | ||||||
|  | 
 | ||||||
| | Model                         | Provider  | Supported names                                                                          | Default name           | Default config                       | | | Model                         | Provider  | Supported names                                                                          | Default name           | Default config                       | | ||||||
| | ----------------------------- | --------- | ---------------------------------------------------------------------------------------- | ---------------------- | ------------------------------------ | | | ----------------------------- | --------- | ---------------------------------------------------------------------------------------- | ---------------------- | ------------------------------------ | | ||||||
| | `spacy.GPT-4.v1`              | OpenAI    | `["gpt-4", "gpt-4-0314", "gpt-4-32k", "gpt-4-32k-0314"]`                                 | `"gpt-4"`              | `{}`                                 | | | `spacy.GPT-4.v1`              | OpenAI    | `["gpt-4", "gpt-4-0314", "gpt-4-32k", "gpt-4-32k-0314"]`                                 | `"gpt-4"`              | `{}`                                 | | ||||||
|  | @ -1036,6 +1032,8 @@ These models all take the same parameters: | ||||||
| > name = "llama2-7b-hf" | > name = "llama2-7b-hf" | ||||||
| > ``` | > ``` | ||||||
| 
 | 
 | ||||||
|  | Currently, these models are provided as part of the core library: | ||||||
|  | 
 | ||||||
| | Model                | Provider        | Supported names                                                                                              | HF directory                           | | | Model                | Provider        | Supported names                                                                                              | HF directory                           | | ||||||
| | -------------------- | --------------- | ------------------------------------------------------------------------------------------------------------ | -------------------------------------- | | | -------------------- | --------------- | ------------------------------------------------------------------------------------------------------------ | -------------------------------------- | | ||||||
| | `spacy.Dolly.v1`     | Databricks      | `["dolly-v2-3b", "dolly-v2-7b", "dolly-v2-12b"]`                                                             | https://huggingface.co/databricks      | | | `spacy.Dolly.v1`     | Databricks      | `["dolly-v2-3b", "dolly-v2-7b", "dolly-v2-12b"]`                                                             | https://huggingface.co/databricks      | | ||||||
|  | @ -1044,8 +1042,6 @@ These models all take the same parameters: | ||||||
| | `spacy.StableLM.v1`  | Stability AI    | `["stablelm-base-alpha-3b", "stablelm-base-alpha-7b", "stablelm-tuned-alpha-3b", "stablelm-tuned-alpha-7b"]` | https://huggingface.co/stabilityai     | | | `spacy.StableLM.v1`  | Stability AI    | `["stablelm-base-alpha-3b", "stablelm-base-alpha-7b", "stablelm-tuned-alpha-3b", "stablelm-tuned-alpha-7b"]` | https://huggingface.co/stabilityai     | | ||||||
| | `spacy.OpenLLaMA.v1` | OpenLM Research | `["open_llama_3b", "open_llama_7b", "open_llama_7b_v2", "open_llama_13b"]`                                   | https://huggingface.co/openlm-research | | | `spacy.OpenLLaMA.v1` | OpenLM Research | `["open_llama_3b", "open_llama_7b", "open_llama_7b_v2", "open_llama_13b"]`                                   | https://huggingface.co/openlm-research | | ||||||
| 
 | 
 | ||||||
| See the "HF directory" for more details on each of the models. |  | ||||||
| 
 |  | ||||||
| Note that Hugging Face will download the model the first time you use it - you | Note that Hugging Face will download the model the first time you use it - you | ||||||
| can | can | ||||||
| [define the cached directory](https://huggingface.co/docs/huggingface_hub/main/en/guides/manage-cache) | [define the cached directory](https://huggingface.co/docs/huggingface_hub/main/en/guides/manage-cache) | ||||||
|  |  | ||||||
|  | @ -1299,9 +1299,9 @@ correct type. | ||||||
| 
 | 
 | ||||||
| ```python {title="functions.py",highlight="1"} | ```python {title="functions.py",highlight="1"} | ||||||
| @spacy.registry.tokenizers("bert_word_piece_tokenizer") | @spacy.registry.tokenizers("bert_word_piece_tokenizer") | ||||||
| def create_whitespace_tokenizer(vocab_file: str, lowercase: bool): | def create_bert_tokenizer(vocab_file: str, lowercase: bool): | ||||||
|     def create_tokenizer(nlp): |     def create_tokenizer(nlp): | ||||||
|         return BertWordPieceTokenizer(nlp.vocab, vocab_file, lowercase) |         return BertTokenizer(nlp.vocab, vocab_file, lowercase) | ||||||
| 
 | 
 | ||||||
|     return create_tokenizer |     return create_tokenizer | ||||||
| ``` | ``` | ||||||
|  |  | ||||||
		Loading…
	
		Reference in New Issue
	
	Block a user