From 9cc469f72d6ec7e8d4191bcdb9c55e17791a01c2 Mon Sep 17 00:00:00 2001 From: MKhalusova Date: Mon, 19 Feb 2024 10:08:26 -0500 Subject: [PATCH 1/3] updates to Derek's notebook --- notebooks/en/_toctree.yml | 4 +-- ...ic_embedding_tei_inference_endpoints.ipynb | 27 +++---------------- 2 files changed, 6 insertions(+), 25 deletions(-) diff --git a/notebooks/en/_toctree.yml b/notebooks/en/_toctree.yml index ccf8078b..587c2e06 100644 --- a/notebooks/en/_toctree.yml +++ b/notebooks/en/_toctree.yml @@ -1,9 +1,9 @@ - title: Open-Source AI Cookbook sections: - - local: automatic_embedding_tei_inference_endpoints - title: Automatic Embeddings with TEI through Inference Endpoints - local: index title: Open-Source AI Cookbook + - local: automatic_embedding_tei_inference_endpoints + title: Automatic Embeddings with TEI through Inference Endpoints - local: faiss_with_hf_datasets_and_clip title: Embedding multimodal data for similarity search - local: fine_tuning_code_llm_on_single_gpu diff --git a/notebooks/en/automatic_embedding_tei_inference_endpoints.ipynb b/notebooks/en/automatic_embedding_tei_inference_endpoints.ipynb index f9c0ee66..f9d609c4 100644 --- a/notebooks/en/automatic_embedding_tei_inference_endpoints.ipynb +++ b/notebooks/en/automatic_embedding_tei_inference_endpoints.ipynb @@ -6,9 +6,8 @@ "metadata": {}, "source": [ "# How to use Inference Endpoints to Embed Documents\n", - "[Derek Thomas](https://huggingface.co/derek-thomas)\n", - "[LinkedIn](https://www.linkedin.com/in/dthomas/)\n", - "[Blog](https://datavistics.github.io)\n", + "\n", + "_Authored by: [Derek Thomas](https://huggingface.co/derek-thomas)\n", "\n", "## Goal\n", "I have a dataset I want to embed for semantic search (or QA, or RAG), I want the easiest way to do embed this and put it in a new dataset.\n", @@ -37,22 +36,12 @@ }, { "cell_type": "code", - "execution_count": 1, + "execution_count": null, "id": "35386f72-32cb-49fa-a108-3aa504e20429", "metadata": { "tags": [] }, - "outputs": [ - { - "name": "stdout", - "output_type": "stream", - "text": [ - "\n", - "\u001B[1m[\u001B[0m\u001B[34;49mnotice\u001B[0m\u001B[1;39;49m]\u001B[0m\u001B[39;49m A new release of pip is available: \u001B[0m\u001B[31;49m23.0.1\u001B[0m\u001B[39;49m -> \u001B[0m\u001B[32;49m23.3.1\u001B[0m\n", - "\u001B[1m[\u001B[0m\u001B[34;49mnotice\u001B[0m\u001B[1;39;49m]\u001B[0m\u001B[39;49m To update, run: \u001B[0m\u001B[32;49mpip install --upgrade pip\u001B[0m\n" - ] - } - ], + "outputs": [], "source": [ "!pip install -q aiohttp==3.8.3 datasets==2.14.6 pandas==1.5.3 requests==2.31.0 tqdm==4.66.1 huggingface-hub>=0.20" ] @@ -806,14 +795,6 @@ "else:\n", " print('Delete Endpoint in manually') " ] - }, - { - "cell_type": "code", - "execution_count": null, - "id": "5db1b1c3-16c3-403a-9472-a97e730826d5", - "metadata": {}, - "outputs": [], - "source": [] } ], "metadata": { From 38430e3b8da63065e942d5ba48fdd0973e1753d9 Mon Sep 17 00:00:00 2001 From: MKhalusova Date: Mon, 19 Feb 2024 10:21:34 -0500 Subject: [PATCH 2/3] fixed a typo --- notebooks/en/automatic_embedding_tei_inference_endpoints.ipynb | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/notebooks/en/automatic_embedding_tei_inference_endpoints.ipynb b/notebooks/en/automatic_embedding_tei_inference_endpoints.ipynb index f9d609c4..3aefd95b 100644 --- a/notebooks/en/automatic_embedding_tei_inference_endpoints.ipynb +++ b/notebooks/en/automatic_embedding_tei_inference_endpoints.ipynb @@ -7,7 +7,7 @@ "source": [ "# How to use Inference Endpoints to Embed Documents\n", "\n", - "_Authored by: [Derek Thomas](https://huggingface.co/derek-thomas)\n", + "_Authored by: [Derek Thomas](https://huggingface.co/derek-thomas_\n", "\n", "## Goal\n", "I have a dataset I want to embed for semantic search (or QA, or RAG), I want the easiest way to do embed this and put it in a new dataset.\n", From b445e26312be2e172365d0fdcfcc5e702ff0ceb8 Mon Sep 17 00:00:00 2001 From: MKhalusova Date: Mon, 19 Feb 2024 10:21:54 -0500 Subject: [PATCH 3/3] fixed another typo --- notebooks/en/automatic_embedding_tei_inference_endpoints.ipynb | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/notebooks/en/automatic_embedding_tei_inference_endpoints.ipynb b/notebooks/en/automatic_embedding_tei_inference_endpoints.ipynb index 3aefd95b..6ec9a2c9 100644 --- a/notebooks/en/automatic_embedding_tei_inference_endpoints.ipynb +++ b/notebooks/en/automatic_embedding_tei_inference_endpoints.ipynb @@ -7,7 +7,7 @@ "source": [ "# How to use Inference Endpoints to Embed Documents\n", "\n", - "_Authored by: [Derek Thomas](https://huggingface.co/derek-thomas_\n", + "_Authored by: [Derek Thomas](https://huggingface.co/derek-thomas)_\n", "\n", "## Goal\n", "I have a dataset I want to embed for semantic search (or QA, or RAG), I want the easiest way to do embed this and put it in a new dataset.\n",