Movatterモバイル変換


[0]ホーム

URL:


Skip to content

Navigation Menu

Sign in
Appearance settings

Search code, repositories, users, issues, pull requests...

Provide feedback

We read every piece of feedback, and take your input very seriously.

Saved searches

Use saved searches to filter your results more quickly

Sign up
Appearance settings

Commitcce1d44

Browse files
authored
Montana/docfix (#1376)
1 parent59977c3 commitcce1d44

File tree

16 files changed

+40
-41
lines changed

16 files changed

+40
-41
lines changed

‎pgml-cms/blog/introducing-the-openai-switch-kit-move-from-closed-to-open-source-ai-in-minutes.md

Lines changed: 5 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -21,18 +21,18 @@ December 1, 2023
2121

2222
###Introduction
2323

24-
Last week's whirlwind of events with OpenAI CEO and founder Sam Altman stirred up quite a buzz in the industry. The whole deal left many of us scratching our heads about where OpenAI is headed. Between the corporate drama, valid worries about privacy and transparency, and ongoing issues around model performance, censorship, and the use of marketing scare tactics; it's no wonder there's a growing sense of dissatisfaction and distrust in proprietary models. 
24+
Last week's whirlwind of events with OpenAI CEO and founder Sam Altman stirred up quite a buzz in the industry. The whole deal left many of us scratching our heads about where OpenAI is headed. Between the corporate drama, valid worries about privacy and transparency, and ongoing issues around model performance, censorship, and the use of marketing scare tactics; it's no wonder there's a growing sense of dissatisfaction and distrust in proprietary models.
2525

2626
On the bright side, the open-source realm has emerged as a potent contender, not just in reaction to OpenAI's shortcomings but as a genuine advancement in its own right. We're all about making the benefits of open-source models accessible to as many folks as possible. So, we've made switching from OpenAI to open-source as easy as possible with a drop-in replacement. It lets users specify any model they’d like in just a few lines of code. We call it the OpenAI Switch Kit. Read on to learn more about why we think you’ll like it, or just try it now and see what you think.
2727

2828
###Is switching to open-source AI right for you?
2929

3030
We think so. Open-source models have made remarkable strides, not only catching up to proprietary counterparts but also surpassing them across multiple domains. The advantages are clear:
3131

32-
***Performance & reliability:** Open-source models are increasingly comparable or superior across a wide range of tasks and performance metrics. Mistral and Llama-based models, for example, are easily faster than GPT 4. Reliability is another concern you may reconsider leaving in the hands of OpenAI. OpenAI’s API has suffered from several recent outages, and their rate limits can interrupt your app if there is a surge in usage. Open-source models enable greater control over your model’s latency, scalability and availability. Ultimately, the outcome of greater control is that your organization can produce a more dependable integration and a highly reliable production application. 
33-
***Safety & privacy:** Open-source models are the clear winner when it comes to security sensitive AI applications. There are[enormous risks](https://www.infosecurity-magazine.com/news-features/chatgpts-datascraping-scrutiny/) associated with transmitting private data to external entities such as OpenAI. By contrast, open-source models retain sensitive information within an organization's own cloud environments. The data never has to leave your premises, so the risk is bypassed altogether – it’s enterprise security by default. At PostgresML, we offer such private hosting of LLM’s in your own cloud. 
34-
***Model censorship:** A growing number of experts inside and outside of leading AI companies argue that model restrictions have gone too far. The Atlantic recently published an[article on AI’s “Spicy-Mayo Problem''](https://www.theatlantic.com/ideas/archive/2023/11/ai-safety-regulations-uncensored-models/676076/) which delves into the issues surrounding AI censorship. The titular example describes a chatbot refusing to return commands asking for a “dangerously spicy” mayo recipe. Censorship can affect baseline performance, and in the case of apps for creative work such as Sudowrite, unrestricted open-source models can actually be a key differentiating value for users. 
35-
***Flexibility & customization:** Closed-source models like GPT3.5 Turbo are fine for generalized tasks, but leave little room for customization. Fine-tuning is highly restricted. Additionally, the headwinds at OpenAI have exposed the[dangerous reality of AI vendor lock-in](https://techcrunch.com/2023/11/21/openai-dangers-vendor-lock-in/). Open-source models such as MPT-7B, Llama V2 and Mistral 7B are designed with extensive flexibility for fine tuning, so organizations can create custom specifications and optimize model performance for their unique needs. This level of customization and flexibility opens the door for advanced techniques like DPO, PPO LoRa and more. 
32+
***Performance & reliability:** Open-source models are increasingly comparable or superior across a wide range of tasks and performance metrics. Mistral and Llama-based models, for example, are easily faster than GPT 4. Reliability is another concern you may reconsider leaving in the hands of OpenAI. OpenAI’s API has suffered from several recent outages, and their rate limits can interrupt your app if there is a surge in usage. Open-source models enable greater control over your model’s latency, scalability and availability. Ultimately, the outcome of greater control is that your organization can produce a more dependable integration and a highly reliable production application.
33+
***Safety & privacy:** Open-source models are the clear winner when it comes to security sensitive AI applications. There are[enormous risks](https://www.infosecurity-magazine.com/news-features/chatgpts-datascraping-scrutiny/) associated with transmitting private data to external entities such as OpenAI. By contrast, open-source models retain sensitive information within an organization's own cloud environments. The data never has to leave your premises, so the risk is bypassed altogether – it’s enterprise security by default. At PostgresML, we offer such private hosting of LLM’s in your own cloud.
34+
***Model censorship:** A growing number of experts inside and outside of leading AI companies argue that model restrictions have gone too far. The Atlantic recently published an[article on AI’s “Spicy-Mayo Problem''](https://www.theatlantic.com/ideas/archive/2023/11/ai-safety-regulations-uncensored-models/676076/) which delves into the issues surrounding AI censorship. The titular example describes a chatbot refusing to return commands asking for a “dangerously spicy” mayo recipe. Censorship can affect baseline performance, and in the case of apps for creative work such as Sudowrite, unrestricted open-source models can actually be a key differentiating value for users.
35+
***Flexibility & customization:** Closed-source models like GPT3.5 Turbo are fine for generalized tasks, but leave little room for customization. Fine-tuning is highly restricted. Additionally, the headwinds at OpenAI have exposed the[dangerous reality of AI vendor lock-in](https://techcrunch.com/2023/11/21/openai-dangers-vendor-lock-in/). Open-source models such as MPT-7B, Llama V2 and Mistral 7B are designed with extensive flexibility for fine tuning, so organizations can create custom specifications and optimize model performance for their unique needs. This level of customization and flexibility opens the door for advanced techniques like DPO, PPO LoRa and more.
3636

3737
###Try it now
3838

‎pgml-cms/blog/postgresml-as-a-memory-backend-to-auto-gpt.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -88,7 +88,7 @@ Adding PostgresML as a memory backend to Auto-GPT is a relatively simple process
8888
POSTGRESML_TABLENAME =autogpt_text_embeddings
8989
```
9090

91-
If you are using PostgresML cloud, use the hostname and credentials from the cloud platform. 
91+
If you are using PostgresML cloud, use the hostname and credentials from the cloud platform.
9292

9393
!!! note
9494

‎pgml-cms/careers/product-manager.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -4,7 +4,7 @@ tags: [engineering]
44
---
55
#Product Manager
66

7-
PostgresML provides cloud hosted AI application databases, that bring the latest machine learning and vector capabilities to the heart of everyone’s favorite tech stack. We're looking for a Head of Growth, with a Technical Product Manager skill set to help shape the core product, inside and outside the company. 
7+
PostgresML provides cloud hosted AI application databases, that bring the latest machine learning and vector capabilities to the heart of everyone’s favorite tech stack. We're looking for a Head of Growth, with a Technical Product Manager skill set to help shape the core product, inside and outside the company.
88

99
Reach out if you want to:
1010

‎pgml-cms/docs/README.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -6,9 +6,9 @@ description: The key concepts that make up PostgresML.
66

77
PostgresML is a complete MLOps platform built on PostgreSQL.
88

9-
>_Move the models to thedatabase_,_rather than continuously moving the data to the models._
9+
>_Move the models to thedatabase, rather than continuously moving the data to the models._
1010
11-
The data for ML & AI systems is inherently larger and more dynamic than the models. It's more efficient, manageable and reliable to move the models to the database, rather than continuously moving the data to the models\_.\_ PostgresML allows you to take advantage of the fundamental relationship between data and models, by extending the database with the following capabilities and goals:
11+
The data for ML & AI systems is inherently larger and more dynamic than the models. It's more efficient, manageable and reliable to move the models to the database, rather than continuously moving the data to the models. PostgresML allows you to take advantage of the fundamental relationship between data and models, by extending the database with the following capabilities and goals:
1212

1313
***Model Serving** -_**GPU accelerated**_ inference engine for interactive applications, with no additional networking latency or reliability costs.
1414
***Model Store** - Download_**open-source**_ models including state of the art LLMs from HuggingFace, and track changes in performance between versions.

‎pgml-cms/docs/api/client-sdk/document-search.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -118,7 +118,7 @@ results = await collection.search(
118118

119119
Just like`vector_search`,`search` takes in two arguments. The first is a`JSON` object specifying the`query` and`limit` and the second is the`Pipeline`. The`query` object can have three fields:`full_text_search`,`semantic_search` and`filter`. Both`full_text_search` and`semantic_search` function similarly. They take in the text to compare against, titled`query`, an optional`boost` parameter used to boost the effectiveness of the ranking, and`semantic_search` also takes in an optional`parameters` key which specify parameters to pass to the embedding model when embedding the passed in text.
120120

121-
Lets break this query down a little bit more. We are asking for a maximum of 10 documents ranked by`full_text_search` on the`abstract` and`semantic_search` on the`abstract` and`body`. We are also filtering out all documents that do not have the key`user_id` equal to`1`. The`full_text_search` provides a score for the`abstract`, and`semantic_search` provides scores for the`abstract` and the`body`. The`boost` parameter is a multiplier applied to these scores before they are summed together and sorted by`score``DESC`. 
121+
Lets break this query down a little bit more. We are asking for a maximum of 10 documents ranked by`full_text_search` on the`abstract` and`semantic_search` on the`abstract` and`body`. We are also filtering out all documents that do not have the key`user_id` equal to`1`. The`full_text_search` provides a score for the`abstract`, and`semantic_search` provides scores for the`abstract` and the`body`. The`boost` parameter is a multiplier applied to these scores before they are summed together and sorted by`score``DESC`.
122122

123123
The`filter` is structured the same way it is when performing`vector_search` see[filtering with vector\_search](https://postgresml.org/docs/api/client-sdk/search)[](https://postgresml.org/docs/api/client-sdk/search#metadata-filtering)for more examples on filtering documents.
124124

‎pgml-cms/docs/api/client-sdk/pipelines.md

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -1,10 +1,10 @@
11
#Pipelines
22

3-
`Pipeline`s define the schema for the transformation of documents. Different`Pipeline`s can be used for different tasks. 
3+
`Pipeline`s define the schema for the transformation of documents. Different`Pipeline`s can be used for different tasks.
44

55
##Defining Schema
66

7-
New`Pipeline`s require schema. Here are a few examples of variations of schema along with common use cases. 
7+
New`Pipeline`s require schema. Here are a few examples of variations of schema along with common use cases.
88

99
For the following section we will assume we have documents that have the structure:
1010

@@ -100,7 +100,7 @@ pipeline = Pipeline(
100100
{% endtab %}
101101
{% endtabs %}
102102

103-
This`Pipeline` splits and embeds the`body` text enabling semantic search using vectors. This is a very popular`Pipeline` for RAG. 
103+
This`Pipeline` splits and embeds the`body` text enabling semantic search using vectors. This is a very popular`Pipeline` for RAG.
104104

105105
We support most every open source model on[Hugging Face](https://huggingface.co/), and OpenAI's embedding models. To use a model from OpenAI specify the`source` as`openai`, and make sure and set the environment variable`OPENAI_API_KEY`.
106106

@@ -210,7 +210,7 @@ pipeline = Pipeline("test_pipeline")
210210

211211
##Searching with Pipelines
212212

213-
There are two different forms of search that can be done after adding a`Pipeline` to a`Collection` 
213+
There are two different forms of search that can be done after adding a`Pipeline` to a`Collection`
214214

215215
*[Vector Search](https://postgresml.org/docs/api/client-sdk/search)
216216
*[Document Search](https://postgresml.org/docs/api/client-sdk/document-search)

‎pgml-cms/docs/api/client-sdk/tutorials/semantic-search.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -4,7 +4,7 @@ description: Example for Semantic Search
44

55
#Semantic Search
66

7-
This tutorial demonstrates using the`pgml` SDK to create a collection, add documents, build a pipeline for vector search, make a sample query, and archive the collection when finished. 
7+
This tutorial demonstrates using the`pgml` SDK to create a collection, add documents, build a pipeline for vector search, make a sample query, and archive the collection when finished.
88

99
[Link to full JavaScript implementation](../../../../../pgml-sdks/pgml/javascript/examples/semantic\_search.js)
1010

‎pgml-cms/docs/api/sql-extension/pgml.transform/fill-mask.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -4,7 +4,7 @@ description: Task to fill words in a sentence that are hidden
44

55
#Fill Mask
66

7-
Fill-mask refers to a task where certain words in a sentence are hidden or "masked", and the objective is to predict what words should fill in those masked positions. Such models are valuable when we want to gain statistical insights about the language used to train the model. 
7+
Fill-mask refers to a task where certain words in a sentence are hidden or "masked", and the objective is to predict what words should fill in those masked positions. Such models are valuable when we want to gain statistical insights about the language used to train the model.
88

99
```sql
1010
SELECTpgml.transform(

‎pgml-cms/docs/api/sql-extension/pgml.transform/text-to-text-generation.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1,6 +1,6 @@
11
#Text-to-Text Generation
22

3-
Text-to-text generation methods, such as T5, are neural network architectures designed to perform various natural language processing tasks, including summarization, translation, and question answering. T5 is a transformer-based architecture pre-trained on a large corpus of text data using denoising autoencoding. This pre-training process enables the model to learn general language patterns and relationships between different tasks, which can be fine-tuned for specific downstream tasks. During fine-tuning, the T5 model is trained on a task-specific dataset to learn how to perform the specific task. 
3+
Text-to-text generation methods, such as T5, are neural network architectures designed to perform various natural language processing tasks, including summarization, translation, and question answering. T5 is a transformer-based architecture pre-trained on a large corpus of text data using denoising autoencoding. This pre-training process enables the model to learn general language patterns and relationships between different tasks, which can be fine-tuned for specific downstream tasks. During fine-tuning, the T5 model is trained on a task-specific dataset to learn how to perform the specific task.
44

55
_Translation_
66

‎pgml-cms/docs/introduction/getting-started/connect-your-app.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -4,7 +4,7 @@ description: PostgresML is compatible with all standard PostgreSQL clients
44

55
#Connect your app
66

7-
You can connect to your database from any Postgres compatible client. PostgresML is intended to serve in the traditional role of an application database, along with it's extended role as an MLOps platform to make it easy to build and maintain AI applications. 
7+
You can connect to your database from any Postgres compatible client. PostgresML is intended to serve in the traditional role of an application database, along with it's extended role as an MLOps platform to make it easy to build and maintain AI applications.
88

99
##Application SDKs
1010

0 commit comments

Comments
 (0)

[8]ページ先頭

©2009-2025 Movatter.jp