What's New
New updates and improvements to Shiro AI
GPT-4 32K now available through Microsoft Azure OpenAI
New
We've added the GPT-4 32K model which is now available through Microsoft Azure OpenAI. The GPT-4 32K model is available on both our Essential and Premium plans.
Microsoft Azure OpenAI Service is now supported
New
We've just added support for Microsoft Azure OpenAI Service APIs. Shiro AI can now use hosted versions of OpenAI's cutting edge models with additional layers of trust and security beyond the standard OpenAI models.
Microsoft's integrated safety system provides protection from undesirable inputs and outputs and monitors for misuse. Microsoft Azure OpenAI Service models use built-in responsible AI and enterprise-grade Azure security, read more about Microsoft data privacy.
Your prompts (inputs) and completions (outputs), your embeddings, and your training data:
Microsoft's integrated safety system provides protection from undesirable inputs and outputs and monitors for misuse. Microsoft Azure OpenAI Service models use built-in responsible AI and enterprise-grade Azure security, read more about Microsoft data privacy.
Your prompts (inputs) and completions (outputs), your embeddings, and your training data:
- are NOT available to other customers.
- are NOT available to OpenAI.
- are NOT used to improve OpenAI models.
- are NOT used to improve any Microsoft or 3rd party products or services.
- are NOT used for automatically improving Azure OpenAI models for your use in your resource (The models are stateless, unless you explicitly fine-tune models with your training data).
- Your fine-tuned Azure OpenAI models are available exclusively for your use.
The Azure OpenAI Service is fully controlled by Microsoft; Microsoft hosts the OpenAI models in Microsoft’s Azure environment and the Service does NOT interact with any services operated by OpenAI (e.g. ChatGPT, or the OpenAI API).
OpenAI GPT-3.5 Turbo now available
New
We've updated our base models with both OpenAI and Microsoft Azure to GPT-3.5 Turbo!
4x faster data embeddings and enhanced security with pgvector
Improvement
We've just made an upgrade to store our vector embeddings in our own PostgreSQL database using pgvector. Previously we had been using a 3rd party hosted vector database with Pinecone.
This enhancement makes it 4 times faster to process the Datasets our customers upload! This also improves security since we can now store this data in our own database with our own managed database backups. We no longer have to trust Pinecone with keeping this data secure and properly backed up.
Overall the reasons we chose to bring this functionality in house were:
This enhancement makes it 4 times faster to process the Datasets our customers upload! This also improves security since we can now store this data in our own database with our own managed database backups. We no longer have to trust Pinecone with keeping this data secure and properly backed up.
Overall the reasons we chose to bring this functionality in house were:
- Postgres provides a wide range of vector-relevant features, including database backups, row-level security, client library support, and support for Object-Relational Mapping (ORM) in 18 programming languages. It also boasts complete ACID compliance and offers efficient bulk updates and deletes, with metadata updates taking just seconds.
- Consolidating all data within a single Postgres instance reduces the number of roundtrips in a production environment and allows for the convenient running of the entire developer setup on a local machine.
- Reducing the number of external databases reduces operational complexity and lowers the learning curve for managing data.
- Postgres is a battle-tested and robust database system, whereas many specialized vector databases have yet to establish a track record of reliability.