OpenAI's models are now available on Amazon Web Services (AWS) for the first time, allowing users to integrate these advanced AI capabilities into their applications. This partnership aims to enhance the accessibility and scalability of OpenAI's technologies for developers and organizations.
AWS has introduced two new OpenAI models with open weights, gpt-oss-120b and gpt-oss-20b, available through Amazon Bedrock and SageMaker JumpStart. These models excel in text generation, coding, and reasoning tasks, offering developers greater control and flexibility in building AI applications. They support extensive customization and integration within AWS's ecosystem, enhancing the capabilities for various use cases.
An OpenAI-compatible API can be effectively deployed using AWS Lambda and an Application Load Balancer (ALB) to bypass the limitations of API Gateway's authentication requirements. By setting up the ALB to route traffic directly to the Lambda function, developers can maintain a seamless integration with the OpenAI Python client, ensuring a consistent API experience. This approach offers flexibility and security when exposing custom AI services.