OpenAI has introduced a fine-tuning option for GPT-3.5 Turbo, allowing artificial intelligence (AI) developers to improve performance on specific tasks using dedicated data. However, developers have both criticized and expressed enthusiasm for the development.
OpenAI clarified that developers can tailor the capabilities of GPT-3.5 Turbo to their needs through fine-tuning. Using a data set derived from the client’s business operations, a developer could, for instance, fine-tune GPT-3.5 Turbo to generate customized code or expertly summarize German legal documents.
The recent announcement has prompted developers to respond with caution. According to a comment attributed to X user Joshua Segeren, adding fine-tuning to GPT-3.5 Turbo is intriguing but not a comprehensive remedy.
According to his observations, enhancing prompts, employing vector databases for semantic queries, or transitioning to GPT-4 typically yields superior results to custom training. In addition, there are additional factors to consider, such as installation and ongoing maintenance expenses.
The base GPT-3.5 Turbo models begin at $0.0004 per one thousand tokens (the basic units processed by extensive language models). However, due to fine-tuning, the refined versions cost $0.012 per 1,000 input tokens and $0.016 per 1,000 output tokens. In addition, an initial training fee based on data volume is charged.
This feature is essential for businesses and developers who wish to create personalized user interactions. For example, companies can fine-tune the model to align with their brand’s voice, ensuring that the chatbot has a consistent personality and tone that complements its identity.
To guarantee responsible use of the fine-tuning facility, their moderation API and the GPT-4-powered moderation system review the training data used for fine-tuning. This is performed to preserve the security attributes of the default model throughout the procedure for fine-tuning.
The system identifies and removes potentially hazardous training data, ensuring the refined output conforms to OpenAI’s established security standards. This also indicates that OpenAI has some control over the data that users input into its models.