Tenyx Breakthrough AI Fine-tuning Technology Surpasses GPT-4

2024-05-08

According to reports, Itamar Arel, the founder and CEO of AI startup Tenyx, has recently made a significant breakthrough in the field of natural language processing. Tenyx successfully fine-tuned Meta's open-source Llama-3 language model (now renamed Tenyx-70B), surpassing OpenAI's GPT-4 in certain areas, marking the first time an open-source model has surpassed proprietary technology.


"We have developed a fine-tuning technique that allows us to take a base model and optimize or train it to surpass its original training scope," explained Arel. "We are very excited about this technology because it allows us to leverage the redundancy in large models to achieve potentially better continuous learning or incremental learning."


Overcoming "catastrophic forgetting"

Tenyx's new fine-tuning method addresses the issue of "catastrophic forgetting," where a model may forget previously learned knowledge when exposed to new data. By selectively updating only a small portion of the model's parameters, Tenyx is able to efficiently train the model to learn new information without affecting its existing capabilities.

"If you end up changing only about 5% of the model's parameters while keeping the rest unchanged, you can update more aggressively without the risk of distorting other things," said Arel. This selective parameter updating method also enables Tenyx to achieve fast training times, fine-tuning the Llama-3 model with 700 billion parameters in just 15 hours using only 100 GPUs.


Commitment to open-source AI

Tenyx's commitment to open-source AI is reflected in their decision to release their fine-tuned model, Tenyx-70B, under the same license agreement as the original Llama-3. "We strongly support open-source models," said Arel. "The more progress shared within the community, the more amazing applications can be created, which is better for everyone."

Tenyx's post-training optimization technology has a wide range of potential applications, from creating highly specialized chatbots for specific industries to enabling deployed models to undergo more frequent incremental updates, keeping them up-to-date between major releases.

Reshaping the AI landscape

Tenyx's breakthrough has profound implications, potentially balancing the competitive landscape by providing access to state-of-the-art language models for businesses and researchers, while avoiding the high costs and limitations associated with proprietary products. This development may also inspire further innovation within the open-source community, as other developers seek to build upon Tenyx's success.

"It makes you wonder, what does this mean for the industry? What does it mean for the leading OpenAI globally?" pondered Arel. As the AI arms race continues to heat up, Tenyx's achievements in fine-tuning open-source models could reshape the AI industry and change the way businesses approach natural language processing tasks.

Although Tenyx's optimized Llama-3 model inherits some limitations from the base model, such as occasional unreasonable or unfounded responses, its performance improvement is significant. Arel specifically noted impressive progress in mathematical and reasoning tasks, with an accuracy rate of nearly 96% compared to the base model's 85%.

As Tenyx ushers in a new era of innovation for open-source AI, the impact of its breakthrough on the AI ecosystem remains to be seen. However, one thing is certain: Tenyx has proven that open-source models can compete with and even surpass proprietary models, laying a solid foundation for a more open and collaborative future in the field of artificial intelligence.