Tech News

A new kind of AI Model allows data owners to take over

New kind The main language model, developed by researchers in Allen Institute for Ai (A2), enables us to control how the training data is used even after the model is built.

The new model, called the Flexelomo, can challenge the current industry graphic partnership company from the web, books, and other sources – have small models. When the data is tilted in AI model today, it is to remove it from a slight model as an attempt to recover from the eggs from the finished cake.

Ali Farhadi, Parthadi says: “At the meeting, your information is inward or out.” When I train that data, you lose control. Also you have no way out, unless you press me to get through the training cycle of multiple dollars. “

AI2’s AVANDE-GARDE method is dividing training so that data owners can control. Those who want to enter data into the Floxolo model can do that by starting to copy the shared model in public known as “Anchor.” Then train the second model using their data, including the effect with anchor model, and offer the result back to anyone who form a third and last model.

Contributing this way means that the details itself has never been transferred. And because the model of the data owner is associated with the last, it is possible to remove information later. For example, the publisher magazine, offering the illegal history of its articles or later remove the lower trained model in that data if there is a legal dispute or if the company works how to use the Model.

“Training is the fullest science of asynchronous,” said Secon Min, researcher researching AI2 who led the work of technology. “Data owners do not have to link, and training can be done independently.”

Flexolmo Model Architecture is what is known as “experts known as” a famous design, often used to combine at the same time to combine several few models. Basic Establishment from A2 is a way to combine small principles trained independently. This is available using a new system of modeling model to keep its skills mixed with others when the last integrated model is driven.

Examining the way, FLaxomo researchers built a dataset call flexmix on the sources of related resources including books and websites. They used a flexolomo design to create a model with 37 billion models, about ten of the main source model size in Meta. They had compared their model several others. They found that any individual model passed through all the activities and announced 10 points of better benchmarks there are two other ways to combine trained models independently.

The result is a way to have your cake – and get your eggs back, too. “You can just get out of the system without a big damage and the measuring time,” said Farhadi. “It is the whole new way of thinking about training these kinds.”

Related Articles

Leave a Reply

Your email address will not be published. Required fields are marked *

Back to top button