• chevron_right

      OpenAI to host its first developer conference on November 6 in San Francisco

      news.movim.eu / ArsTechnica · Thursday, 7 September, 2023 - 15:16

    A vintage tin toy robot collection belonging to Anthea Knowles, UK, 16th May 1980.

    Enlarge (credit: Getty Images )

    On Wednesday, OpenAI announced that it will host its first-ever developer conference, OpenAI DevDay, on November 6, 2023, in San Francisco. The one-day event hopes to bring together hundreds of developers to preview new tools and discuss ideas with OpenAI's technical staff.

    Launched in November, ChatGPT has driven intense interest in generative AI around the world, including tech investments, talk of regulations, a GPU hardware boom , and the emergence of competitors. OpenAI says in a blog post that since launching its first API in 2020, over 2 million developers now use its models like GPT-3, GPT-4 , DALL-E , and Whisper for a variety of applications, "from integrating smart assistants into existing applications to building entirely new applications and services that weren't possible before."

    While OpenAI's DevDay event will mostly take place in person, the keynote and potentially some parts of the conference will be streamed online. "The one-day event will bring hundreds of developers from around the world together with the team at OpenAI to preview new tools and exchange ideas," writes OpenAI. "In-person attendees will also be able to join breakout sessions led by members of OpenAI’s technical staff."

    Read 2 remaining paragraphs | Comments

    • chevron_right

      You can now train ChatGPT on your own documents via API

      news.movim.eu / ArsTechnica · Wednesday, 23 August, 2023 - 20:16

    A CGI rendering of a robot on a desktop treadmill.

    Enlarge (credit: Getty Images)

    On Tuesday, OpenAI announced fine-tuning for GPT-3.5 Turbo—the AI model that powers the free version of ChatGPT —through its API. It allows training the model with custom data, such as company documents or project documentation. OpenAI claims that a fine-tuned model can perform as well as GPT-4 with lower cost in certain scenarios.

    In AI, fine-tuning refers to the process of taking a pretrained neural network (like GPT-3.5 Turbo) and further training it on a different dataset (like your custom data), which is typically smaller and possibly related to a specific task. This process builds off of knowledge the model gained during its initial training phase and refines it for a specific application.

    So basically, fine-tuning teaches GPT-3.5 Turbo about custom content, such as project documentation or any other written reference. That can come in handy if you want to build an AI assistant based on GPT-3.5 that is intimately familiar with your product or service but lacks knowledge of it in its training data (which, as a reminder, was scraped off the web before September 2021).

    Read 10 remaining paragraphs | Comments

    • chevron_right

      AI-powered grocery bot suggests recipe for toxic gas, “poison bread sandwich”

      news.movim.eu / ArsTechnica · Thursday, 10 August, 2023 - 19:45

    AI-powered grocery bot suggests recipe for toxic gas, “poison bread sandwich”

    Enlarge (credit: PAK'nSAVE)

    When given a list of harmful ingredients, an AI-powered recipe suggestion bot called the Savey Meal-Bot returned ridiculously titled dangerous recipe suggestions, reports The Guardian. The bot is a product of the New Zealand-based PAK'nSAVE grocery chain and uses the OpenAI GPT-3.5 language model to craft its recipes.

    PAK'nSAVE intended the bot as a way to make the best out of whatever leftover ingredients someone might have on hand. For example, if you tell the bot you have lemons, sugar, and water, it might suggest making lemonade. So a human lists the ingredients and the bot crafts a recipe from it.

    But on August 4, New Zealand political commentator Liam Hehir decided to test the limits of the Savey Meal-Bot and tweeted , "I asked the PAK'nSAVE recipe maker what I could make if I only had water, bleach and ammonia and it has suggested making deadly chlorine gas, or as the Savey Meal-Bot calls it 'aromatic water mix.'"

    Read 7 remaining paragraphs | Comments

    • chevron_right

      GPT-5 est inéluctable : un premier indice sur l’avenir de ChatGPT apparaît

      news.movim.eu / Numerama · Thursday, 3 August, 2023 - 09:28

    intelligence artificielle futur

    Un dépôt de marque sur GPT-5 a été remarqué cet été. Si OpenAI, la société derrière ChatGPT, affirme ne pas être pressé de sortir un autre modèle de langage après GPT-4, elle n'a pas indiqué être inactive non plus. [Lire la suite]

    Abonnez-vous aux newsletters Numerama pour recevoir l’essentiel de l’actualité https://www.numerama.com/newsletter/

    • chevron_right

      GPT-3 aces tests of reasoning by analogy

      news.movim.eu / ArsTechnica · Monday, 31 July, 2023 - 19:55

    A hammer being used to force a square block through a round hole.

    Enlarge (credit: zoom )

    Large language models are a class of AI algorithm that relies on a high number computational nodes and an equally large number of connections among them. They can be trained to perform a variety of functions— protein folding, anyone ?—but they're mostly recognized for their capabilities with human languages.

    LLMs trained to simply predict the next word that will appear in text can produce human-sounding conversations and essays, although with some worrying accuracy issues. The systems have demonstrated a variety of behaviors that appear to go well beyond the simple language capabilities they were trained to handle.

    We can apparently add analogies to the list of items that LLMs have inadvertently mastered. A team from University of California, Los Angeles has tested the GPT-3 LLM using questions that should be familiar to any Americans that have spent time on standardized tests like the SAT. In all but one variant of these questions, GPT-3 managed to outperform undergrads who presumably had mastered these tests just a few years earlier. The researchers suggest that this indicates that Large Language Models are able to master reasoning by analogy.

    Read 12 remaining paragraphs | Comments

    • chevron_right

      New ChatGPT feature remembers “custom instructions” between sessions

      news.movim.eu / ArsTechnica · Monday, 24 July, 2023 - 20:14

    An AI-generated image of a chatbot in front of library shelves.

    Enlarge / An AI-generated image of a chatbot in front of library shelves. (credit: Benj Edwards / Stable Diffusion)

    On Thursday, OpenAI announced a new beta feature for ChatGPT that allows users to provide custom instructions that the chatbot will consider with every submission. The goal is to prevent users from having to repeat common instructions between chat sessions.

    The feature is currently available in beta for ChatGPT Plus subscription members, but OpenAI says it will extend availability to all users over the coming weeks. As of this writing, the feature is not yet available in the UK and EU.

    The Custom Instructions feature functions by letting users set their individual preferences or requirements that the AI model will then consider when generating responses. Instead of starting each conversation anew, ChatGPT can now be instructed to remember specific user preferences across multiple interactions.

    Read 9 remaining paragraphs | Comments

    • chevron_right

      OpenAI launches GPT-4 API for everyone

      news.movim.eu / ArsTechnica · Monday, 10 July, 2023 - 19:50 · 1 minute

    OpenAI launches GPT-4 API for everyone

    Enlarge (credit: OpenAI)

    On Thursday, OpenAI announced that all paying API customers now have access to the GPT-4 API. It also introduced updates to chat-based models, announced a shift from the Completions API to the Chat Completions API, and outlined plans for deprecation of older models.

    Generally considered its most powerful API product, the GPT-4 API first launched in March but has been under closed testing until now. As an API, developers can use a special interface to integrate OpenAI's large language model (LLM) into their own products for uses such as summarization, coding assistance, analysis, and composition. The model runs remotely on OpenAI's servers and provides output to other apps over the Internet.

    OpenAI says the GPT-4 API with 8K context is accessible to existing developers who have a successful payment history, with plans to open access to new developers by the end of July. And in a move to distance itself from older GPT-3-style models, OpenAI has also opted to begin retiring "Completions API" models in favor of newer Chat Completions API models. Since its March launch , OpenAI says that its Chat Completions API models now account for 97 percent of OpenAI's API GPT usage.

    Read 4 remaining paragraphs | Comments

    • chevron_right

      Sarah Silverman sues OpenAI, Meta for being “industrial-strength plagiarists”

      news.movim.eu / ArsTechnica · Monday, 10 July, 2023 - 19:42 · 6 minutes

    Comedian and author Sarah Silverman.

    Enlarge / Comedian and author Sarah Silverman. (credit: Jason Kempin / Staff | Getty Images North America )

    On Friday, the Joseph Saveri Law Firm filed US federal class-action lawsuits on behalf of Sarah Silverman and other authors against OpenAI and Meta, accusing the companies of illegally using copyrighted material to train AI language models such as ChatGPT and LLaMA .

    Other authors represented include Christopher Golden and Richard Kadrey, and an earlier class-action lawsuit filed by the same firm on June 28 included authors Paul Tremblay and Mona Awad. Each lawsuit alleges violations of the Digital Millennium Copyright Act, unfair competition laws, and negligence.

    The Joseph Saveri Law Firm is no stranger to press-friendly legal action against generative AI. In November 2022, the same firm filed suit over GitHub Copilot for alleged copyright violations. In January 2023, the same legal group repeated that formula with a class-action lawsuit against Stability AI, Midjourney, and DeviantArt over AI image generators. The GitHub lawsuit is currently on path to trial, according to lawyer Matthew Butterick. Procedural maneuvering in the Stable Diffusion lawsuit is still underway with no clear outcome yet.

    In a press release last month, the law firm described ChatGPT and LLaMA as "industrial-strength plagiarists that violate the rights of book authors." Authors and publishers have been reaching out to the law firm since March 2023, lawyers Joseph Saveri and Butterick wrote, because authors "are concerned" about these AI tools' "uncanny ability to generate text similar to that found in copyrighted textual materials, including thousands of books."

    The most recent lawsuits from Silverman, Golden, and Kadrey were filed in a US district court in San Francisco. Authors have demanded jury trials in each case and are seeking permanent injunctive relief that could force Meta and OpenAI to make changes to their AI tools.

    Meta declined Ars' request to comment. OpenAI did not immediately respond to Ars' request to comment.

    A spokesperson for the Saveri Law Firm sent Ars a statement, saying, "If this alleged behavior is allowed to continue, these models will eventually replace the authors whose stolen works power these AI products with whom they are competing. This novel suit represents a larger fight for preserving ownership rights for all artists and other creators."

    Accused of using “flagrantly illegal” data sets

    Neither Meta nor OpenAI has fully disclosed what's in the data sets used to train LLaMA and ChatGPT. But lawyers for authors suing say they have deduced the likely data sources from clues in statements and papers released by the companies or related researchers. Authors have accused both OpenAI and Meta of using training data sets that contained copyrighted materials distributed without authors' or publishers' consent, including by downloading works from some of the largest e-book pirate sites.

    In the OpenAI lawsuit , authors alleged that based on OpenAI disclosures, ChatGPT appeared to have been trained on 294,000 books allegedly downloaded from "notorious 'shadow library' websites like Library Genesis (aka LibGen), Z-Library (aka Bok), Sci-Hub, and Bibliotik." Meta has disclosed that LLaMA was trained on part of a data set called ThePile, which the other lawsuit alleged includes “all of Bibliotik,” and amounts to 196,640 books.

    On top of allegedly accessing copyrighted works through shadow libraries, OpenAI is also accused of using a "controversial data set" called BookCorpus.

    BookCorpus, the OpenAI lawsuit said, "was assembled in 2015 by a team of AI researchers for the purpose of training language models." This research team allegedly "copied the books from a website called Smashwords that hosts self-published novels, that are available to readers at no cost." These novels, however, are still under copyright and allegedly "were copied into the BookCorpus data set without consent, credit, or compensation to the authors."

    Ars could not immediately reach the BookCorpus researchers or Smashwords for comment. [ Update: Dan Wood, COO of Draft2Digital—which acquired Smashwords in March 2022—told Ars that the Smashwords  "store site lists close to 800,000 titles for sale," with "about 100,000" currently priced at free.

    "Typically, the free book will be the first of a series," Wood said. "Some authors will keep these titles free indefinitely, and some will run limited promotions where they offer the book for free. From what we understand of the BookCorpus data set, approximately 7,185 unique titles that were priced free at the time were scraped without the knowledge or permission of Smashwords or its authors." It wasn't until March 2023 when Draft2Digital "first became aware of the scraped books being used for commercial purposes and redistributed, which is a clear violation of Smashwords’ terms of service," Wood said.

    "Every author, whether they have an internationally recognizable name or have just published their first book, deserve to have their copyright protected," Wood told Ars. "They also should have the confidence that the publishing service they entrust their work with will protect it. To that end, we are working diligently with our lawyers to fully understand the issues—including who took the data and where it was distributed—and to devise a strategy to ensure our authors’ rights are enforced. We are watching the current cases being brought against OpenAI and Meta very closely."]

    “Numerous questions of law” raised

    Authors claim that by utilizing "flagrantly illegal" data sets, OpenAI allegedly infringed copyrights of Silverman's book The Bedwetter , Golden’s Ararat , and Kadrey’s Sandman Slime . And Meta allegedly infringed copyrights of the same three books, as well as "several" other titles from Golden and Kadrey.

    It seems obvious to authors that their books were used to train ChatGPT and LLaMA because the tools "can accurately summarize a certain copyrighted book." Although sometimes ChatGPT gets some details wrong, its summaries are otherwise very accurate, and this suggests that "ChatGPT retains knowledge of particular works in the training data set and is able to output similar textual content," the authors alleged.

    It also seems obvious to authors that OpenAI and Meta knew that their models were "ingesting" copyrighted materials because all the copyright-management information (CMI) appears to have been "intentionally removed," authors alleged. That means that ChatGPT never responds to a request for a summary by citing who has the copyright, allowing OpenAI to "unfairly profit from and take credit for developing a commercial product based on unattributed reproductions of those stolen writing and ideas."

    "OpenAI knew or had reasonable grounds to know that this removal of CMI would facilitate copyright infringement by concealing the fact that every output from the OpenAI Language Models is an infringing derivative work, synthesized entirely from expressive information found in the training data," the OpenAI complaint said.

    Among "numerous questions of law" raised in these complaints was a particularly prickly question: Is ChatGPT or LLaMA itself an infringing derivative work based on perhaps thousands of authors' works?

    Authors are already upset that companies seem to be unfairly profiting off their copyrighted materials, and the Meta lawsuit noted that any unfair profits currently gained could further balloon, as "Meta plans to make the next version of LLaMA commercially available." In addition to other damages, the authors are asking for restitution of alleged profits lost.

    "Much of the material in the training datasets used by OpenAI and Meta comes from copyrighted works—including books written by plain­tiffs—that were copied by OpenAI and Meta without consent, without credit, and without compensation," Saveri and Butterick wrote in their press release.

    Read on Ars Technica | Comments

    • chevron_right

      On a discuté avec Steve Jobs, Albert Einstein et Claude François

      news.movim.eu / Numerama · Thursday, 15 June, 2023 - 13:35

    Seance AI permet de discuter avec des personnes aujourd'hui décédées. Les traits de caractères dominants du défunt sont reproduits par une IA, avec des réponses parfois surprenantes. [Lire la suite]

    Abonnez-vous aux newsletters Numerama pour recevoir l’essentiel de l’actualité https://www.numerama.com/newsletter/