The AI Book
    Facebook Twitter Instagram
    The AI BookThe AI Book
    • Home
    • Categories
      • AI Media Processing
      • AI Language processing (NLP)
      • AI Marketing
      • AI Business Applications
    • Guides
    • Contact
    Subscribe
    Facebook Twitter Instagram
    The AI Book
    Daily AI News

    Generative AI Systems Aren’t Just Open or Closed Source

    24 May 2023No Comments4 Mins Read

    [ad_1]

    Recently, a leaked document, allegedly from Google, claimed that open-source AI will outcompete Google and OpenAI. The leak brought to the fore ongoing conversations in the AI community about how an AI system and its many components should be shared with researchers and the public. Even with the slew of recent generative AI system releases, this issue remains unresolved.

    Many people think of this as a binary question: Systems can either be open source or closed source. Open development decentralizes power so that many people can collectively work on AI systems to make sure they reflect their needs and values, as seen with BigScience’s BLOOM. While openness allows more people to contribute to AI research and development, the potential for harm and misuse—especially from malicious actors—increases with more access. Closed-source systems, like Google’s original LaMDA release, are protected from actors outside the developer organization but cannot be audited or evaluated by external researchers.

    I’ve been leading and researching generative AI system releases, including OpenAI’s GPT-2, since these systems first started to become available for widespread use, and I now focus on ethical openness considerations at Hugging Face. Doing this work, I’ve come to think of open source and closed source as the two ends of a gradient of options for releasing generative AI systems, rather than a simple either/or question.

    Illustration: Irene Solaiman

    At one extreme end of the gradient are systems that are so closed they are not known to the public. It’s hard to cite any concrete examples of these, for obvious reasons. But just one step over on the gradient, publicly announced closed systems are becoming increasingly common for new modalities, such as video generation. Because video generation is a relatively recent development, there is less research and information about the risks it presents and how best to mitigate them. When Meta announced its Make-a-Video model in September 2022, it cited concerns like the ease with which anyone could make realistic, misleading content as reasons for not sharing the model. Instead, Meta stated that it will gradually allow access to researchers.

    In the middle of the gradient are the systems casual users are most familiar with. Both ChatGPT and Midjourney, for instance, are publicly accessible hosted systems where the developer organization, OpenAI and Midjourney respectively, shares the model through a platform so the public can prompt and generate outputs. With their broad reach and a no-code interface, these systems have proved both useful and risky. While they can allow for more feedback than a closed system, because people outside the host organization can interact with the model, those outsiders have limited information and cannot robustly research the system by, for example, evaluating the training data or the model itself.

    On the other end of the gradient, a system is fully open when all components, from the training data to the code to the model itself, are fully open and accessible to everyone. Generative AI is built on open research and lessons from early systems like Google’s BERT, which was fully open. Today, the most-used fully open systems are pioneered by organizations focused on democratization and transparency. Initiatives hosted by Hugging Face (to which I contribute)—like BigScience and BigCode, co-led with ServiceNow—and by decentralized collectives like EleutherAI are now popular case studies for building open systems to include many languages and peoples worldwide.

    There is no definitively safe release method or standardized set of release norms. Neither is there any established body for setting standards. Early generative AI systems like ELMo and BERT were largely open until GPT-2’s staged release in 2019, which sparked new discussions about responsibly deploying increasingly powerful systems, such as what the release or publication obligations ought to be. Since then, systems across modalities, especially from large organizations, have shifted toward closedness, raising concern about the concentration of power in the high-resource organizations capable of developing and deploying these systems.

    [ad_2]

    Source link

    Previous ArticleAdobe integrates generative AI directly into Photoshop with new Firefly capabilities
    Next Article Generative AI Podcasts Are Here. Prepare to Be Bored
    The AI Book

    Related Posts

    Daily AI News

    Adobe Previews New GenAI Tools for Video Workflows

    16 April 2024
    Daily AI News

    Exciting Updates From Stanford HAI’s Seventh Annual AI Index Report

    15 April 2024
    Daily AI News

    8 Reasons to Make the Switch

    15 April 2024
    Add A Comment

    Leave A Reply Cancel Reply

    • Privacy Policy
    • Terms and Conditions
    • About Us
    • Contact Form
    © 2025 The AI Book.

    Type above and press Enter to search. Press Esc to cancel.