close
  • chevron_right

    1960s chatbot ELIZA beat OpenAI’s GPT-3.5 in a recent Turing test study

    news.movim.eu / ArsTechnica · Friday, 1 December - 21:27 · 1 minute

An illustration of a man and a robot sitting in boxes, talking.

Enlarge / An artist's impression of a human and a robot talking. (credit: Getty Images | Benj Edwards)

In a preprint research paper titled "Does GPT-4 Pass the Turing Test?", two researchers from UC San Diego pitted OpenAI's GPT-4 AI language model against human participants, GPT-3.5, and ELIZA to see which could trick participants into thinking it was human with the greatest success. But along the way, the study, which has not been peer-reviewed, found that human participants correctly identified other humans in only 63 percent of the interactions—and that a 1960s computer program surpassed the AI model that powers the free version of ChatGPT.

Even with limitations and caveats, which we'll cover below, the paper presents a thought-provoking comparison between AI model approaches and raises further questions about using the Turing test to evaluate AI model performance.

British mathematician and computer scientist Alan Turing first conceived the Turing test as "The Imitation Game" in 1950 . Since then, it has become a famous but controversial benchmark for determining a machine's ability to imitate human conversation. In modern versions of the test, a human judge typically talks to either another human or a chatbot without knowing which is which. If the judge cannot reliably tell the chatbot from the human a certain percentage of the time, the chatbot is said to have passed the test. The threshold for passing the test is subjective, so there has never been a broad consensus on what would constitute a passing success rate.

Read 13 remaining paragraphs | Comments

  • chevron_right

    ChatGPT is one year old. Here’s how it changed the world.

    news.movim.eu / ArsTechnica · Thursday, 30 November - 17:01 · 1 minute

A toy tin robot blowing out a birthday candle.

Enlarge / An artist's interpretation of what ChatGPT might look like if embodied in the form of a robot toy blowing out a birthday candle. (credit: Aurich Lawson | Getty Images)

One year ago today, on November 30, 2022, OpenAI released ChatGPT . It's uncommon for a single tech product to create as much global impact as ChatGPT in just one year.

Imagine a computer that can talk to you. Nothing new, right? Those have been around since the 1960s . But ChatGPT, the application that first bought large language models (LLMs) to a wide audience, felt different. It could compose poetry, seemingly understand the context of your questions and your conversation, and help you solve problems. Within a few months, it became the fastest-growing consumer application of all time. And it created a frenzy.

During these 365 days, ChatGPT has broadened the public perception of AI, captured imaginations, attracted critics , and stoked existential angst. It emboldened and reoriented Microsoft, made Google dance , spurred fears of AGI taking over the world, captivated world leaders , prompted attempts at government regulation , helped add words to dictionaries , inspired conferences and copycats , led to a crisis for educators, hyper-charged automated defamation , embarrassed lawyers by hallucinating, prompted lawsuits over training data, and much more.

Read 12 remaining paragraphs | Comments

  • chevron_right

    Sam Altman officially back as OpenAI CEO: “We didn’t lose a single employee”

    news.movim.eu / ArsTechnica · Thursday, 30 November - 14:37 · 1 minute

A glowing OpenAI logo on a light blue background.

Enlarge (credit: OpenAI / Benj Edwards)

On Wednesday, OpenAI announced that Sam Altman has officially returned to the ChatGPT-maker as CEO—accompanied by Mira Murati as CTO and Greg Brockman as president—resuming their roles from before the shocking firing of Altman that threw the company into turmoil two weeks ago. Altman says the company did not lose a single employee or customer throughout the crisis.

"I have never been more excited about the future. I am extremely grateful for everyone’s hard work in an unclear and unprecedented situation, and I believe our resilience and spirit set us apart in the industry," wrote Altman in an official OpenAI news release . "I feel so, so good about our probability of success for achieving our mission."

In the statement, Altman formalized plans that have been underway since last week: ex-Salesforce co-CEO Bret Taylor and economist Larry Summers have officially begun their tenure on the "new initial" OpenAI board of directors. Quora CEO Adam D’Angelo is keeping his previous seat on the board. Also on Wednesday, previous board members Tasha McCauley and Helen Toner officially resigned . In addition, a representative from Microsoft (a key OpenAI investor) will have a non-voting observer role on the board of directors.

Read 8 remaining paragraphs | Comments

  • chevron_right

    Amazon unleashes Q, an AI assistant for the workplace

    news.movim.eu / ArsTechnica · Wednesday, 29 November - 17:13

The Amazon Q logo.

Enlarge / The Amazon Q logo. (credit: Amazon)

On Tuesday, Amazon unveiled Amazon Q , an AI chatbot similar to ChatGPT that is tailored for corporate environments. Developed by Amazon Web Services (AWS), Q is designed to assist employees with tasks like summarizing documents, managing internal support tickets, and providing policy guidance, differentiating itself from consumer-focused chatbots. It also serves as a programming assistant.

According to The New York Times , the name "Q" is a play on the word “question" and a reference to the character Q in the James Bond novels, who makes helpful tools. (And there's apparently a little bit of Q from Star Trek: The Next Generation thrown in, although hopefully the new bot won't cause mischief on that scale.)

Amazon Q's launch positions it against existing corporate AI tools like Microsoft's Copilot , Google's Duet AI , and ChatGPT Enterprise . Unlike some of its competitors, Amazon Q isn't built on a singular AI large language model (LLM). Instead, it uses a platform called Bedrock, integrating multiple AI systems, including Amazon's Titan and models from Anthropic and Meta .

Read 5 remaining paragraphs | Comments

  • chevron_right

    Microsoft offers legal protection for AI copyright infringement challenges

    news.movim.eu / ArsTechnica · Friday, 8 September - 22:40

A man in an armor helmet sitting at a desk with a protective glowing field around him.

Enlarge (credit: Getty Images / Benj Edwards )

On Thursday, Microsoft announced that it will provide legal protection for customers who are sued for copyright infringement over content generated by the company's AI systems. This new policy, called the Copilot Copyright Commitment, is an expansion of Microsoft's existing intellectual property indemnification coverage, Reuters reports .

Microsoft's announcement comes as generative AI tools like ChatGPT have raised concerns about reproducing copyrighted material without proper attribution. Microsoft has heavily invested in AI through products like GitHub Copilot and Bing Chat that can generate original code, text, and images on demand. Its AI models have gained these capabilities by scraping publicly available data off of the Internet without seeking express permission from copyright holders.

By offering legal protection, Microsoft aims to give customers confidence in deploying its AI systems without worrying about potential copyright issues. The policy covers damages and legal fees, providing customers with an added layer of protection as generative AI sees rapid adoption across the tech industry.

Read 5 remaining paragraphs | Comments

  • chevron_right

    The AI-assistant wars heat up with Claude Pro, a new ChatGPT Plus rival

    news.movim.eu / ArsTechnica · Friday, 8 September - 20:37

The Anthropic Claude logo on a purple background.

Enlarge / The Anthropic Claude logo. (credit: Anthropic / Benj Edwards)

On Thursday, AI-maker and OpenAI competitor Anthropic launched Claude Pro , a subscription-based version of its Claude.ai web-based AI assistant, which functions similarly to ChatGPT. It's available for $20/month in the US or 18 pounds/month in the UK, and it promises five-times-higher usage limits, priority access to Claude during high-traffic periods, and early access to new features as they emerge.

Like ChatGPT, Claude Pro can compose text, summarize, do analysis, solve logic puzzles, and more.

Claude.ai is what Anthropic offers as its conversational interface for its Claude 2 AI language model, similar to how ChatGPT provides an application wrapper for the underlying models GPT-3.5 and GPT-4. In February, OpenAI chose a subscription route for ChatGPT Plus , which for $20 a month also gives early access to new features, but it also unlocks access to GPT-4, which is OpenAI's most powerful language model.

Read 9 remaining paragraphs | Comments

  • chevron_right

    OpenAI admits that AI writing detectors don’t work

    news.movim.eu / ArsTechnica · Friday, 8 September - 15:42

A photo of a teacher covering his eyes.

Enlarge (credit: Getty Images )

Last week, OpenAI published tips for educators in a promotional blog post that shows how some teachers are using ChatGPT as an educational aid, along with suggested prompts to get started. In a related FAQ , they also officially admit what we already know: AI writing detectors don't work, despite frequently being used to punish students with false positives.

In a section of the FAQ titled "Do AI detectors work?", OpenAI writes , "In short, no. While some (including OpenAI) have released tools that purport to detect AI-generated content, none of these have proven to reliably distinguish between AI-generated and human-generated content."

In July, we covered in depth why AI writing detectors such as GPTZero don't work, with experts calling them "mostly snake oil." These detectors often yield false positives due to relying on unproven detection metrics. Ultimately, there is nothing special about AI-written text that always distinguishes it from human-written, and detectors can be defeated by rephrasing. That same month, OpenAI discontinued its AI Classifier, which was an experimental tool designed to detect AI-written text. It had an abysmal 26 percent accuracy rate.

Read 5 remaining paragraphs | Comments

  • chevron_right

    OpenAI to host its first developer conference on November 6 in San Francisco

    news.movim.eu / ArsTechnica · Thursday, 7 September - 15:16

A vintage tin toy robot collection belonging to Anthea Knowles, UK, 16th May 1980.

Enlarge (credit: Getty Images )

On Wednesday, OpenAI announced that it will host its first-ever developer conference, OpenAI DevDay, on November 6, 2023, in San Francisco. The one-day event hopes to bring together hundreds of developers to preview new tools and discuss ideas with OpenAI's technical staff.

Launched in November, ChatGPT has driven intense interest in generative AI around the world, including tech investments, talk of regulations, a GPU hardware boom , and the emergence of competitors. OpenAI says in a blog post that since launching its first API in 2020, over 2 million developers now use its models like GPT-3, GPT-4 , DALL-E , and Whisper for a variety of applications, "from integrating smart assistants into existing applications to building entirely new applications and services that weren't possible before."

While OpenAI's DevDay event will mostly take place in person, the keynote and potentially some parts of the conference will be streamed online. "The one-day event will bring hundreds of developers from around the world together with the team at OpenAI to preview new tools and exchange ideas," writes OpenAI. "In-person attendees will also be able to join breakout sessions led by members of OpenAI’s technical staff."

Read 2 remaining paragraphs | Comments