“Some of the toughest questions and comments toward Mr. Altman came from Dr. Marcus, who noted OpenAI hasn’t been transparent about the data its uses to develop its systems. He expressed doubt in Mr. Altman’s prediction that new jobs will replace those killed off by A.I. “We have unprecedented opportunities here but we are also facing a perfect storm of corporate irresponsibility, widespread deployment, lack of adequate regulation and inherent unreliability,” Dr. Marcus said. Tech companies have argued that Congress should be careful with any broad rules that lump different kinds of A.I. together. In Tuesday’s hearing, Ms. Montgomery of IBM called for an A.I. law that is similar to Europe’s proposed regulations, which outlines various levels of risk. She called for rules that focus on specific uses, not regulating the technology itself.”
“Un sourire sur les lèvres, Sam Altman a dû rentrer très satisfait de son escapade à Washington. On imagine le directeur d’OpenAI, éditeur de ChatGPT, taper dans les mains de ses collaborateurs de retour à San Francisco. Mission accomplie, sans doute au-delà de ses espérances, après avoir passé mardi soir trois heures devant une commission parlementaire. Non seulement Sam Altman a charmé les sénateurs. Mais en plus, il s’est posé comme le noble défenseur d’une régulation de l’intelligence artificielle (IA). Un sommet d’hypocrisie, à notre sens. La partie a été beaucoup trop facile pour le directeur d’OpenAI. On a été loin, très loin des interrogatoires musclés de directeurs d’autres géants de la tech, questionnés ces dernières années sur les effets nocifs des réseaux sociaux ou sur des fuites de données. Non, Sam Altman a pu tranquillement présenter les avantages et inconvénients de l’IA et une nouvelle fois appeler à sa régulation.”
“At the beginning of March the open source community got their hands on their first really capable foundation model, as Meta’s LLaMA was leaked to the public. It had no instruction or conversation tuning, and no RLHF. Nonetheless, the community immediately understood the significance of what they had been given. A tremendous outpouring of innovation followed, with just days between major developments (see The Timeline for the full breakdown). Here we are, barely a month later, and there are variants with instruction tuning, quantization, quality improvements, human evals, multimodality, RLHF, etc. etc. many of which build on each other. Most importantly, they have solved the scaling problem to the extent that anyone can tinker. Many of the new ideas are from ordinary people. The barrier to entry for training and experimentation has dropped from the total output of a major research organization to one person, an evening, and a beefy laptop.”
“Amazon Bedrock provides you the flexibility to choose from a wide range of FMs built by leading AI startups and Amazon so you can find the model that is best suited for what you are trying to get done. With Bedrock’s serverless experience, you can get started quickly, privately customize FMs with your own data, and easily integrate and deploy them into your applications using the AWS tools and capabilities you are familiar with”
“Since ChatGPT launched three months ago it’s taken the world by storm. People are using it to create poems, essays for school, and song lyrics. It’s also making inroads in the workplace. According to data from Cyberhaven’s product, as of March 21, 8.2% of employees have used ChatGPT in the workplace and 6.5% have pasted company data into it since it launched. Some knowledge workers say that using the tool makes them 10 times more productive. But companies like JP Morgan and Verizon are blocking access to ChatGPT over concerns about confidential data.”
“Contemporary AI systems are now becoming human-competitive at general tasks, and we must ask ourselves: Should we let machines flood our information channels with propaganda and untruth? Should we automate away all the jobs, including the fulfilling ones? Should we develop nonhuman minds that might eventually outnumber, outsmart, obsolete and replace us? Should we risk loss of control of our civilization? Such decisions must not be delegated to unelected tech leaders. Powerful AI systems should be developed only once we are confident that their effects will be positive and their risks will be manageable. This confidence must be well justified and increase with the magnitude of a system’s potential effects. OpenAI’s recent statement regarding artificial general intelligence, states that « At some point, it may be important to get independent review before starting to train future systems, and for the most advanced efforts to agree to limit the rate of growth of compute used for creating new models. » We agree. That point is now.”
“We’ve created GPT-4, the latest milestone in OpenAI’s effort in scaling up deep learning. GPT-4 is a large multimodal model (accepting image and text inputs, emitting text outputs) that, while less capable than humans in many real-world scenarios, exhibits human-level performance on various professional and academic benchmarks.”
“Why are –seeds useful in MJ? Using the same prompt + seed in MJ v4 (the current default) will produce identical images. This is VERY useful when building a prompt, b/c it lets you visualize the impact of any addition/change you make (like here, with lighting)”
“But an AI bot cannot always distinguish between helpful and hateful content. According to George Washington University’s Ding, after ChatGPT was trained by digesting the 175 billion parameters that inform it, parent company OpenAI still needed to employ several dozen human contractors to teach it not to regurgitate racist and misogynist speech or to give instructions on how to do things like build a bomb. This human-trained version, called InstructGPT, is the framework behind the chat bot. No similar effort has been announced for Baidu’s Ernie Bot or any of the other Chinese projects in the works, Ding said. Even with a robust content management team in place at Baidu, it may not be enough. Zhao, the former Baidu employee, said the company originally dedicated just a handful of engineers to the development of its AI framework. “Baidu’s AI research was slowed by a lack of commitment in a risk-ridden field that promised little return in the short term,” she said.”
« While the prospect of ChatGPT-based cheating has alarmed teachers and the academic profession, Matt Glanville, the IB’s head of assessment principles and practice, said the chatbot should be embraced as “an extraordinary opportunity”. However, Glanville told the Times, the responses must be treated as any other source in essays.
“The clear line between using ChatGPT and providing original work is exactly the same as using ideas taken from other people or the internet. As with any quote or material adapted from another source, it must be credited in the body of the text and appropriately referenced in the bibliography,” he said.
The IB is taken by thousands of children every year in the UK at more than 120 schools. Glanville said essay writing would feature less prominently in the qualifications process in the future because of the rise of chatbot technology.
“Essay writing is, however, being profoundly challenged by the rise of new technology and there’s no doubt that it will have much less prominence in the future.”
He added: “When AI can essentially write an essay at the touch of a button, we need our pupils to master different skills, such as understanding if the essay is any good or if it has missed context, has used biased data or if it is lacking in creativity. These will be far more important skills than writing an essay, so the assessment tasks we set will need to reflect this.” »