Connect with us

Technology

The Three Biggest Advancements in AI for 2023

Published

on

The Three Biggest Advancements in AI for 2023

In many ways, the year 2023 marked the beginning of people’s understanding of artificial intelligence (AI) and its potential. That was the year governments started to take AI risk seriously and the year chatbots went viral for the first time. These advancements weren’t so much new inventions as they were concepts and technologies that were coming of age after a protracted gestation period.

However, there were also a lot of fresh inventions. These are the top three from the previous year:

Differentiation

Although the term “multimodality” may sound technical, it’s important to know that it refers to an AI system’s capacity to handle a wide variety of data types, including audio, video, images, and text.

This year marked the first time that robust multimodal AI models were made available to the general public. The first of these, GPT-4 from OpenAI, let users upload images in addition to text inputs. With its ability to “see” images, GPT-4 offers up a plethora of possibilities. For instance, you could ask it to decide what to have for dinner based on a picture of what’s in your refrigerator. OpenAI released the capability for users to communicate with ChatGPT via voice and text in September.

Announced in December, Google DeepMind’s most recent model, Gemini, is also capable of processing audio and images. In a Google launch video, the model was shown using a post-it note with a line drawing to identify a duck. In the same video, Gemini came up with an image of a pink and blue plush octopus after being shown a picture of pink and blue yarn and asked what they could make. (The promotional film gave the impression that Gemini was watching moving images and reacting to voice commands in real time. However, Google stated in a blog post on its website that the video had been trimmed for brevity and that the model was being prompted with text prompts rather than audio and still images, even though the model does have

“I think the next landmark that people will think back to, and remember, is [AI systems] going much more fully multimodal,” Google DeepMind co-founder Shane Legg said on a podcast in October. “It’s early days in this transition, and when you start really digesting a lot of video and other things like that, these systems will start having a much more grounded understanding of the world.” In an interview with TIME in November, OpenAI CEO Sam Altman said multimodality in the company’s new models would be one of the key things to watch out for next year.

Multimodality offers benefits beyond making models more practical. The models can also be trained on a wealth of new data sets, including audio, video, and images, which together contain more information about the world than text can. Many of the world’s leading AI companies hold the view that these models will become more powerful or capable as a result of this new training data. It is a step toward “artificial general intelligence,” the kind of system that can equal human intellect, producing labor that is economically valuable and leading to new scientific discoveries. This is the hope held by many AI scientists.

AI under the Constitution

How to integrate AI with human values is one of the most important unsolved issues in the field. If artificial intelligence and power surpass that of humans, these systems have the potential to unleash immense damage on our species—some even predict its extinction—unless they are somehow restrained by laws that prioritize human well-being.

The method that OpenAI employed to align ChatGPT (in order to steer clear of the racist and sexist tendencies of previous models) was successful, but it necessitated a significant amount of human labor. This method is called “reinforcement learning with human feedback,” or RLHF. If the AI’s response was beneficial, safe, and adhered to OpenAI’s list of content guidelines, human raters would evaluate it and award it the computational equivalent of a dog treat. OpenAI created a reasonably safe and efficient chatbot by rewarding the AI for good behavior and punishing it for bad behavior.

However, the RLHF process’s scalability is seriously questioned due to its heavy reliance on human labor. It costs a lot. It is susceptible to the prejudices or errors committed by certain raters. The longer the list of rules, the greater the likelihood of failure. And it doesn’t seem like it will work for AI systems that get so strong that they start doing things that are incomprehensible to humans.

Constitutional AI, which was initially introduced in a December 2022 paper by researchers at the prestigious AI lab Anthropic, aims to solve these issues by utilizing the fact that AI systems are now able to comprehend natural language. The concept is very straightforward. You start by creating a “constitution” that outlines the principles you want your AI to uphold. Subsequently, the AI is trained to grade responses according to how closely they adhere to the constitution. The model is then given incentives to produce responses that receive higher scores. Reward learning from AI feedback has replaced reinforcement learning from human feedback. The Anthropic researchers stated that “these methods make it possible to control AI behavior more precisely and with far fewer human labels.” Anthropic’s 2023 response to ChatGPT, Claude, was aligned using constitutional AI. (Among the investors in Anthropic is Salesforce, whose CEO and co-chair of TIME is Marc Benioff.)

“With constitutional AI, you’re explicitly writing down the normative premises with which your model should approach the world,” Jack Clark, Anthropic’s head of policy, told TIME in August. “Then the model is training on that.” There are still problems, like the difficulty of making sure the AI has understood both the letter and the spirit of the rules, (“you’re stacking your chips on a big, opaque AI model,” Clark says,) but the technique is a promising addition to a field where new alignment strategies are few and far between.

Naturally, Constitutional AI does not address the issue of whose values AI ought to be in line with. However, Anthropic is attempting to make that decision more accessible to all. The lab conducted an experiment in October wherein it asked a representative sample of one thousand Americans to assist in selecting rules for a chatbot. The results showed that, despite some polarization, it was still possible to draft a functional constitution based on statements that the group reached a consensus on. These kinds of experiments may pave the way for a time when the general public has far more influence over AI policy than it does now, when regulations are set by a select group of Silicon Valley executives.

Text to Video

The rapidly increasing popularity of text-to-video tools is one obvious result of the billions of dollars that have been invested in AI this year. Text-to-image technologies had just begun to take shape a year ago; today, a number of businesses are able to convert sentences into moving pictures with ever-increasing precision.

One of those businesses is Runway, an AI video startup with offices in Brooklyn that aims to enable anyone to make movies. With its most recent model, Gen-2, users can perform video-to-video editing—that is, altering an already-existing video’s style in response to a text prompt, such as transforming a picture of cereal boxes on a tabletop into a nighttime cityscape.

“Our mission is to build tools for human creativity,” Runway’s CEO Cristobal Valenzuela told TIME in May. He acknowledges that this will have an impact on jobs in the creative industries, where AI tools are quickly making some forms of technical expertise obsolete, but he believes the world on the other side is worth the upheaval. “Our vision is a world where human creativity gets amplified and enhanced, and it’s less about the craft, and the budget, and the technical specifications and knowledge that you have, and more about your ideas.” (Investors in Runway include Salesforce, where TIME co-chair and owner Marc Benioff is CEO.)

Pika AI, another startup in the text-to-video space, claims to be producing millions of new videos every week. The startup, which is headed by two Stanford dropouts, debuted in April but has already raised money valued at between $200 and $300 million, according to Forbes. Free tools like Pika, aimed more at the average user than at professional filmmakers, are attempting to change the face of user-generated content. Though text-to-video tools are computationally expensive, don’t be shocked if they start charging for access once the venture capital runs out. That could happen as soon as 2024.

Technology

Windows 11 PCs with Arm Processors now have an Official ISO for Clean Installations

Published

on

Power users occasionally prefer to start over when they acquire a new computer, so they follow the pro-gamers’ advice and reinstall Windows using a brand-new ISO image that comes straight from Microsoft and is free of bloatware and needlessly complex “driver management programs.” Up until recently, the new Snapdragon laptops’ more specialized version of Windows 11 didn’t support that.

The Windows 11 build on these new laptops is unusual because of the Arm64-based hardware, which differs from the typical x86 and x64 innards found in most laptops and desktops. Microsoft has finally released a disk image (or ISO file) for these devices after several months of waiting. To perform a direct reinstallation or make a bootable flash drive for a different device, you may now download it straight from Microsoft’s website. It is identical to the installation media utility that is currently available.

Be aware that there may be some glitches if you use this method for a fresh install. Compared to previous designs, the Snapdragon X system-on-a-chip has a lot fewer hardware variables, but because it’s so new, Windows Update might not include all the necessary components. You may need to use an Ethernet connection or the old-fashioned sneakernet to manually load drivers from another computer. You may also need to do some Googling to locate all the files you require for that.

Continue Reading

Technology

OPPO Reno 13 series will debut in China shortly, with India following in 2025

Published

on

According to reports, OPPO, a Chinese firm, is getting ready to introduce its Reno 13 series smartphones in its native nation this month. As per 91Mobiles, the OPPO Reno 13 and Reno 13 Pro models are anticipated to debut in China on November 25. The Indian launch is probably set for January 2025. The smartphone series that debuted in July of this year, the Reno 12 series, will be replaced by the Reno 13 series.

Information regarding the specifications of the new Reno 13 and Reno 13 Pro smartphones has leaked online, although the business has not yet confirmed the launch date. These are the specifics:

OPPO Reno 13 Series: Anticipations

It is anticipated that the OPPO Reno 13 Pro would have a 6.78-inch, quad-curved OLED screen with 1.5K resolution. In contrast, the slightly smaller 6.7-inch display with FHD+ resolution is found on the OPPO Reno 12 Pro. In China, the Pro model is probably going to be powered by the MediaTek Dimensity 8350 chipset, while in India, it might have a different processor. A 50MP primary camera, an 8MP ultrawide sensor, and a 50MP telephoto sensor with 3x optical zoom are anticipated to be included in the OPPO Reno 13 Pro’s photographic setup. Most likely, the front camera will include a 50MP sensor.

With a 5,900mAh battery as opposed to the 5,000mAh battery on the Reno 12 Pro, the Reno 13 Pro is anticipated to significantly increase battery capacity. Additionally, it is anticipated that the smartphone would support both 50W wireless and 80W wired charging. Additionally, an IP68/IP69 designation for water and dust protection could increase its durability.

Although the price of the smartphones in the Reno 13 series is not well known, it is anticipated to be similar to that of its predecessor. For comparison, the 12GB RAM + 256GB storage version of the OPPO Reno 12 Pro launched at Rs 36,999, while the 8GB RAM + 256GB storage version of the vanilla model cost Rs 32,999.

OPPO Reno 13 Pro: Anticipated features

  • Display: 6.78-inch OLED, quad-curved, with a refresh rate of 120 Hz and a resolution of 1.5K
  • processor: MediaTek Dimensity 8350
  • rear camera: 50MP primary, 8MP ultra-wide, and 50MP telephoto (3x zoom)
  • front camera: 50MP
  • Battery: 5,900mAh
  •  Charging: 50W wireless and 80W wired
  • IP rating: IP68/IP69; operating system: ColorOS 15 based on Android 15

Continue Reading

Technology

Apple has released Final Cut Pro 11, an AI-powered program

Published

on

Apple introduced Final Cut X thirteen years ago. Considering that the video-editing program marked its 25th birthday this April, that represents just over half of its lifetime. Some have questioned whether the corporation has discreetly withdrawn the offering due to its multiple lifetimes in the consumer software industry.

Final Cut Pro finally reaches level 11, after 13 years of waiting, and Apple is no longer playing around. On Wednesday, the program will be accessible for download. After a 90-day trial period, new users will need to pay $300 to buy Final Cut Pro 11 from the Mac App Store, while current users will receive it as a free update.

What specifically justified the much anticipated move to 11? AI is two letters. The business is using AI to power new features just weeks after releasing Apple Intelligence for iOS, iPadOS, and MacOS.

Magnetic Mask is at the top of the list because it makes it simple to crop objects and people out of videos without using a green screen.

According to Apple, “This powerful and precise automatic analysis provides additional flexibility to customize backgrounds and environments,” “Editors can also combine Magnetic Mask with color correction and video effects, allowing them to precisely control and stylize each project.”

Transcribe to Captions, which basically adds text to Final Cut’s timeline, is the second standout AI-based tool here. The company claims that its in-house large language model (LLM) powers that feature.

Apple’s problematic mixed-reality headset is the subject of this article’s other major headline. The most recent iPhones now have the capability to record Spatial Video, and Final Cut may be used to edit that footage. It is possible to add effects, color correct the video, and change the titles’ depth placement.

Apple is reportedly working on a more inexpensive variant, even though CEO Tim Cook has acknowledged that the $3,500 headgear isn’t the mainstream consumer product the company wanted. Along with the iPhone 15 Pro and all iPhone 16 models, the Vision Pro itself can record spatial video. Additionally, Canon just unveiled a new twin lens that works with R7 cameras.

Additionally, there are various time-saving features in the new Final Cut. For example, Magnetic Timeline allows you to swiftly rearrange clips while maintaining audio and video synchronization.

According to Apple, Final Cut Pro 11 was developed especially for the M-series of CPUs, which are its first-party silicon. This includes having more simultaneous 4K and 8K playback capabilities.

Apple claims that the M-series of chips, their first-party silicon, were the reason behind the creation of Final Cut Pro 11. This includes the capacity to play back several 4K and 8K ProRes video streams at once.

Final Cut Pro for iPad 2.1 is being released by Apple concurrently with the eagerly anticipated release of Pro 11. The brightness and color of the touched-based interface will be increased, and the workflow will be enhanced as well. Starting on Wednesday, current users can also obtain that for free.

Continue Reading

Trending

error: Content is protected !!