Connect with us

Technology

Gen AI without the dangers

Published

on

It’s understandable that ChatGPT, Stable Diffusion, and DreamStudio-Generative AI are making headlines. The outcomes are striking and getting better geometrically. Already, search and information analysis, as well as code creation, network security, and article writing, are being revolutionized by intelligent assistants.

Gen AI will play a critical role in how businesses run and provide IT services, as well as how business users complete their tasks. There are countless options, but there are also countless dangers. Successful AI development and implementation can be a costly and risky process. Furthermore, the workloads associated with Gen AI and the large language models (LLMs) that drive it are extremely computationally demanding and energy-intensive.Dr. Sajjad Moazeni of the University of Washington estimates that training an LLM with 175 billion or more parameters requires an annual energy expenditure for 1,000 US households, though exact figures are unknown. Over 100 million generative AI questions answered daily equate to one gigawatt-hour of electricity use, or about 33,000 US households’ daily energy use.

How even hyperscalers can afford that much electricity is beyond me. It’s too expensive for the typical business. How can CIOs provide reliable, accurate AI without incurring the energy expenses and environmental impact of a small city?

Six pointers for implementing Gen AI economically and with less risk

Retraining generative AI to perform particular tasks is essential to its applicability in business settings. Expert models produced by retraining are smaller, more accurate, and require less processing power. So, in order to train their own AI models, does every business need to establish a specialized AI development team and a supercomputer? Not at all.

Here are six strategies to create and implement AI without spending a lot of money on expensive hardware or highly skilled personnel.

Start with a foundation model rather than creating the wheel.

A company might spend money creating custom models for its own use cases. But the expenditure on data scientists, HPC specialists, and supercomputing infrastructure is out of reach for all but the biggest government organizations, businesses, and hyperscalers.

Rather, begin with a foundation model that features a robust application portfolio and an active developer ecosystem. You could use an open-source model like Meta’s Llama 2, or a proprietary model like OpenAI’s ChatGPT. Hugging Face and other communities provide a vast array of open-source models and applications.

Align the model with the intended use

Models can be broadly applicable and computationally demanding, such as GPT, or more narrowly focused, like Med-BERT (an open-source LLM for medical literature). The time it takes to create a viable prototype can be shortened and months of training can be avoided by choosing the appropriate model early in the project.

However, exercise caution. Any model may exhibit biases in the data it uses to train, and generative AI models are capable of lying outright and fabricating responses. Seek models trained on clean, transparent data with well-defined governance and explicable decision-making for optimal trustworthiness.

Retrain to produce more accurate, smaller models

Retraining foundation models on particular datasets offers various advantages. The model sheds parameters it doesn’t need for the application as it gets more accurate on a smaller field. One way to trade a general skill like songwriting for the ability to assist a customer with a mortgage application would be to retrain an LLM in financial information.

With a more compact design, the new banking assistant would still be able to provide superb, extremely accurate services while operating on standard (current) hardware.

Make use of your current infrastructure

A supercomputer with 10,000 GPUs is too big for most businesses to set up. Fortunately, most practical AI training, retraining, and inference can be done without large GPU arrays.

  • Training up to 10 billion: at competitive price/performance points, contemporary CPUs with integrated AI acceleration can manage training loads in this range. For better performance and lower costs, train overnight during periods of low demand for data centers.
  • Retraining up to 10 billion models is possible with modern CPUs; no GPU is needed, and it takes only minutes.
  • With integrated CPUs, smaller models can operate on standalone edge devices, with inferencing ranging from millions to less than 20 billion. For models with less than 20 billion parameters, such as Llama 2, CPUs can respond as quickly and precisely as GPUs.

Execute inference with consideration for hardware

Applications for inference can be fine-tuned and optimized for improved performance on particular hardware configurations and features. Similar to training a model, optimizing one for a given application means striking a balance between processing efficiency, model size, and accuracy.

One way to increase inference speeds four times while maintaining accuracy is to round down a 32-bit floating point model to the nearest 8-bit fixed integer (INT8). Utilizing host accelerators such as integrated GPUs, Intel® Advanced Matrix Extensions (Intel® AMX), and Intel® Advanced Vector Extensions 512 (Intel® AVX-512), tools such as Intel® Distribution of OpenVINOTM toolkit manage optimization and build hardware-aware inference engines.

Monitor cloud utilization

A quick, dependable, and expandable route is to offer AI services through cloud-based AI applications and APIs. Customers and business users alike benefit from always-on AI from a service provider, but costs can rise suddenly. Everyone will use your AI service if it is well-liked by all.

Many businesses that began their AI journeys entirely in the cloud are returning workloads to their on-premises and co-located infrastructure that can function well there. Pay-as-you-go infrastructure-as-a-service is becoming a competitive option for cloud-native enterprises with minimal or no on-premises infrastructure in comparison to rising cloud costs.

You have choices when it comes to Gen AI. Generative AI is surrounded by a lot of hype and mystery, giving the impression that it’s a cutting-edge technology that’s only accessible to the wealthiest companies. Actually, on a typical CPU-based data center or cloud instance, hundreds of high-performance models, including LLMs for generative AI, are accurate and performant. Enterprise-grade generative AI experimentation, prototyping, and deployment tools are rapidly developing in both open-source and proprietary communities.

By utilizing all of their resources, astute CIOs can leverage AI that transforms businesses without incurring the expenses and hazards associated with in-house development.

Technology

OPPO Reno 13 series will debut in China shortly, with India following in 2025

Published

on

According to reports, OPPO, a Chinese firm, is getting ready to introduce its Reno 13 series smartphones in its native nation this month. As per 91Mobiles, the OPPO Reno 13 and Reno 13 Pro models are anticipated to debut in China on November 25. The Indian launch is probably set for January 2025. The smartphone series that debuted in July of this year, the Reno 12 series, will be replaced by the Reno 13 series.

Information regarding the specifications of the new Reno 13 and Reno 13 Pro smartphones has leaked online, although the business has not yet confirmed the launch date. These are the specifics:

OPPO Reno 13 Series: Anticipations

It is anticipated that the OPPO Reno 13 Pro would have a 6.78-inch, quad-curved OLED screen with 1.5K resolution. In contrast, the slightly smaller 6.7-inch display with FHD+ resolution is found on the OPPO Reno 12 Pro. In China, the Pro model is probably going to be powered by the MediaTek Dimensity 8350 chipset, while in India, it might have a different processor. A 50MP primary camera, an 8MP ultrawide sensor, and a 50MP telephoto sensor with 3x optical zoom are anticipated to be included in the OPPO Reno 13 Pro’s photographic setup. Most likely, the front camera will include a 50MP sensor.

With a 5,900mAh battery as opposed to the 5,000mAh battery on the Reno 12 Pro, the Reno 13 Pro is anticipated to significantly increase battery capacity. Additionally, it is anticipated that the smartphone would support both 50W wireless and 80W wired charging. Additionally, an IP68/IP69 designation for water and dust protection could increase its durability.

Although the price of the smartphones in the Reno 13 series is not well known, it is anticipated to be similar to that of its predecessor. For comparison, the 12GB RAM + 256GB storage version of the OPPO Reno 12 Pro launched at Rs 36,999, while the 8GB RAM + 256GB storage version of the vanilla model cost Rs 32,999.

OPPO Reno 13 Pro: Anticipated features

  • Display: 6.78-inch OLED, quad-curved, with a refresh rate of 120 Hz and a resolution of 1.5K
  • processor: MediaTek Dimensity 8350
  • rear camera: 50MP primary, 8MP ultra-wide, and 50MP telephoto (3x zoom)
  • front camera: 50MP
  • Battery: 5,900mAh
  •  Charging: 50W wireless and 80W wired
  • IP rating: IP68/IP69; operating system: ColorOS 15 based on Android 15

Continue Reading

Technology

Apple has released Final Cut Pro 11, an AI-powered program

Published

on

Apple introduced Final Cut X thirteen years ago. Considering that the video-editing program marked its 25th birthday this April, that represents just over half of its lifetime. Some have questioned whether the corporation has discreetly withdrawn the offering due to its multiple lifetimes in the consumer software industry.

Final Cut Pro finally reaches level 11, after 13 years of waiting, and Apple is no longer playing around. On Wednesday, the program will be accessible for download. After a 90-day trial period, new users will need to pay $300 to buy Final Cut Pro 11 from the Mac App Store, while current users will receive it as a free update.

What specifically justified the much anticipated move to 11? AI is two letters. The business is using AI to power new features just weeks after releasing Apple Intelligence for iOS, iPadOS, and MacOS.

Magnetic Mask is at the top of the list because it makes it simple to crop objects and people out of videos without using a green screen.

According to Apple, “This powerful and precise automatic analysis provides additional flexibility to customize backgrounds and environments,” “Editors can also combine Magnetic Mask with color correction and video effects, allowing them to precisely control and stylize each project.”

Transcribe to Captions, which basically adds text to Final Cut’s timeline, is the second standout AI-based tool here. The company claims that its in-house large language model (LLM) powers that feature.

Apple’s problematic mixed-reality headset is the subject of this article’s other major headline. The most recent iPhones now have the capability to record Spatial Video, and Final Cut may be used to edit that footage. It is possible to add effects, color correct the video, and change the titles’ depth placement.

Apple is reportedly working on a more inexpensive variant, even though CEO Tim Cook has acknowledged that the $3,500 headgear isn’t the mainstream consumer product the company wanted. Along with the iPhone 15 Pro and all iPhone 16 models, the Vision Pro itself can record spatial video. Additionally, Canon just unveiled a new twin lens that works with R7 cameras.

Additionally, there are various time-saving features in the new Final Cut. For example, Magnetic Timeline allows you to swiftly rearrange clips while maintaining audio and video synchronization.

According to Apple, Final Cut Pro 11 was developed especially for the M-series of CPUs, which are its first-party silicon. This includes having more simultaneous 4K and 8K playback capabilities.

Apple claims that the M-series of chips, their first-party silicon, were the reason behind the creation of Final Cut Pro 11. This includes the capacity to play back several 4K and 8K ProRes video streams at once.

Final Cut Pro for iPad 2.1 is being released by Apple concurrently with the eagerly anticipated release of Pro 11. The brightness and color of the touched-based interface will be increased, and the workflow will be enhanced as well. Starting on Wednesday, current users can also obtain that for free.

Continue Reading

Technology

Apple has revealed a revamped Mac Mini with an M4 chip

Published

on

A smaller but no less powerful Mac Mini was recently unveiled by Apple as part of the company’s week of Mac-focused announcements. It now has Apple’s most recent M4 silicon, enables ray tracing for the first time, and comes pre-installed with 16GB of RAM, which seems to be the new standard in the age of Apple Intelligence. While the more potent M4 Pro model starts at $1,399, the machine still starts at $599 with the standard M4 CPU. The Mac Mini is available for preorder right now and will be in stores on November 8th, just like the updated iMac that was revealed yesterday.

The new design will be the first thing you notice. The Mini has reportedly been significantly reduced in size, although it was already a comparatively small desktop computer. It is now incredibly small, with dimensions of five inches for both length and width. Apple claims that “an innovative thermal architecture, which guides air to different levels of the system, while all venting is done through the foot” and the M4’s efficiency are the reasons it keeps things cool.

Nevertheless, Apple has packed this device with a ton of input/output, including a 3.5mm audio jack and two USB-C connections on the front. Three USB-C/Thunderbolt ports, Ethernet, and HDMI are located around the back. Although the USB-A ports are outdated, it’s important to remember that the base M2 Mini only featured two USB-A connectors and two Thunderbolt 4 ports. You get a total of five ports with the M4. You get an additional Thunderbolt port but lose native USB-A.

Depending on the M4 processor you select, those Thunderbolt connectors will have varying speeds. While the M4 Pro offers the most recent Thunderbolt 5 throughput, the standard M4 processor comes with Thunderbolt 4.

With its 14 CPU and 20 GPU cores, the M4 Pro Mac Mini also offers better overall performance. The standard M4 can have up to 32GB of RAM, while the M4 Pro can have up to 64GB. The maximum storage capacity is an astounding 8TB. Therefore, even though the Mini is rather little, if you have the money, you can make it really powerful. For those who desire it, 10 gigabit Ethernet is still an optional upgrade.

Apple has a big week ahead of it. On Monday, the company released the M4 iMac and its first Apple Intelligence software features for iOS, iPadOS, and macOS. (More AI functionality will be available in December, such as ChatGPT integration and image production.) As Apple completes its new hardware, those updated MacBook Pros might make their appearance tomorrow. The business will undoubtedly highlight its newest fleet of Macs when it releases its quarterly profits on Thursday.

Continue Reading

Trending

error: Content is protected !!