Oracle has partnered with the Tamil Nadu Skill Development Corporation (TNSDC) to launch a specialised skills development program aimed at empowering over 200,000 students in the state with the latest IT skills.
This collaboration falls under Tamil Nadu’s ambitious skill enhancement initiative, Naan Mudhalvan, which focuses on equipping youth with employment-linked training in cutting-edge technologies.
The program, launched in 2024, targets students across various educational backgrounds, including engineering, arts, and science streams, from more than 900 colleges statewide.
These students will receive training and certification in modern technologies such as Oracle Cloud Infrastructure (OCI), data science, artificial intelligence (AI), machine learning (ML), and blockchain.
“Tamil Nadu, with its burgeoning youth population, is a priority state for us. Through Naan Mudhalvan, we aim to provide a platform for youth to enhance their career prospects,” said Tmt J. Innocent Divya, managing director of TNSDC.
The initiative, already reaching 1.3 million students, aims to leverage Oracle’s expertise to train an additional 200,000 students, opening doors to promising career opportunities.
The training curriculum, integrated into college curricula, will be delivered by experienced educators and industry professionals. Digital learning modules via Oracle MyLearn, the company’s robust training platform, will complement on-campus instruction.
This blended approach ensures comprehensive learning in cloud computing and related domains, tailored to individual learning levels and career aspirations.
Shailender Kumar, senior vice president and regional managing director of Oracle India and NetSuite JAPAC, expressed enthusiasm about the collaboration. “Oracle’s Skills Development Initiative aligns with the Indian government’s digital economy vision. Our certification programs are industry-recognized, enhancing employability and stability for tech professionals,” he said.
Oracle University, renowned for its technology training and certification programs, has already impacted over 3.1 million certified professionals globally.
The post Oracle to Train 200,000 Students in AI and Machine Learning in Tamil Nadu appeared first on AIM.
This year’s most-awaited Apple WorldWide Developers Conference (WWDC 24) marked a significant shift for the company, with the opening address almost entirely dedicated to the theme of AI or ‘Apple Intelligence’.
After falling behind significant competitors such as OpenAI, Google, and Microsoft, Apple introduced a host of AI features across its most popular operating systems. The company also revealed this year’s software updates for the iPhone, iPad, Apple Watch, Mac, and Vision Pro.
Contrary to expectations, Apple did not unveil any new hardware during the event, leaving many enthusiasts and developers disappointed.
What Did They Miss Out?
WWDC is a developer-focused event, which Apple has previously used to unveil new goods, such as the Apple Vision Pro and 15-inch MacBook Air showcased last year.
Those who expected Apple to release its next M4 chip for the MacBook Pro and Mac platform at WWDC 2024 were the most disappointed. At WWDC 2023, the firm also unveiled the M2 Max/Ultra Mac Studio M2 Ultra Mac Pro. It is still being determined when the M4 will be available for Macs, particularly the MacBook Pro.
Prior to WWDC 2024, there were talks that Apple may bring a significant redesign to iOS 18 to bring it more in line with what visionOS looks like. Unfortunately though, it didn’t happen.
Several anticipated hardware updates did not find a mention at WWDC, including the future of Apple’s HomePod line, upgrades to AirTag 2 and AirPods Max 2, and the potential unveiling of affordable Apple TV devices.
As with the iPhone hardware, Apple made no remarks about the upcoming Apple Watch model. Users only got a sneak peak at what’s in store with watchOS 11.
These missed opportunities have left many wondering about Apple’s hardware strategy.
Also, there has been much discussion over Apple not introducing a subscription service. However, many believe Apple needs to gather more user feedback on Apple Intelligence before putting any new capabilities behind a barrier (vs. free now).
Furthermore, during the keynote, there was no clarification on the economics of the OpenAl relationship. Apart from building in-house Apple Intelligence, the iPhone maker has partnered with OpenAI to integrate ChatGPT-powered by GPT-4o into iOS.
It’s worth noting that ChatGPT isn’t integrated directly into Siri, rather, it’s gaining access to it. Further, information on how Apple Intelligence may affect Apple devices’ battery life was omitted.
Where is RCS?
During the one-hour and forty-seven-minute-long keynote, Apple briefly discussed iOS 18. Among the many new features, one of the most significant changes was the introduction of RCS or Rich Communication Service.
RCS was created as a replacement for SMS. RCS includes typing indicators and read receipts, enables lengthier communications, and supports higher-quality photos and videos. In other words, it’s very similar to iMessage.
RCS messaging was addressed briefly at the end of the ‘Messages’ presentation, among other upgrades. It was essentially a bullet point, among many others, leaving many questions unanswered like: What RCS capabilities will Apple truly support? When strictly will this be implemented? And, maybe most significantly, will SMS from Android users still be green?
Too Late to the Party
It’s no secret that Apple has been slipping in the AI race over the past few years. While Microsoft spent $10 billion on OpenAI to kickstart an AI renaissance, Apple was wearing Vision Pro blinkers. According to Bloomberg, it increased spending on AI research late last year, allocating a billion each year for the effort.
Apple, aware of its position in the AI race, has made a significant stride at this year’s WWDC. It has unveiled a multitude of AI implementations in its products, including the much-anticipated enhancement to Siri and the integration of AI capabilities into text writing and photo editing.
Apple showed several experiments with the Apple Pencil, which could smoothen your handwriting and finish arithmetic equations as you wrote them. Kelsey Peterson, the director of machine learning and AI at Apple, walked the attendees through a demo during the pre-recorded keynote.
These are UX curiosities, but they appear to be designed to demonstrate technology rather than solve real-world problems.
The post Where Did WWDC Go Wrong? appeared first on AIM.
Stability AI has been a key player in the artificial intelligence (AI) image generator space thanks to its open-source Stable Diffusion models, which set the bar for quality, customization, and speed. Now, the company is adding to its family of models with its most advanced text-to-image generator yet.
On Wednesday, Stability AI launched Stable Diffusion 3 Medium, which the company claims is its "most sophisticated" image generation model. The two-billion-parameter model boasts several upgrades from its predecessors, resulting in higher-quality generations.
Also: How to use Stable Diffusion AI to create amazing images
For example, the new model can overcome typically difficult tasks for image generators, including generating photorealistic images (even of hands and faces) and accurate text without artifacts or spelling errors. It can also adhere to complex prompts and understand spatial relationships, as seen in the image below.
According to the company, Stable Diffusion 3 Medium is a smaller model, making it a good candidate for running on both individual computing systems and enterprise-tier GPUs. Stability AI added that the model is also ideal for customization due to its ability to gather "nuanced details from small datasets."
Also: The best AI image generators of 2024: Tested and reviewed
Stable Diffusion 3 Medium's weights remain open-sourced and accessible to all users with a free non-commercial license via Hugging Face. Those interested in using the commercial model are encouraged to contact Stability AI for licensing information.
Stable Diffusion 3 Medium is available on Stability AI's API, Stable Assistant, the company's chatbot, and Discord via Stable Artisan.
Apple’s Worldwide Developer Conference (WWDC 2024) was all about ‘Apple Intelligence’ announcements. And, the internet just couldn’t seem to get over it.
While intense conversations centred around the security and privacy of Apple Intelligence, with Elon Musk even threatening to ban the devices at his companies, there’s no denying that with its impressive AI reveals, the company might be rendering various apps and startups useless all at once!
Apple Intelligence on the Hunt to Kill Apps
Unlike what some are saying, Apple Intelligence is not a ChatGPT wrapper and Apple did develop a lot of its own AI. The LLMs built into Apple Intelligence deliver deep natural language understanding, making so many of our day-to-day tasks faster and easier.
It powers brand-new writing tools that can rewrite, suggest, proofread, and summarise text for you whether you’re working on an article or a blog post.
With AI being deeply integrated into the Apple OS, you can access these tools automatically across mail, notes, Safari, Pages, Keynote and even your third-party apps.
Many are calling it a hard time for writing tools like Grammarly, QuillBot, copy.ai, and more. Then, we have Genmoji, or AI-generated emoji. Leveraging the power of Apple Intelligence, now you can create new emojis on your device for any moment or emotion.
You can literally type it like a prompt and the diffusion model will create that new emoji from scratch in the same style like all the other emojis. Now, with this by their side, why would anyone use apps like Newji and Bitmoji?
The ‘Image Playground’ feature allows users to generate images on-device and share them across apps, making it ideal for quickly responding to your friends with just the right image. Since Apple Intelligence understands personal context, AI-generated images can be personalised to users. So, you can also create illustrations with your face in them.
This serves the purpose of many GenAI image generator apps.
9. Apple's new Image playground allows users to generate images on-device and share them across appspic.twitter.com/yyFTB6crZL
— Rowan Cheung (@rowancheung) June 10, 2024
While impressive, the images are not yet close enough to those generated by Midjourney or Dall-E as these apps are more accurate to realistic images whereas Apple seems to have taken a playful route. Another interesting feature, Image Wand, lets users convert rough sketches into high quality images.
Built into Apple’s photos app is a cleanup tool which will identify background items. It will let you circle items you don’t like in your photos and just get rid of them super quick, challenging apps like Facetune.
Additionally, Apple went from not having a calculator to having the best calculator in iPads overnight, surpassing all the other such apps.
If we're doing it, we're doing it the best way ~ Apple iPad didn't have built-in calculator until now pic.twitter.com/a0nbH1YpYD
— Shiva Rapolu (@shivarapolu01) June 11, 2024
This clearly seems to be the end of all the calculator apps out there.
Source: X
The Calculator app on iPad also introduced the Math Notes feature. You can use the Apple Pencil to write out expressions and have the Calculator app solve the equation in your notes! With this, notepad calculator apps like Soulver will have a hard time staying relevant.
With the new improvements in the WatchOS, Apple seems to be going after fitness apps like Whoop, Athlytic, Strava, and more. And, with hiking routes and directions now native in Apple Maps, it’s a direct competition to apps like AllTrails.
Apple now lets you record your call on iOS 18. You can easily transcribe your calls, turn them into summaries, and share them across your apps and emails. So, there’s no need to download a separate app anymore to record and transcribe a phone call. This makes apps like TapeACall irrelevant.
MacOS also finally got window snapping, making third-party apps like Spectacle, Magnet, and Rectangle redundant.
The new iPhone mirroring feature lets you put your iPhone right on your Mac and use any app. This might be the end of apps like Bezel.
You can now tap to pay for Apple Cash in iOS 18 without a phone number or email. This feature might hurt apps like Venmo, Revolut, and CashApp.
Venmo/ Revolut: we let friends to send money to each other easily Apple: hold my beer pic.twitter.com/byGvQaMTW7
— Ash Arora (@0xashesonchain) June 10, 2024
Finally, now that Apple Intelligence will allow Siri to have on-screen awareness and take actions on a user’s behalf, and, it might be game-over for Humane and Rabbit as well.
The post Apple is Killing More Apps Than OpenAI appeared first on AIM.
In 2023, India filed over 90,000 patents, which is an average of 247 patents per day, marking a notable 17% increase – the highest in two decades. However, it still does not rank in the top 10 countries for patent filings, raising the question of its position in the global innovation sector.
“Many patents attributed to Western companies are the result of work conducted in India, However, these patents are registered under the names of Western companies due to the global operations,” said TV Mohandas Pai, chairman at Aarin Capital, at the recently concluded IGIC 2024 in Bengaluru.
Pai highlighted that the Indian research culture does not consider patent filing as an objective, which has been a hindrance. While in the US, a significant number of patents are filed defensively driven by broader strategic reasons rather than for immediate utility.
“For any innovation system, there are three main ingredients – human capital, physical capital, and financial capital,” Pai noted.
India’s overall Digital Competitiveness Score of 60 (out of 100), places it ahead of all BRICS nations except China, reflecting the growing tech talent in the country. According to NASSCOM, India has approximately 4.5 million IT professionals, including those working in the software export industry.
“India is second only to the US in the number of highly skilled technology professionals. Notably, 60% of the professionals in the export industry work for American companies,” Pai mentioned.
Additionally, India has a significant digital infrastructure, with 1.31 billion mobile phone users and substantial data consumption. The country has achieved impressive milestones in digital public infractures, especially with UPI’s daily transactions expected to hit a billion by 2026.
However, Pai pointed out that India lacks cutting-edge research labs and a robust hardware industry, crucial for leading global innovation.
Hurdles to Cross
Coming down to financial capital, Pai noted that over the past decade, $145 billion has been invested in India, a contrast to China’s $835 billion and the US $2.3 trillion. The scarcity of financial capital is a significant constraint for India.
“Public funding for research is insufficient, with Indian universities receiving around $800 million annually, less than what a single prominent American university might receive,” he said.
In 2020, India filed nearly 57,000 patents, which is a mere 4% of the 1.49 million applications filed in China and 9.5% of the 5,97,000 applications in the US. Similarly, India grants 23,361 patents, compared to 5,30,000 in China and 3,50,000 in the US.
On an average, it takes about 58 months to process a patent application in India, while in China it takes about 20 months, and 23 in the US.
Source: World Intellectual Property Indicators 2021 report
The delay is primarily due to a shortage of manpower in the Indian patent office. As of the end of March 2022, the Indian patent employed only 860 people, including examiners and controllers.
In contrast, China employed 13,704 people, and the US employed 8,132. As of March 31, 2022, approximately 1,64,000 applications were pending at the controller level in India.
Slowly Catching Up
Despite the challenges, there have been some positive developments. In 2023, central universities like JNU, DU, Jamia, AMU, Maulana Azad National Urdu University, and UoH received an additional Rs 826.28 crore, increasing their total allocation from Rs 3,191.55 crore in 2021-22 to Rs 4,017.83 crore in 2022-23.
Similarly, if you look at the generative AI startups emerging in India, the numbers only indicate how far the country is progressing.
Currently, approximately 70 Indian startups are creating products and solutions focused on generative AI. It may not be too long before India is also considered an innovation capital of the world.
The post Why is India Not Leading the Innovation Sector? appeared first on AIM.
Are you one of the people who seemingly used the GPT Builder in Microsoft's AI-based Copilot Pro? If so, you can kiss it goodbye. Come July 10, Microsoft will squash the tool permanently — at least for the consumer version of Copilot.
On a new support page, Microsoft said that it will remove the ability to create GPTs starting July 10 and then delete all GPTs — those created by the company and its customers — from July 10 through July 14. In a nod toward privacy, any data associated with or collected by the GPTs also will be gone.
Also: How to use ChatGPT to create an app
Launched just three months ago in mid-March, Copilot's GPT Builder works similarly to OpenAI's ChatGPT Plus GPT Builder. Both tools let you create your own custom GPT chatbots to use yourself or share with others. Creating your own task-specific chatbot this way requires no coding skills or knowledge as the tool talks you through the process.
The difference is that the ChatGPT Plus GPT Builder lets you use your GPTs privately, share them with specific people, or publish them to the GPT store. With Copilot Pro's GPT Builder, you can use them only privately or share them with specific people.
OK, but why kill off the tool after only three months of life? Here's the explanation from Microsoft:
"We are continuing to evaluate our strategy for consumer Copilot extensibility and are prioritizing core product experiences, while remaining committed to developer opportunities," the company said on the support page. "To this end, we are shifting our focus on GPTs to Commercial and Enterprise scenarios and are stopping GPT efforts in consumer Copilot."
This makes it sound as if custom GPTs and the GPT Builder may have a future for commercial and enterprise customers, though Microsoft didn't elaborate on this possibility. Otherwise, the tool may have simply failed to catch on with subscribers, and the company might feel it's no longer worth the time and effort to support it.
Also: ChatGPT vs. Microsoft Copilot vs. Gemini: Which is the best AI chatbot?
In the meantime, what do you do if you've created custom Copilot GPTs and don't want to lose them? Here's what Microsoft advises:
Open your custom GPT in Edit mode and select the Configure tab. Copy the instructions and paste them somewhere else.
That last step seems vague since Microsoft didn't indicate how and where you could use those instructions. But at least, the data will be retained in some format. Certainly, you can save the information and then copy it back to a regular Copilot prompt when you want to run that particular request.
In today’s rapidly evolving world, leadership has transcended traditional boundaries. Leaders are no longer just stewards of organisational success but also torchbearers of societal change. As the world changes, companies must evolve or risk extinction, and embracing Diversity, Equity, and Inclusion (DEI) is key to this evolution.
DEI matters more than ever, especially in times of crisis when diverse teams prove to be more resilient and adaptable. They bring varied perspectives that help navigate complex challenges and drive innovation. Inclusive practices ensure that all team members feel valued and supported, boosting morale and productivity.
Equity ensures that the benefits of growth and success are shared by all, leading to a just and prosperous society.
DEI at Tesco: A Strategic Priority
For Tesco, DEI is a strategic priority because it aligns with our core values of treating people the way they want to be treated and creating a place where everybody is welcome. By fostering an inclusive workplace, Tesco harnesses the diverse perspectives and ideas of our colleagues, driving innovation and better decision-making. The CEO and senior leadership team visibly champion diversity and inclusion whilst setting bold ambitions on where we would like to be.
Mentoring and Sponsorship Programmes
Tesco has established active mentoring and sponsorship programs designed to support the professional growth and development of its colleagues. Through these initiatives, we pair colleagues with experienced mentors who provide guidance, share knowledge, and help mentees navigate their career paths. Sponsorship programs go a step further by having senior leaders advocate for their protégés, opening doors to new opportunities and facilitating career advancements.
These programs are integral to our commitment to fostering a culture of continuous learning and development. By investing in mentoring and sponsorship, Tesco not only empowers its workforce but also builds a pipeline of diverse and capable leaders for the future.
Embedding Inclusion in Everything We Do
Inclusivity starts with inclusive hiring. Inclusive job descriptions, active sourcing for diversity, and encouraging an inclusive approach to referrals are good starts. We are keen on fair employment practices, pay parity, strong anti-discrimination policies and being a disability-confident employer.
Inclusive Insurance Policies
Our comprehensive policies and initiatives show our commitment to fostering an inclusive and supportive environment. At Tesco Bengaluru, our inclusive approach begins with extensive insurance policies tailored to meet the diverse needs of our workforce. Recognising the importance of supporting colleagues through significant life changes, we proudly offer gender reassignment support, ensuring individuals undergoing transition receive the necessary care and respect that they deserve.
Acknowledging the evolving nature of family structures, our policies extend coverage to live-in partners and siblings. We are also committed to equality, providing coverage for same-gender partners and ensuring equal access to benefits for all colleagues.
Furthermore, our fertility treatment coverage supports those on the journey to parenthood. We also offer robust support for specially-abled colleagues and their family members. These inclusive policies highlight our dedication to the well-being and equality of our workforce, reinforcing our commitment to creating a supportive environment for all.
Listen, Learn, Act
Periodic feedback through employee surveys helps us measure how inclusive we are and guides us to take the right steps to drive it in a positive direction. Colleague-driven groups focussing on empowering women, ally networks and our male allies encourage everyone to bring their whole selves to work while fostering well-being and psychological safety.
Celebrating Diversity Through Events and Initiatives
At Tesco Bengaluru, we actively celebrate inclusion through dedicated events and initiatives that reflect our commitment to diversity and equality. Each year, we honour Pride Month, International Men’s Day, and International Women’s Day with a plethora of activities and discussions. These events celebrate our diverse workforce and promote understanding and respect among colleagues.
During the Inclusion Week, we engage in workshops, panel discussions, and cultural celebrations to reinforce our inclusive values. Comprehensive inclusion training for all colleagues ensures that everyone is equipped with the knowledge and skills to foster a welcoming environment.
Additionally, Disability Awareness Month allows us to focus on the contributions and needs of our specially-abled colleagues, ensuring they feel supported and valued. These celebrations and initiatives are more than just events; they are integral to our culture, underscoring our dedication to creating an inclusive and empowering workplace for everyone.
At Tesco Bengaluru, listening, learning, and acting are at the heart of our inclusive practices. Our comprehensive policies, celebratory events, and robust mentoring programs demonstrate our dedication to fostering an environment where everyone feels cherished. We believe that by embracing diversity and promoting equality, we can create a more innovative, dynamic, and successful workplace for all.
The post Transformative Leadership Strategies for Inclusive Innovation in Tech appeared first on AIM.
Nvidia CEO Jensen Huang surprised the audience at his keynote on Sunday for the Computex event by teasing the next chip architecture from his company, "Rubin."
Chip giant Nvidia has long dominated what's known as the "training" of neural networks, the compute-intensive task of fashioning and refashioning the neural "weights," or, "parameters," of the network until they reach optimal performance. The company has always had competition, from a variety of chip makers including giants such as Intel and Advanced Micro Devices, and startups such as Graphcore.
The latest benchmark tests of speed, however, suggest Nvidia really has no competition, if competition means parties that meaningfully challenge the best the company can do.
The MLCommons, the industry consortium that compiles multiple benchmark reports each year on AI chip performance, on Wednesday announced numbers for how different chips perform when training neural nets for a variety of tasks, including training Meta's Llama large language model to make predictions, and training the Stable Diffusion image model to produce pictures.
Nvidia swept the benchmarks, getting the top score, and also the second-best, in all nine competitions. Competitors such as AMD, Intel and Google's cloud computing division, didn't even come close.
It was the third time in a row that Nvidia had no competition for the top scores, but even the encouraging results of past rounds by competitors failed to materialize this time around.
The Training 4.0 test, which totals nine separate tasks, records the time to tune a neural network by having its settings refined in multiple experiments. It is one half of neural network performance, the other half being so-called inference, where the finished neural network makes predictions as it receives new data. Inference is covered in separate releases from MLCommons.
Most of MLPerf's tasks are by now well-established neural nets that have been in development for years, such as 3-D U-Net, a program for studying volumetric data for things such as solid tumor detection, which was introduced by Google's DeepMind back in 2016.
However, MLCommons continues to periodically update the benchmarks with new tasks to reflect emerging workloads. This round of training was the first time the submitters competed on the time to "fine-tune" a version of Meta's Llama language model, where the AI model is retrained, after its initial training, by using a more focused training data set. Also added was a "graph neural network" task, training the neural net to traverse a set of associated data points, which can be useful for things such as drug discovery.
In the test to fine-tune Meta's Llama 2 70B, Nvidia took just a minute and a half, with a collection of 1,024 of its "H100" GPU chips, a mainstream part currently powering AI workloads across the industry. The Nvidia chips scored the top twenty-three results, with Intel's "Guadi" AI accelerator showing up in twenty-fourth place.
Even when adjusted for the number of chips, substantial challenges fail to materialize. In eight-chip configurations, for example, which is more common than a 1,024-chip system, as far as enterprises are concerned — a configuration where Intel had promising results last summer — Intel's best — and only — submission this time around was for the Llama 70B task.
It took Intel's system, aided by two of Intel's XEON CPUs, 78 minutes to fine-tune Llama. An 8-way Nvidia H100 system, aided by two of Advanced Micro Devices' EPYC processor, assembled by open-source vendor Red Hat, took less than half the time, just over 31 minutes.
In the test to train OpenAI's GPT-3 for things such as chat, Intel was able to use just 1,024 Gaudi chips, a tenth of the number of chips Nvidia used, 11,616 H100s. But Intel's score, 67 minutes, took more than twenty times as long to train as Nvidia's leading score, 3.4 minutes. Of course, some enterprises may find that a difference of an hour to train versus three minutes is negligible considering the cost savings of using far fewer chips, and given that much of the work in training AI models can be factors other than the strict wall-clock time to train, such as the time required for data prep.
Other vendors had an equally hard time catching Nvidia. On the venerable test of image recognition, with the neural net known as Resnet, Advanced Micro Devices took 167 minutes to train the network using six of its "RADEON RX 7900 XTX" accelerator chips, versus just 122 minutes for a six-way Nvidia "GeForce RTX 4090" system.
Google's four submissions of its "TPU" version 5 chip, all for the GPT-3 test, achieved scores far below Nvidia's, between 12 and 114 minutes to train versus Nvidia's 3.4 minutes. Past competitors such as Graphcore have since bowed out of the race.
Also conspicuous in the results are Nvidia's dominance as a system vendor. All of its winning scores were made with systems engineered by Nvidia itself, even though a raft of system vendors participated, including Dell, Fujitsu, Hewlett Packard Enterprise, Juniper Networks, and Lenovo. Of the second-place scores won by Nvidia, three were built by the company's close partner in systems design, Supermicro.
An interesting future development could be systems using Nvidia's "Grace" CPU. All of the chip results submitted, whether from Nvidia, Intel, AMD or Google, continue to use only one of two x86 CPUs, Intel's XEON or AMD's EPYC. With Nvidia aiming to sell more complete computing systems using Grace, it seems only a matter of time before the CPU gets joined with Nvidia's GPUs. That could have an interesting impact on Nvidia's already substantial lead.
An interesting first this time around for the benchmark suite was the inclusion of a measurement of the energy consumed to train neural nets. The Singapore company Firmus Technologies featured results of its cloud platform, Sustainable Metal Cloud, running tens and hundreds of Nvidia H100s, and offered the total "energy-to-train" measured in Jules.
To run the Llama 2 70B fine-tuning, for example, Firmus's cloud computing system took between 45 million and 46 million Jules to train the network using 512 H100 chips. That training run, took two minutes, a little longer than Nvidia's best time on its own. It required four times as much energy as an 8-chip system that took fifteen times as long to train, or, 29 minutes, demonstrating the remarkable increase in energy consumed with giant training systems.
The cost of training AI has been a hot-button issue in recent years, both in terms of cost burden to companies in dollar terms, and the environmental burden. It remains to be seen whether other submitters to the MLPerf results will join Firmus in offering energy measurement in the next round of benchmarks.
Samsung on Wednesday unveiled an updated roadmap for its most advanced chip nodes catered for AI chips. The South Korean tech giant also introduced a new turnkey service that leverages its multiple chip business areas to entice companies such as Nvidia and AMD to use its foundry, or contract chip production, service for their AI chips.
The announcement marks the shifting focus of Samsung Foundry, the company's contract chip-making business unit, to chips for AI and high-performance computing (HPC) rather than processors on mobile devices.
Samsung Foundry's AI sales have increased by 80% over the past year and it was making significant strides in diversifying its customer base and application areas amid evolving market demand, the company noted. The tech giant aims to have over 50% of its foundry revenue brought in outside of mobile, it also said.
At Samsung Foundry Forum, its annual event for foundry themed this year as Empowering the AI Revolution, held at San Jose, the company showed off its new 2-nanometer (nm) and 4nm process nodes, called SF2Z and SF4U, respectively.
SF2Z incorporates what is called a backside power delivery network (BSPDN) to a conventional 2nm node (Samsung's SF2), where the power rails are placed on the backside of the wafer. Current chips have all the components in a chip such as memory, logic, and power rails on the front side of the wafer. Other contract chip makers are also preparing their own BSPDN technologies __ Intel calls it PoweVia and TSMC refers it to as Super PowerRail which they are also planning to adopt for their 2nm or under node chips.
According to Samsung, putting the power rails on the backside instead enhances the power, performance, and area (PPA) and voltage drop. SF2Z is aimed at high-performance computing (HPC) and AI chips and will roll out in 2027. Samsung already said previously that SF2 will launch in 2025 prior to the forum.
Samsung was the first to start production of a 3nm gate-all-around (GAA) node in 2022. At the forum, Samsung noted that its GAA process was maturing in terms of performance and yield and a second-generation 3nm node called SF3 will launch later this year. GAA will also be adopted to its 2nm node launching next year, the company said.
Samsung is adding various variants to its 2nm node, with a particular focus on AI. Image: Samsung
SF4U, meanwhile, is a variant of its 4nm, SF4 process that offers PPA boosts using optical shrink, where an existing die design by the customer is scaled down to fit into the newer node. This saves them in cost as no major architectural changes in the chip's design is required for them to migrate into the more advanced node. SF4U is launching in 2025 while Samsung already offers SF4 to customers. The South Korean tech giant also reiterated that its 1.4nm node (SF1.4) will launch in 2027 and was also preparing for chips below 1.4nm.
Samsung also unveiled its new turnkey foundry platform dubbed Samsung AI Solutions. This platform is offered together by Samsung's three business units in its chip division, Foundry, Memory, and Advanced Package (AVP).
According to the company, the platform integrates the "unique strengths" of each of these business units that will allow Samsung to offer solutions for customers tailored for the specific requirements for their AI chips. Overall, it will provide more bandwidth in a compact form factor, reduce power consumption, and improve signal integrity, Samsung said.
Because it is the only chipmaker that can offer memory chips to go with the customer's chip, manufacture these chips as well as package them, Samsung said this streamlines supply chain management and reduces time to market for the customer. The company said Samsung AI Solutions offers a 20% improvement in total turnaround time for customers. Samsung AI Solutions will be offered with co-packaged optics (CPO) __ where even the optics are packaged __ in 2027, the South Korean tech giant said.
Samsung's decision to add more variants to its process nodes makes sense as a strategy for AI, a general term that in fact encompasses a wide variety of different chips designed for different tasks and scales. Some are made for ChatGPT and other large language models. Some are for vision and image processing for use in drones and display devices. Others are AI accelerators for data centers.
According to market research firm Omdia, the global foundry market is expected to grow 18.1% a year on average from 2023 when it was worth $103.55 billion to 2027, when it is expected to be worth $201.28 billion. Growth in cutting-edge nodes of 3nm and under will have the most marked growth at 92.3% per year on average over the time period, the research firm's forecast noted.
Samsung is the world's largest memory chip producer and the second-largest contract chipmaker. Its Memory Business is attempting to win Nvidia as a customer for its HBM3E chips, which are used together with the GPU maker's AI accelerators.
Anthropic recently launched Claude in the European Union and updated its ToS (terms of service). The company highlighted policy refinements, high-risk use cases and certain disclosure requirements within its usage policy, possibly to align with the EU regulations.
Interestingly, the policy changes applied to users worldwide. Soon after, complaints about the model’s performance began surfacing from across the globe.
Why the Change?
Users noticed a marked change in the way Claude reacted to certain prompts and questioning. While there have been several theories as to why the company decided to shuffle things, the most believable seems to be that Anthropic is trying to anticipate the upcoming EU AI Act, thanks to its recent deployment in the region.
Like one Reddit user said, the rest “is just a cheap conspiracy. The new ToS is because they are finally deploying to the EU, and therefore need to comply with this,” pointing to the EU’s Artificial Intelligence Act (AIA).
Anthropic has gone all-in on creating a more holistic policy, ahead of their launch in the EU as well as more recently in Canada. However, other big tech companies have faced similar problems in the EU.
OpenAI, Meta and Others Follow
Now, Anthropic making overarching policy changes to fit in with EU standards isn’t unwarranted. The region has been notorious for cracking down on companies not following through with the regulations.
Case in point, OpenAI was recently in hot water when an Italian regulatory body accused the company of violating the EU privacy laws. In January this year, the company was subjected to a fact-finding initiative by Italy’s Data Protection Authority (DPA), where they alleged that user data had been used to train OpenAI’s ChatGPT.
This, they said, was in violation of the EU General Data Protection Regulation (GDPR).
Similarly, Meta updated its privacy policy, stating, “To properly serve our European communities, the models that power AI at Meta need to be trained on relevant information that reflects the diverse languages, geography and cultural references of the people in Europe who use them.”
However, this was also flagged by an Austrian privacy organisation, NYOB, stating that this also violated EU GDPR.
With countries in the EU closely following AI companies on how they implement their policies, Anthropic’s need for such a drastic change makes sense. But whether this change is doing good overall is up for debate.
How Bad is the Change?
As per the updated usage policy, Anthropic prohibits the usage of its services in compromising child safety, critical infrastructure, and personal identities. They have also barred making use of their products to create emotionally and psychological harmful content, as well as misinformation, including those used in elections.
There are several other changes made to the policy, as well as their ToS and privacy policies, including the right to request deletion of personal data and the option to opt out in case of data selling to third parties.
While most would be happy about stricter data privacy policies, users have reported that Claude is performing significantly worse this year. Particularly, with respect to the use cases in the updated usage policy.
“Some stuff that’s very open to interpretation or just outright dumb. Want to write some facts about the well-documented health risks of obesity? You’d be violating the “body shaming” rule. You can’t create anything that could be considered ‘emotionally harmful’,” one Reddit user said.
Further, they said that this would be worse to determine, considering there is no guarantee that those reviewing violations would be unbiased or neutral in terms of political misinformation.
Additionally, sexually-explicit content generation has also been significantly restricted. One user said that a story they had been working on with Claude had stopped progressing because Claude refused to continue, stating that it was uncomfortable with the prompt.
This was further backed by several users who stated the same issue, including one who said that Claude refused to comply with providing quotes from certain fictional characters, citing copyright infringement.
“You can’t ‘promote or advocate for a particular political candidate, party, issue or position’. Want to write a persuasive essay about an issue that can be construed as political? Better not use Claude,” they said.
What’s the Damage?
At the moment, users are willing to give both Claude and Anthropic the benefit of the doubt. With the updated policies, seemingly also due to the EU AI Act, Anthropic has made it easier to flag issues with their products and data privacy concerns.
This includes two emails, including one for Anthropic’s Data Protection Officer (DPO), to raise complaints or offer feedback, which was not present in the previous iteration of their policy.
Similarly, users believe that while Claude seems to have been handicapped by the new ToS, this could be reverted if given enough time and if the issues are raised by the users. “Anthropic does seem willing to listen to user feedback – and we’ve seen with the release of the Claude 3 models the dialling back of the refusals. So I think, at some point in the future, Anthropic will loosen up on things like that,” another user said.
Whether this can actually happen or if Anthropic will stick to its guns to preserve a user base in the EU and Canada is yet to be seen.
It’s no surprise to conclude that the noose is only tightening around big tech companies, and Claude seems to be the actual first in a long line of victims of over-regulation.
The post Anthropic’s Claude Performance Tanks After EU Updates appeared first on AIM.