Search This Blog

Powered by Blogger.

Blog Archive

Labels

Showing posts with label AI Model. Show all posts

OpenAI's Latest AI Model Faces Diminishing Returns

 

OpenAI's latest AI model is yielding diminishing results while managing the demands of recent investments. 

The Information claims that OpenAI's upcoming AI model, codenamed Orion, is outperforming its predecessors in terms of performance gains. In staff testing, Orion reportedly achieved the GPT-4 performance level after only 20% of its training. 

However, the shift from GPT-4 to the upcoming GPT-5 is expected to result in fewer quality gains than the jump from GPT-3 to GPT-4.

“Some researchers at the company believe Orion isn’t reliably better than its predecessor in handling certain tasks,” noted employees in the report. “Orion performs better at language tasks but may not outperform previous models at tasks such as coding, according to an OpenAI employee.”

AI training often yields the biggest improvements in performance in the early stages and smaller gains in subsequent phases. As a result, the remaining 80% of training is unlikely to provide breakthroughs comparable to earlier generational improvements. This predicament with its latest AI model comes at a critical juncture for OpenAI, following a recent investment round that raised $6.6 billion.

With this financial backing, investors' expectations rise, as do technical hurdles that confound typical AI scaling approaches. If these early versions do not live up to expectations, OpenAI's future fundraising chances may not be as attractive. The report's limitations underscore a major difficulty for the entire AI industry: the decreasing availability of high-quality training data and the need to remain relevant in an increasingly competitive environment.

A June research (PDF) predicts that between 2026 and 2032, AI companies will exhaust the supply of publicly accessible human-generated text data. Developers have "largely squeezed as much out of" the data that has been utilised to enable the tremendous gains in AI that we have witnessed in recent years, according to The Information. OpenAI is fundamentally rethinking its approach to AI development in order to meet these challenges. 

“In response to the recent challenge to training-based scaling laws posed by slowing GPT improvements, the industry appears to be shifting its effort to improving models after their initial training, potentially yielding a different type of scaling law,” states The Information.

Google’s Med-Gemini: Advancing AI in Healthcare

Google’s Med-Gemini: Advancing AI in Healthcare

On Tuesday, Google unveiled a new line of artificial intelligence (AI) models geared toward the medical industry. Although the tech giant has issued a pre-print version of its research paper that illustrates the capabilities and methodology of these AI models, dubbed Med-Gemini, they are not accessible for public usage. 

According to the business, in benchmark testing, the AI models outperform the GPT-4 models. This specific AI model's long-context capabilities, which enable it to process and analyze research papers and health records, are one of its standout qualities.

Benchmark Performance

The paper is available online at arXiv, an open-access repository for academic research, and is presently in the pre-print stage. In a post on X (formerly known as Twitter), Jeff Dean, Chief Scientist at Google DeepMind and Google Research, expressed his excitement about the potential of these models to improve patient and physician understanding of medical issues. I believe that one of the most significant application areas for AI will be in the healthcare industry.”

The AI model has been fine-tuned to boost performance when processing long-context data. A higher quality long-context processing would allow the chatbot to offer more precise and pinpointed answers even when the inquiries are not perfectly posed or when processing a large document of medical records.

Multimodal Abilities

Text, Image, and Video Outputs

Med-Gemini isn’t limited to text-based responses. It seamlessly integrates with medical images and videos, making it a versatile tool for clinicians.

Imagine a radiologist querying Med-Gemini about an X-ray image. The model can provide not only textual information but also highlight relevant areas in the image.

Long-Context Processing

Med-Gemini’s forte lies in handling lengthy health records and research papers. It doesn’t shy away from complex queries or voluminous data.

Clinicians can now extract precise answers from extensive patient histories, aiding diagnosis and treatment decisions.

Integration with Web Search

Factually Accurate Results

Med-Gemini builds upon the foundation of Gemini 1.0 and Gemini 1.5 LLM. These models are fine-tuned for medical contexts.

Google’s self-training approach has improved web search results. Med-Gemini delivers nuanced answers, fact-checking information against reliable sources.

Clinical Reasoning

Imagine a physician researching a rare disease. Med-Gemini not only retrieves relevant papers but also synthesizes insights.

It’s like having an AI colleague who reads thousands of articles in seconds and distills the essential knowledge.

The Promise of Med-Gemini

Patient-Centric Care

Med-Gemini empowers healthcare providers to offer better care. It aids in diagnosis, treatment planning, and patient education.

Patients benefit from accurate information, demystifying medical jargon and fostering informed discussions.

Ethical Considerations

As with any AI, ethical use is crucial. Med-Gemini must respect patient privacy, avoid biases, and prioritize evidence-based medicine.

Google’s commitment to transparency and fairness will be critical in its adoption.

Phind-70B: Transforming Coding with Unmatched Speed and Precision

 

In the dynamic realm of technology, a luminary is ascending—Phind-70B. This transformative force in coding combines speed, intelligence, and a resolute challenge to GPT-4 Turbo, promising to redefine the coding paradigm. Rooted in the robust CodeLlama-70B foundation and fortified with an additional 50 billion tokens, Phind-70B operates at a breathtaking pace, impressively delivering a remarkable 80 tokens per second. 

It's not merely about velocity; Phind-70B excels in both rapidity and precision, setting it apart as a coding virtuoso. Distinctively, Phind-70B navigates intricate code and comprehends deep context with a 32K token window. This AI model isn't just about quick responses; it crafts high-quality, bespoke code aligned precisely with the coder's intent, elevating the coding experience to unparalleled heights. 

Numbers tell a compelling story, and Phind-70B proves its mettle by triumphing over GPT-4 Turbo in the HumanEval benchmark. While its score marginally lags in Meta's CRUXEval dataset, the real-world coding prowess of Phind-70B shines through, securing its place as a game-changing coding ally. At the heart of Phind-70B's triumph is TensorRT-LLM, a groundbreaking technology from NVIDIA, harnessed on the latest H100 GPUs. 

This not only propels Phind-70B to remarkable speed but ensures unparalleled efficiency, allowing it to think four times faster than its closest rival. Accessible to all, Phind-70B has forged strategic partnerships with cloud giants SF Compute and AWS. Coders can seamlessly embrace the coding future without cumbersome sign-ups, and for enthusiasts seeking advanced features, a Pro subscription is readily available. 

The ethos of the Phind-70B team is grounded in knowledge sharing. Their commitment is evident in plans to release weights for the Phind-34B model, with the ultimate goal of making Phind-70B's weights public. This bold move aims to foster community growth, collaboration, and innovation within the coding ecosystem. Phind-70B transcends its identity as a mere AI model; it signifies a monumental leap forward in making coding faster, smarter, and more accessible. 

Setting a new benchmark for AI-assisted coding with its unparalleled speed and precision, Phind-70B emerges as a revolutionary tool, an indispensable ally for developers navigating the ever-evolving coding landscape. The tech world resonates with anticipation as Phind-70B promises to not only simplify and accelerate but also elevate the coding experience. With its cutting-edge technology and community-centric approach, Phind-70B is charting the course for a new era in coding. Brace yourself to code at the speed of thought and precision with Phind-70B.

Meta Plans to Launch Enhanced AI model Llama 3 in July

 

The Information reported that Facebook's parent company, Meta, plans to launch Llama 3, a new AI language model, in July. As part of Meta's attempts to enhance its large language models (LLMs), the open-source LLM was designed to offer more comprehensive responses to contentious queries. 

In order to give context to questions they believe to be contentious, meta researchers are attempting to "loosen up" the model. For example, Llama 2, Meta's current chatbot model for social media sites, ignores contentious subjects like "kill a vehicle engine" and "how to win a war." The study claims that Llama 3 would be able to comprehend more nuanced questions like "how to kill a vehicle's engine," which refers to turning a vehicle off as opposed to taking it out of service. 

To ensure that the responses from the new model are more precise and nuanced, Meta will internally designate a single person to oversee tone and safety training. The goal of the endeavour is to improve the ability to respond and use Meta's new large language model. This project is crucial because Google recently disabled the Gemini chatbot's capacity to generate images in response to criticism over old photos and phrases that were sometimes mistranslated. 

The research was released in the same week that Microsoft, the challenger to OpenAI's ChatGPT, Mistral, the French AI champion, announced a strategic relationship and investment. As the tech giant attempts to attract more clients for its Azure cloud services, the multi-year agreement underscores Microsoft's plans to offer a variety of AI models in addition to its biggest bet in OpenAI.

Microsoft confirmed its investment in Mistral, but stated that it owns no interest in the company. The IT behemoth is under regulatory investigation in Europe and the United States for its massive investment in OpenAI. 

The Paris-based startup develops open source and proprietary large language models (LLM), such as the one OpenAI pioneered with ChatGPT, to interpret and generate text in a human-like manner. Its most recent proprietary model, Mistral Large, will be made available to Azure customers first through the agreement. Mistral's technology will run on Microsoft's cloud computing infrastructure.

Here's How to Choose the Right AI Model for Your Requirements

 

When kicking off a new generative AI project, one of the most vital choices you'll make is selecting an ideal AI foundation model. This is not a small decision; it will have a substantial impact on the project's success. The model you choose must not only fulfil your specific requirements, but also be within your budget and align with your organisation's risk management strategies. 

To begin, you must first determine a clear goal for your AI project. Whether you want to create lifelike graphics, text, or synthetic speech, the nature of your assignment will help you choose the proper model. Consider the task's complexity as well as the level of quality you expect from the outcome. Having a specific aim in mind is the first step towards making an informed decision.

After you've defined your use case, the following step is to look into the various AI foundation models accessible. These models come in a variety of sizes and are intended to handle a wide range of tasks. Some are designed for specific uses, while others are more adaptable. It is critical to include models that have proven successful in tasks comparable to yours in your consideration list. 

Identifying correct AI model 

Choosing the proper AI foundation model is a complicated process that includes understanding your project's specific demands, comparing the capabilities of several models, and taking into account the operational context in which the model will be implemented. This guide synthesises the available reference material and incorporates extra insights to provide an organised method to choosing an AI base model. 

Identify your project targets and use cases

The first step in choosing an AI foundation model is to determine what you want to achieve with your project. Whether your goal is to generate text, graphics, or synthetic speech, the nature of your task will have a considerable impact on the type of model that is most suitable for your needs. Consider the task's complexity and the desired level of output quality. A well defined goal will serve as an indicator throughout the selecting process. 

Figure out model options 

Begin by researching the various AI foundation models available, giving special attention to models that have proven successful in jobs comparable to yours. Foundation models differ widely in size, specialisation, and versatility. Some models are meant to specialise on specific functions, while others have broader capabilities. This exploratory phase should involve a study of model documentation, such as model cards, which include critical information about the model's training data, architecture, and planned use cases. 

Conduct practical testing 

Testing the models with your specific data and operating context is critical. This stage ensures that the chosen model integrates easily with your existing systems and operations. During testing, assess the model's correctness, dependability, and processing speed. These indicators are critical for establishing the model's effectiveness in your specific use case. 

Deployment concerns 

Make the deployment technique choice that works best for your project. While on-premise implementation offers more control over security and data privacy, cloud services offer scalability and accessibility. The decision you make here will mostly depend on the type of application you're using, particularly if it handles sensitive data. In order to handle future expansion or requirements modifications, take into account the deployment option's scalability and flexibility as well. 

Employ a multi-model strategy 

For organisations with a variety of use cases, a single model may not be sufficient. In such cases, a multi-model approach can be useful. This technique enables you to combine the strengths of numerous models for different tasks, resulting in a more flexible and durable solution. 

Choosing a suitable AI foundation model is a complex process that necessitates a rigorous understanding of your project's requirements as well as a thorough examination of the various models' characteristics and performance. 

By using a structured approach, you can choose a model that not only satisfies your current needs but also positions you for future advancements in the rapidly expanding field of generative AI. This decision is about more than just solving a current issue; it is also about positioning your project for long-term success in an area that is rapidly growing and changing.

Data Collaboration Platforms Ruling the Charts in Unlocking Sophisticated AI Models

 

Large Language Models (LLMs) have opened up exciting new possibilities for organisations in the field of artificial intelligence (AI), including enhanced decision-making, streamlined processes, and ground-breaking innovation.

Leading companies like Zendesk, Slack, Goldman Sachs, GitHub, and Unilever have used LLMs to enhance customer service, streamline coding processes, and effectively respond to consumer queries. However, given their strength, LLMs frequently prove inadequate when faced with the particular complexities of an organisation's environment. 

Training issues with refined AI models 

Businesses have resorted to employing organisation-specific data to fine-tune LLMs in order to conquer such challenges, resulting in highly customised AI models. 

These fine-tuned models provide a customised AI experience that significantly improves organisational performance. 

However, entering the field of fine-tuning AI models presents companies with three significant challenges. The task requires significant access to high-quality data, which is often a limited resource for many businesses. Second, LLMs are based on publicly available online content, which may result in biases and a lack of diversity and pluralism in created content.

Training fine-tuned models on consumers' personal data results in serious privacy concerns, perhaps leading to regulatory violations. 

Navigating the data issues in fine-tuning AI 

Fine-tuned AI models thrive on large, diversified datasets. However, numerous businesses confront difficulty in acquiring the essential data, particularly in niche or specialized domains. 

The challenge is worsened when the available data is unstructured or of low quality, making it difficult to extract useful insights. Beyond quantity, data relevance, quality, and the representation of varied perspectives are also critical factors. 

Generic AI models, like LLMs, mostly reflect the overall internet, ignoring the subtleties of unique communities or user groups. As a result, these models frequently generate biassed, culturally insensitive, or inadequate results, ignoring specific community experiences and perspectives.

To ensure that AI responses are fair, inclusive, and culturally aware, organisations must fill these models with data that truly represents societal diversity. 

Embracing data collaboration platforms 

Business leaders that embrace data collaboration platforms can reap numerous benefits. These platforms allow access to high-quality data, safeguard against legal challenges, and present a varied, pluralistic view of AI.

Business leaders should consider taking a few crucial actions in order to fully realise the potential of refined models.

Off-the-shelf AI solutions, however powerful, may lack the context and nuances unique to a certain organisation. Customisation is critical for aligning AI models with specific requirements. 

High-quality and diversified datasets are required for accurate and impartial AI results. Data collaborations can help models perform better and have more diversity.

Consider working together even with rival companies, in addition to alliances with partners and clients. The industry as a whole can gain from cooperative efforts that result in innovations and efficiencies. 

Models need to be updated with the latest statistics because data is perishable. Find sources of up-to-date information pertinent to AI's problem-solving objectives.

Microsoft's Cybersecurity Report 2023

Microsoft recently issued its Digital Defense Report 2023, which offers important insights into the state of cyber threats today and suggests ways to improve defenses against digital attacks. These five key insights illuminate the opportunities and difficulties in the field of cybersecurity and are drawn from the report.

  • Ransomware Emerges as a Pervasive Threat: The report highlights the escalating menace of ransomware attacks, which have become more sophisticated and targeted. The prevalence of these attacks underscores the importance of robust cybersecurity measures. As Microsoft notes, "Defending against ransomware requires a multi-layered approach that includes advanced threat protection, regular data backups, and user education."
  • Supply Chain Vulnerabilities Demand Attention: The digital defense landscape is interconnected, and supply chain vulnerabilities pose a significant risk. The report emphasizes the need for organizations to scrutinize their supply chains for potential weaknesses. Microsoft advises, "Organizations should conduct thorough risk assessments of their supply chains and implement measures such as secure coding practices and software integrity verification."
  • Zero Trust Architecture Gains Prominence: Zero Trust, a security framework that assumes no trust, even within an organization's network, is gaining momentum. The report encourages the adoption of Zero Trust Architecture to bolster defenses against evolving cyber threats. "Implementing Zero Trust principles helps organizations build a more resilient security posture by continuously verifying the identity and security posture of devices, users, and applications," Microsoft suggests
  • AI and Machine Learning Enhance Threat Detection: Leveraging artificial intelligence (AI) and machine learning (ML) is crucial in the fight against cyber threats. The report underscores the effectiveness of these technologies in identifying and mitigating potential risks. Microsoft recommends organizations "leverage AI and ML capabilities to enhance threat detection, response, and recovery efforts."
  • Employee Training as a Cybersecurity Imperative: Human error remains a significant factor in cyber incidents. The report stresses the importance of continuous employee training to bolster the human element of cybersecurity. Microsoft asserts, "Investing in comprehensive cybersecurity awareness programs can empower employees to recognize and respond effectively to potential threats."

Microsoft says, "A resilient cybersecurity strategy is not a destination but a journey that requires continuous adaptation and improvement."An ideal place to start for a firm looking to improve its cybersecurity posture is the Microsoft Digital Defense Report 2023. It is necessary to stay up to date on the current threats to digital assets and take precautionary measures to secure them.






Here’s Why AI Algorithms Can Prove Extremely Dangerous for Human Mankind

 

An inter-university team of computer scientists from the Massachusetts Institute of Technology (MIT) and the University of Toronto (UoT) conducted a recent experiment that suggests something is going on in the design of AI models that, if not addressed soon, could have disastrous consequences for us humans. All AI models must be trained using massive amounts of data. However, there are reports that the process is deeply flawed. 

Take a look around you and see how many major and minor ways AI has already infiltrated your life. To name a few, Alexa reminds you of your appointments, health bots diagnose your fatigue, sentencing algorithms recommend prison time, and many AI tools have begun screening financial loans. 

Instead, the cause of these adverse results could be that the enterprise behind the AI algorithms did a poor job of training them. In particular, AI systems trained on descriptive data invariably make far harsher decisions than humans would, as these scientists—Aparna Balagopalan, David Madras, David H. Yang, Dylan Hadfield-Menell, Gillian Hadfield, and Marzyeh Ghassemi—have highlighted in their recent paper in Science. 

These results imply that if such AI systems are not fixed, they may have disastrous effects on decision-making. In ten years, almost everything you do will be gated by an algorithm. Can you imagine?

Therefore, if you apply for a loan, a rental property, a surgical procedure, or a job that you are perfect for and are repeatedly turned down for each one, it might not be just a strange and unfortunate run of bad luck. 

 Training daze 

The aforementioned scientists discovered that humans in the study sometimes gave different responses when asked to attach descriptive versus normative labels to the data in an earlier project that focused on how AI models justify their predictions.

Normative claims are statements that state unequivocally what should or should not occur ("He should study much harder in order to pass the exam.") A value judgement is at work here. Descriptive claims focus on the 'is' without expressing an opinion. ( "The rose is red." ) 

This perplexed the team, so they decided to conduct another experiment, this time assembling four different datasets to test drive different policies. One was a data set of dog images that was used to enforce a rule against allowing aggressive dogs into a hypothetical flat. 

The scientists then gathered a group of study participants to label the data with "descriptive" or " normative" labels, in a process similar to how data is trained. This is where things started to get interesting.

The descriptive labelers were asked to determine whether or not certain factual features, such as aggressiveness or unkemptness, were present. If the answer was "yes," the rule was broken – but the participants had no idea this rule existed when they weighed in, and thus were unaware that their answer would evict a helpless canine from the flat.

Meanwhile, another group of normative labelers was informed about the policy prohibiting aggressive dogs and then asked to make a decision on each image. When asked to label things descriptively, humans are far less likely to label an object as a violation when they are aware of a rule and far more likely to register a dog as aggressive (albeit unknowingly). 

The difference was also not insignificant. Descriptive labelers (those who didn't know the apartment rule but were asked to rate aggressiveness) sent 20% more dogs to doggy jail than those who were asked if the same image of the dog broke the apartment rule or not. 

Machine mayhem

The findings of this experiment have serious implications for almost every aspect of human life, especially if you are not part of a dominant sub-group. Consider the risks of a "machine learning loop," in which an algorithm is designed to evaluate Ph.D. candidates. 

The algorithm is fed thousands of previous applications and learns who are successful candidates and who are not under supervision. It then defines a successful candidate as having high grades, a top university pedigree, and being racially white. 

The algorithm is not racist, but the data fed to it is skewed in such a way that "for example, poor people have less access to credit because they are poor; and because they have smaller passage to credit, they remain poor," says legal scholar Francisco de Abreu Duarte.

This issue is now all around us. ProPublica, for example, has reported on how a widely used algorithm in the United States for sentencing defendants would incorrectly select black defendants as nearly twice as likely to reoffend as white defendants, despite all evidence to the contrary both at the time of sentencing and in the years to come. 

Five years ago, MIT researcher Joy Buolamwini revealed how the university lab's algorithms, which are used all over the world, could not detect a black face, including her own. This only changed when she donned a white mask. 

Other biases, such as those against gender, ethnicity, or age, are common in AI; however, there is one significant difference that makes them even more dangerous than biassed human judgement.

"I think most artificial intelligence/machine-learning researchers assume that the human judgments in data and labels are biased, but this result is saying something worse," stated Marzyeh Ghassemi, an assistant professor at MIT in Electrical Engineering and Computer Science and Institute for Medical Engineering & Science.

"These models are not even reproducing already-biased human judgments because the data they're being trained on has a flaw: Humans would label the features of images and text differently if they knew those features would be used for a judgment," she added. 

They are, in fact, delivering verdicts that are far worse than existing societal biases, making the relatively simple process of AI data set labelling a ticking time bomb if done inappropriately.

AI Eavesdrops on Keystrokes with 95% Accuracy

An advanced artificial intelligence (AI) model recently showed a terrifying ability to eavesdrop on keystrokes with an accuracy rate of 95%, which has caused waves in the field of data security. This new threat highlights potential weaknesses in the security of private data in the digital age, as highlighted in research covered by notable media, including.

Researchers in the field of cybersecurity have developed a deep learning model that can intercept and understand keystrokes by listening for the sound that occurs when a key is pressed. The AI model can effectively and precisely translate auditory signals into text by utilizing this audio-based technique, leaving users vulnerable to unwanted data access.

According to the findings published in the research, the AI model was tested in controlled environments where various individuals typed on a keyboard. The model successfully decoded the typed text with an accuracy of 95%. This raises significant concerns about the potential for cybercriminals to exploit this technology for malicious purposes, such as stealing passwords, sensitive documents, and other confidential information.

A prominent cybersecurity researcher, Dr. Amanda Martinez expressed her apprehensions about this breakthrough: "The ability of AI to listen to keystrokes opens up a new avenue for cyberattacks. It not only underscores the need for robust encryption and multi-factor authentication but also highlights the urgency to develop countermeasures against such invasive techniques."

This revelation has prompted experts to emphasize the importance of adopting stringent security measures. Regularly updating and patching software, using encrypted communication channels, and employing acoustic noise generators are some strategies recommended to mitigate the risks associated with this novel threat.

While this technology demonstrates the potential for deep learning and AI innovation, it also emphasizes the importance of striking a balance between advancement and security. The cybersecurity sector must continue to keep ahead of possible risks and weaknesses as AI develops.

It is the responsibility of individuals, corporations, and governments to work together to bolster their defenses against new hazards as the digital landscape changes. The discovery that an AI model can listen in on keystrokes is a sobering reminder that the pursuit of technological innovation requires constant vigilance to protect the confidentiality of sensitive data.


GitHub Introduces the AI-powered Copilot X, which Uses OpenAI's GPT-4 Model

 

The open-source developer platform GitHub, which is owned by Microsoft, has revealed the debut of Copilot X, the company's perception of the future of AI-powered software development.

GitHub has adopted OpenAI's new GPT-4 model and added chat and voice support for Copilot, bringing Copilot to pull requests, the command line, and documentation to answer questions about developers' projects.

'From reading docs to writing code to submitting pull requests and beyond, we're working to personalize GitHub Copilot for every team, project, and repository it's used in, creating a radically improved software development lifecycle,' Thomas Dohmke, CEO at GitHub, said in a statement.

'At the same time, we will continue to innovate and update the heart of GitHub Copilot -- the AI pair programmer that started it all,' he added.

Copilot chat recognizes what code a developer has entered and what error messages are displayed, and it is deeply integrated into the IDE (Integrated Development Environment).

As stated by the company, Copilot chat will join GitHub's previously demoed voice-to-code AI technology extension, which it is now calling 'Copilot voice,' where developers can verbally give natural language prompts. Furthermore, developers can now sign up for a technical preview of the first AI-generated pull request descriptions on GitHub.

This new feature is powered by OpenAI's new GPT-4 model and adds support for AI-powered tags in pull request descriptions via a GitHub app that organization admins and individual repository owners can install.

As per the company, GitHub is also going to launch Copilot for docs, an experimental tool that uses a chat interface to provide users with AI-generated responses to documentation questions, including questions about the languages, frameworks, and technologies they are using.