🌟 OCTO Revolutionises Cancer Treatment with AI 🌟

🔬 The Oncology Counterfactual Therapeutics Oracle (OCTO) represents a groundbreaking approach in personalized oncology, merging AI and biology to discover new cancer treatments.

🌐 What sets OCTO apart?

Multimodal Data Integration: OCTO analyzes diverse patient data, including proteins, genes, and tumor sequences.

Hypothetical Simulations: It answers critical questions like “What if we increase gene X expression?”

Innovative Learning: OCTO detects cell structures independently and excels in zero-shot learning.

🚀 Practical Impact:

• Predicts new drug targets and tests them in vivo.

• Enhances our understanding of tumor-immune interactions.

🌟 As we advance, AI models like OCTO will play a crucial role in developing personalised, effective cancer therapies.

About OCTO.

🚫 Meta Faces Regulatory Roadblock in Europe: No Multimodal AI Models for EU

Meta will withhold its next multimodal LLaMA AI model — and future ones — from customers in the European Union due to regulatory uncertainties. This decision sets the stage for a confrontation between Meta and EU regulators and highlights a trend of U.S. tech giants withholding products from European markets.

💬 “We will release a multimodal Llama model over the coming months, but not in the EU due to the unpredictable nature of the European regulatory environment,” Meta stated. This move impacts European companies and prevents non-EU companies from offering products in Europe that utilize these models.

🔍 Meta’s issue isn’t with the still-being-finalized AI Act, but rather with how it can train models using data from European customers while complying with GDPR — the EU’s existing data protection law. Meta announced in May its intention to use publicly available posts from Facebook and Instagram to train future models, offering EU users a means to opt out. Despite briefing EU regulators months in advance and receiving minimal feedback, Meta was ordered to pause the training in June, leading to numerous questions from data privacy regulators.

🇬🇧 Interestingly, Meta does not face the same level of regulatory uncertainty in the U.K., which has nearly identical GDPR laws, and plans to launch its new model there.

🌍 The broader picture reveals escalating tensions between U.S.-based tech companies and European regulators, with tech firms arguing that stringent regulations harm both consumers and the competitiveness of European companies.

🔑 A Meta representative highlighted the importance of training on European data to ensure products accurately reflect regional terminology and culture, noting that competitors like Google and OpenAI are already doing so.

Meta will not launch multimodal Llama AI model in EU

ChatGPT-4o Mini: High-Performance AI at a Fraction of the Cost

OpenAI announced the launch of GPT-4o mini, OpenAI most cost-efficient small model yet! 💡

🔹 Cost-Effective Intelligence: Priced at just 15 cents per million input tokens and 60 cents per million output tokens, GPT-4o mini is an order of magnitude more affordable than previous models and over 60% cheaper than GPT-3.5 Turbo! 💸

🔹 Versatile and Powerful: With a context window of 128K tokens and support for up to 16K output tokens per request, GPT-4o mini excels in tasks like chaining multiple API calls, handling large volumes of context, and providing real-time text responses. It’s perfect for customer support chatbots, code analysis, and more! 🔍💬

🔹 Superior Performance: GPT-4o mini outperforms other small models on key benchmarks:

• 82.0% on MMLU (textual intelligence and reasoning)

• 87.0% on MGSM (math reasoning)

• 87.2% on HumanEval (coding performance)

• 59.4% on MMMU (multimodal reasoning)

🔹 Safety First: Built-in safety measures ensure reliable and safe responses, thanks to advanced techniques like reinforcement learning with human feedback (RLHF). 🛡️

🔹 Accessibility: Available now in the Assistants API, Chat Completions API, and Batch API. Free, Plus, and Team users can access GPT-4o mini today, with Enterprise users joining next week.

Why GPT-4o Mini Over GPT-4? 🆚

GPT-4o mini is designed to make AI more accessible and affordable without compromising on performance. It’s perfect for applications requiring low cost and latency, enabling developers to build scalable AI solutions efficiently. While GPT-4 offers more advanced capabilities, GPT-4o mini provides a high-performance alternative at a fraction of the cost, making it ideal for a broader range of use cases. 🌐

Get ready to unlock new possibilities with GPT-4o mini and join us in making AI accessible to all! 🌟

GPT-4o mini: advancing cost-efficient intelligence

🚀 Revolutionising Education with AI: Eureka Labs Launches! 🚀

Andrej Karpathy, former Tesla AI Director and OpenAI co-founder, has just announced his new venture: Eureka Labs! 🌟

Eureka Labs aims to revolutionize education by leveraging Generative AI to create a new kind of AI-native school. With an AI teaching assistant, students will be guided through course materials crafted by real teachers, making high-quality education accessible to all. 📚✨

Karpathy’s vision is to make it easier for anyone to learn anything, anytime, anywhere. His first course, LLM101n, will teach students how to train their own AI, blending digital and physical learning experiences. 🌐🤖

Karpathy’s extensive background in AI and education, from his time at Tesla to his YouTube tutorials and Stanford courses, makes this new venture one to watch! 👀

Stay tuned as Eureka Labs paves the way for a future where AI amplifies human potential. What would you like to learn next? 🤔

Andrej Karpathy unveils Eureka Labs, an AI education start-up

Florence-2, a cutting-edge vision foundation model with a unified, prompt-based representation


🚀 We introduce Florence-2, a cutting-edge vision foundation model with a unified, prompt-based representation for diverse computer vision and vision-language tasks. Unlike existing models, Florence-2 handles various tasks using simple text instructions, covering captioning, object detection, grounding, and segmentation. It relies on FLD-5B, a dataset with 5.4 billion visual annotations on 126 million images, created through automated annotation and model refinement. Florence-2 employs a sequence-to-sequence structure for training, achieving remarkable zero-shot and fine-tuning capabilities. Extensive evaluations confirm its strong performance across numerous tasks.

Florence-2: Advancing a Unified Representation for a Variety of Vision Tasks

Retrieval Augmented Iterative Self-Feedback (RA-ISF) refines RAG by breaking tasks into subtasks.

🚀 Problem: Large Language Models (LLMs) have static knowledge, making updates costly and time-consuming. Retrieval-augmented generation (RAG) helps, but irrelevant info can degrade performance.

🔧 Solution: Retrieval Augmented Iterative Self-Feedback (RA-ISF) refines RAG by breaking tasks into subtasks.

It uses:

1. Task Decomposition: Splits tasks into subtasks.

2. Knowledge Retrieval: Fetches relevant info for each subtask.

3. Response Generation: Integrates info to generate accurate answers.

🌟 What’s Next: RA-ISF reduces hallucinations and boosts performance, enhancing LLMs for complex tasks. As it evolves, expect more powerful, knowledge-enhanced LLMs.

RA-ISF: Learning to Answer and Understand from Retrieval Augmentation via Iterative Self-Feedback

MetRag – Similarity is Not All You Need: Endowing Retrieval Augmented Generation with Multi-Layered Thoughts

Abstract:

Recent advancements in large language models (LLMs) have significantly impacted various domains. However, the delay in knowledge updates and the presence of hallucination issues limit their effectiveness in knowledge-intensive tasks. Retrieval augmented generation (RAG) offers a solution by incorporating external information. Traditional RAG methods primarily rely on similarity to connect queries with documents, following a simple retrieve-then-read process. This work introduces MetRag, a framework that enhances RAG by integrating multi-layered thoughts, moving beyond mere similarity-based approaches. MetRag employs a utility model supervised by an LLM to generate utility-oriented thoughts and combines them with similarity-oriented thoughts for improved performance. It also uses LLMs as task-adaptive summarizers to condense retrieved documents, fostering compactness-oriented thought. This multi-layered approach culminates in a knowledge-augmented generation process, proving superior in extensive experiments on knowledge-intensive tasks.

Key Contributions:

Utility-Oriented Thought: Incorporates a small-scale utility model for better relevance.

Compactness-Oriented Thought: Utilizes LLMs to summarize large sets of retrieved documents.

Knowledge-Augmented Generation: Combines multiple thought layers for enhanced output.

Significance: MetRag demonstrates improved performance in knowledge-intensive tasks by addressing the limitations of traditional RAG methods through a multi-layered thought approach.

Applications: This framework can be applied to various domains requiring up-to-date and accurate knowledge, enhancing the reliability and efficiency of LLMs in real-world tasks.

Similarity is Not All You Need: Endowing Retrieval Augmented Generation with Multi Layered Thoughts

Microsoft presents SpreadsheetLLM

Overview:

The article discusses the introduction of SpreadsheetLLM by Microsoft, a new method for encoding spreadsheets to optimize the performance of large language models (LLMs) when processing spreadsheet data. Spreadsheets are inherently complex due to their two-dimensional grids, various layouts, and diverse formatting options, posing significant challenges for LLMs. SpreadsheetLLM addresses these challenges with a novel encoding method.

Key Innovations:

1. Vanilla Serialization Approach:

• Incorporates cell addresses, values, and formats.

• Limited by LLMs’ token constraints, making it impractical for large-scale applications.

2. SheetCompressor Framework:

Structural-Anchor-Based Compression: Reduces the complexity of the spreadsheet structure for easier processing by LLMs.

Inverse Index Translation: Efficiently maps compressed data back to its original format.

Data-Format-Aware Aggregation: Considers the formatting of data to maintain contextual understanding.

• Significantly improves performance in spreadsheet tasks, with a 25.6% enhancement in table detection in GPT-4’s in-context learning setting.

• Achieves an average compression ratio of 25 times and an F1 score of 78.9%, outperforming existing models by 12.3%.

3. Chain of Spreadsheet:

• Proposes a methodology for downstream tasks involving spreadsheet understanding.

• Validated in a new and demanding spreadsheet QA task.

Real Business Applications for Office Environments

Enhanced Data Analysis and Reporting:

Automated Insights Generation: SpreadsheetLLM can be used to automatically generate insights and reports from complex datasets, saving time and reducing the risk of human error in data analysis.

Improved Financial Modeling: Businesses can utilize the enhanced encoding capabilities to create more accurate financial models, forecasts, and budgeting tools.

Spreadsheet QA Automation: Implementing SpreadsheetLLM for quality assurance tasks can help in identifying errors, inconsistencies, and anomalies in large datasets, ensuring data integrity and reliability.

Streamlined Decision-Making:

Dynamic Dashboard Creation: SpreadsheetLLM can assist in creating dynamic and interactive dashboards that update in real-time, providing managers with up-to-date information for quick decision-making.

Enhanced Collaboration Tools: The improved understanding and compression of spreadsheets facilitate better integration with collaborative tools, allowing multiple users to work on and analyze data simultaneously.

Other Business Applications

Healthcare:

Patient Data Management: Healthcare providers can use SpreadsheetLLM to efficiently encode and analyze patient records, improving the accuracy of diagnoses and treatment plans.

Operational Efficiency: Hospitals can leverage the technology to streamline administrative tasks, such as scheduling, resource allocation, and inventory management.

Education:

Student Performance Analysis: Educational institutions can utilize SpreadsheetLLM to analyze student performance data, identify trends, and personalize learning experiences.

Administrative Automation: Automating administrative tasks like attendance tracking, grading, and scheduling, reducing the workload on educators.

Retail:

Inventory Management: Retail businesses can optimize their inventory management systems by using SpreadsheetLLM to analyze sales data and forecast demand.

Customer Insights: Analyzing customer data to gain insights into buying patterns and preferences, helping in targeted marketing and personalized offers.

Manufacturing:

Production Planning: Manufacturing companies can use SpreadsheetLLM to enhance their production planning processes, ensuring optimal resource utilization and minimizing downtime.

Quality Control: Implementing the technology for quality control tasks, identifying defects, and ensuring product consistency.

Finance:

Risk Assessment: Financial institutions can leverage SpreadsheetLLM to perform more accurate risk assessments and credit scoring.

Regulatory Compliance: Ensuring compliance with regulatory requirements by automating data validation and reporting tasks.

SpreadsheetLLM represents a significant advancement in the ability of LLMs to handle complex spreadsheet data, offering numerous applications across various industries to improve efficiency, accuracy, and decision-making processes.

The Vanilla Serialization Approach is a straightforward method for encoding spreadsheet data into a format that can be processed by large language models (LLMs). Here’s a detailed explanation of its components and limitations:

Key Components:

1. Cell Addresses:

• This refers to the specific location of each cell in the spreadsheet, typically denoted by a combination of letters and numbers (e.g., A1, B2, C3). By incorporating cell addresses, the method ensures that the positional information of data is preserved.

2. Values:

• These are the actual data entries within the cells, such as numbers, text, dates, or formulas. Including values is crucial as it represents the core content of the spreadsheet.

3. Formats:

• This includes the formatting information of each cell, such as font style, color, borders, number formats (e.g., currency, percentage), and conditional formatting. Preserving formatting helps maintain the contextual and visual understanding of the data.

Limitations:

1. Token Constraints:

• LLMs have a limited capacity to process data, often referred to as token constraints. A token is a unit of text that the model reads and processes, and each model has a maximum number of tokens it can handle at once. This limit can be a few thousand tokens, depending on the specific LLM.

2. Impractical for Large-Scale Applications:

• Spreadsheets can contain vast amounts of data, with potentially thousands of rows and columns. When each cell’s address, value, and format are serialized into a linear sequence, the total number of tokens can quickly exceed the LLM’s processing capacity.

• For instance, a spreadsheet with 1,000 rows and 50 columns results in 50,000 cells. If each cell’s address, value, and format contribute multiple tokens, the total number of tokens can become unmanageable, leading to truncation of data or incomplete processing.

• This limitation makes the vanilla serialization approach impractical for large or complex spreadsheets, as it cannot efficiently encode and process all the necessary information within the token constraints of LLMs.

In Summary:

The vanilla serialization approach attempts to capture the complete structure and content of a spreadsheet by including cell addresses, values, and formats. However, due to the token constraints of LLMs, this method becomes impractical for large-scale applications, where the volume of data exceeds the model’s processing capabilities. This necessitates the development of more efficient encoding methods, like SheetCompressor, to handle large and complex spreadsheets effectively.

The F1 score is a measure of a model’s accuracy in binary classification problems, providing a balance between precision and recall. It is particularly useful when the classes are imbalanced. Here’s a breakdown of the key components and the F1 score calculation:

Key Components:

1.  Precision:
•   Precision is the ratio of correctly predicted positive observations to the total predicted positives.
•   Formula:  Precision = True Positives / (True Positives + Positives) 
2.  Recall:
•   Recall (or Sensitivity) is the ratio of correctly predicted positive observations to all observations in the actual class.
•   Formula: Recall = True Positives / (True Positives + False Negatives 

F1 Score Calculation:

•   The F1 score is the harmonic mean of precision and recall.
•   Formula:  F1 Score = 2 x (Precision x Recall) / (Precision + Recall)

Interpretation:

•   The F1 score ranges from 0 to 1, where 1 indicates perfect precision and recall, and 0 indicates the worst performance.
•   It provides a single metric that balances both precision and recall, making it useful when you need to account for both false positives and false negatives.

Example:

If a model has:

•   80 True Positives (TP)
•   20 False Positives (FP)
•   10 False Negatives (FN)

Precision: 80 / (80 + 20) = 0.8
Recall: 80 / (80 + 10) = 0.888

F1 Score: 2 x (0.8 x 0.888) / (0.8 + 0.888) = approx 0.842

In the context of SpreadsheetLLM, a high F1 score (78.9%) indicates that the model is highly effective at accurately detecting and processing spreadsheet data, balancing both precision and recall in its performance.

SpreadsheetLLM: Encoding Spreadsheets for Large Language Models

Alibaba, Baidu, and ByteDance restrict AI access due to a massive chip shortage that could last for years. 🚫🔋

Kuaishou, after launching the test version of their AI model Kling, had to limit user access to prevent a shortage of computing power. ⚙️ Similarly, Moonshot AI has restricted access to its ChatGPT-like model Kimi during peak times, offering paid services for continued use. 💻💸

Alibaba Cloud stopped renting advanced Nvidia chips to regular clients, prioritizing top clients and supported AI startups. 🏢 Alibaba, Baidu, and ByteDance have also told corporate clients needing intensive, long-term AI use to wait in line. ⏳

The chip shortage impacts many Chinese AI startups dependent on Alibaba Cloud, which has invested in several key AI companies. 📉 While Chinese companies have made strides in AI development, they still rely heavily on Nvidia for chips. The US is pressuring chip manufacturers to halt sales to Huawei, further complicating the situation. 🇺🇸🔧

China’s AI companies are reportedly rationing the use of their services because they don’t have enough chips

OpenAI Stages of AI: Scale Ranks Progress Toward ‘Human-Level’ Problem Solving

OpenAI has introduced a five-level classification system to track its progress towards building artificial general intelligence (AGI). This new framework, shared internally with employees and soon with investors, outlines the path to developing AI capable of outperforming humans in most tasks.

Level 1: Chatbots – AI with conversational language abilities.

Level 2: Reasoners – Systems capable of human-level problem solving akin to a Ph.D. holder without tools.

Level 3: Agents – AI that can perform actions over several days.

Level 4: Innovators – AI that aids in invention and innovation.

Level 5: Organizations – AI performing the work of an entire organization.

Currently, OpenAI is at Level 1 but is close to achieving Level 2 with systems like GPT-4 demonstrating human-like reasoning in recent tests. This framework is part of OpenAI’s commitment to transparency and collaboration in the AI community.

For more details, check out Bloomberg’s report on OpenAI’s progress.

Superintelligence alignment refers to the process of ensuring that AI systems, particularly those that surpass human intelligence, act in ways that are consistent with human values and goals. This involves developing methods to guide and control these advanced AI systems so they behave safely and ethically, avoiding unintended consequences.

The concept of superintelligence alignment is critical because superintelligent AI systems could potentially make decisions and take actions that humans might not anticipate or understand. Ensuring alignment means that these AI systems will adhere to predefined ethical guidelines and objectives, preventing harm and ensuring their benefits are maximized for humanity.

Key aspects of superintelligence alignment include:

1. Control Mechanisms: Developing frameworks and techniques to steer AI behavior in desired directions, ensuring they follow human instructions and values.

2. Safety Protocols: Implementing measures to prevent AI from generating harmful or misleading outputs, such as reducing hallucinations where AI creates false information.

3. Collaboration and Transparency: Encouraging open research and collaboration among global experts to address alignment challenges comprehensively.

4. Technical Innovations: Creating AI models and methods that can guide more advanced systems, using simpler AI to supervise and correct the actions of more complex ones 


For more details on the Superalignment initiative, you can visit OpenAI’s Superalignment page.